commit
stringlengths 40
40
| old_file
stringlengths 4
150
| new_file
stringlengths 4
150
| old_contents
stringlengths 0
3.26k
| new_contents
stringlengths 1
4.43k
| subject
stringlengths 15
501
| message
stringlengths 15
4.06k
| lang
stringclasses 4
values | license
stringclasses 13
values | repos
stringlengths 5
91.5k
| diff
stringlengths 0
4.35k
|
---|---|---|---|---|---|---|---|---|---|---|
b8139440a2509d5b197889664f9ec34be9296210
|
form_designer/contrib/cms_plugins/form_designer_form/cms_plugins.py
|
form_designer/contrib/cms_plugins/form_designer_form/cms_plugins.py
|
from form_designer.contrib.cms_plugins.form_designer_form.models import CMSFormDefinition
from form_designer.views import process_form
from form_designer import settings
from cms.plugin_base import CMSPluginBase
from cms.plugin_pool import plugin_pool
from django.utils.translation import ugettext as _
class FormDesignerPlugin(CMSPluginBase):
model = CMSFormDefinition
module = _('Form Designer')
name = _('Form')
admin_preview = False
render_template = False
def render(self, context, instance, placeholder):
if instance.form_definition.form_template_name:
self.render_template = instance.form_definition.form_template_name
else:
self.render_template = settings.DEFAULT_FORM_TEMPLATE
# Redirection does not work with CMS plugin, hence disable:
return process_form(context['request'], instance.form_definition, context, disable_redirection=True)
plugin_pool.register_plugin(FormDesignerPlugin)
|
from form_designer.contrib.cms_plugins.form_designer_form.models import CMSFormDefinition
from form_designer.views import process_form
from form_designer import settings
from cms.plugin_base import CMSPluginBase
from cms.plugin_pool import plugin_pool
from django.utils.translation import ugettext as _
class FormDesignerPlugin(CMSPluginBase):
model = CMSFormDefinition
module = _('Form Designer')
name = _('Form')
admin_preview = False
render_template = False
cache = False
def render(self, context, instance, placeholder):
if instance.form_definition.form_template_name:
self.render_template = instance.form_definition.form_template_name
else:
self.render_template = settings.DEFAULT_FORM_TEMPLATE
# Redirection does not work with CMS plugin, hence disable:
return process_form(context['request'], instance.form_definition, context, disable_redirection=True)
plugin_pool.register_plugin(FormDesignerPlugin)
|
Disable caching so CSRF tokens are not cached.
|
Disable caching so CSRF tokens are not cached.
|
Python
|
bsd-3-clause
|
USGM/django-form-designer,USGM/django-form-designer
|
---
+++
@@ -14,6 +14,7 @@
name = _('Form')
admin_preview = False
render_template = False
+ cache = False
def render(self, context, instance, placeholder):
if instance.form_definition.form_template_name:
|
21f209b618850d15734c476bd3c1b359b9a7426e
|
infosystem/queue.py
|
infosystem/queue.py
|
import flask
from pika import BlockingConnection, PlainCredentials, ConnectionParameters
class RabbitMQ:
def __init__(self):
self.url = flask.current_app.config['ORMENU_QUEUE_URL']
self.port = flask.current_app.config['ORMENU_QUEUE_PORT']
self.virtual_host = \
flask.current_app.config['ORMENU_QUEUE_VIRTUAL_HOST']
self.username = flask.current_app.config['ORMENU_QUEUE_USERNAME']
self.password = flask.current_app.config['ORMENU_QUEUE_PASSWORD']
credentials = PlainCredentials(self.username, self.password)
self.params = ConnectionParameters(
self.url, self.port, self.virtual_host, credentials)
def connect(self):
try:
return BlockingConnection(self.params)
except Exception as e:
raise
class ProducerQueue:
def __init__(self, exchange, exchange_type):
rabbitMQ = RabbitMQ()
self.connection = rabbitMQ.connect()
self.exchange = exchange
self.channel = self.connection.channel()
self.channel.exchange_declare(
exchange=exchange, exchange_type=exchange_type, durable=True)
def publish(self, routing_key):
body = ""
self.channel.basic_publish(
exchange=self.exchange, routing_key=routing_key, body=body)
self.close()
def close(self):
self.channel.close()
self.connection.close()
|
import flask
from pika import BlockingConnection, PlainCredentials, ConnectionParameters
class RabbitMQ:
def __init__(self):
self.url = flask.current_app.config['INFOSYSTEM_QUEUE_URL']
self.port = flask.current_app.config['INFOSYSTEM_QUEUE_PORT']
self.virtual_host = \
flask.current_app.config['INFOSYSTEM_QUEUE_VIRTUAL_HOST']
self.username = flask.current_app.config['INFOSYSTEM_QUEUE_USERNAME']
self.password = flask.current_app.config['INFOSYSTEM_QUEUE_PASSWORD']
credentials = PlainCredentials(self.username, self.password)
self.params = ConnectionParameters(
self.url, self.port, self.virtual_host, credentials)
def connect(self):
try:
return BlockingConnection(self.params)
except Exception as e:
raise
class ProducerQueue:
def __init__(self, exchange, exchange_type):
rabbitMQ = RabbitMQ()
self.connection = rabbitMQ.connect()
self.exchange = exchange
self.channel = self.connection.channel()
self.channel.exchange_declare(
exchange=exchange, exchange_type=exchange_type, durable=True)
def publish(self, routing_key):
body = ""
self.channel.basic_publish(
exchange=self.exchange, routing_key=routing_key, body=body)
self.close()
def close(self):
self.channel.close()
self.connection.close()
|
Use INFOSYSTEM enviroment for Queue
|
Use INFOSYSTEM enviroment for Queue
|
Python
|
apache-2.0
|
samueldmq/infosystem
|
---
+++
@@ -5,12 +5,12 @@
class RabbitMQ:
def __init__(self):
- self.url = flask.current_app.config['ORMENU_QUEUE_URL']
- self.port = flask.current_app.config['ORMENU_QUEUE_PORT']
+ self.url = flask.current_app.config['INFOSYSTEM_QUEUE_URL']
+ self.port = flask.current_app.config['INFOSYSTEM_QUEUE_PORT']
self.virtual_host = \
- flask.current_app.config['ORMENU_QUEUE_VIRTUAL_HOST']
- self.username = flask.current_app.config['ORMENU_QUEUE_USERNAME']
- self.password = flask.current_app.config['ORMENU_QUEUE_PASSWORD']
+ flask.current_app.config['INFOSYSTEM_QUEUE_VIRTUAL_HOST']
+ self.username = flask.current_app.config['INFOSYSTEM_QUEUE_USERNAME']
+ self.password = flask.current_app.config['INFOSYSTEM_QUEUE_PASSWORD']
credentials = PlainCredentials(self.username, self.password)
self.params = ConnectionParameters(
self.url, self.port, self.virtual_host, credentials)
|
305ba7ee3fff41a7d866968c5332394301c0e83f
|
digi/wagtail_hooks.py
|
digi/wagtail_hooks.py
|
from wagtail.contrib.modeladmin.options import \
ModelAdmin, ModelAdminGroup, modeladmin_register
from .models import Indicator, FooterLinkSection
class IndicatorAdmin(ModelAdmin):
model = Indicator
menu_icon = 'user'
class FooterLinkSectionAdmin(ModelAdmin):
model = FooterLinkSection
menu_icon = 'redirect'
class DigiHelAdminGroup(ModelAdminGroup):
label = "DigiHel"
items = (IndicatorAdmin, FooterLinkSectionAdmin)
modeladmin_register(DigiHelAdminGroup)
|
from wagtail.contrib.modeladmin.options import \
ModelAdmin, ModelAdminGroup, modeladmin_register
from .models import Indicator, FooterLinkSection
from django.utils.html import format_html
from wagtail.wagtailcore import hooks
class IndicatorAdmin(ModelAdmin):
model = Indicator
menu_icon = 'user'
class FooterLinkSectionAdmin(ModelAdmin):
model = FooterLinkSection
menu_icon = 'redirect'
class DigiHelAdminGroup(ModelAdminGroup):
label = "DigiHel"
items = (IndicatorAdmin, FooterLinkSectionAdmin)
modeladmin_register(DigiHelAdminGroup)
# Enable editing of raw HTML
@hooks.register('insert_editor_js')
def enable_source_editing():
return format_html(
"""
<script>
registerHalloPlugin('hallohtml');
</script>
"""
)
|
Enable HTML source editing in the content editor
|
Enable HTML source editing in the content editor
|
Python
|
mit
|
terotic/digihel,City-of-Helsinki/digihel,terotic/digihel,City-of-Helsinki/digihel,City-of-Helsinki/digihel,terotic/digihel,City-of-Helsinki/digihel
|
---
+++
@@ -1,6 +1,8 @@
from wagtail.contrib.modeladmin.options import \
ModelAdmin, ModelAdminGroup, modeladmin_register
from .models import Indicator, FooterLinkSection
+from django.utils.html import format_html
+from wagtail.wagtailcore import hooks
class IndicatorAdmin(ModelAdmin):
@@ -18,3 +20,15 @@
items = (IndicatorAdmin, FooterLinkSectionAdmin)
modeladmin_register(DigiHelAdminGroup)
+
+
+# Enable editing of raw HTML
+@hooks.register('insert_editor_js')
+def enable_source_editing():
+ return format_html(
+ """
+ <script>
+ registerHalloPlugin('hallohtml');
+ </script>
+ """
+ )
|
c8fdc9aaea13567a293237e7a4126f83a52f0a22
|
salt/modules/test_virtual.py
|
salt/modules/test_virtual.py
|
# -*- coding: utf-8 -*-
'''
Module for running arbitrary tests with a __virtual__ function
'''
from __future__ import absolute_import
def __virtual__():
return False
def test():
return True
|
# -*- coding: utf-8 -*-
'''
Module for running arbitrary tests with a __virtual__ function
'''
from __future__ import absolute_import
def __virtual__():
return False
def ping():
return True
|
Fix mis-naming from pylint cleanup
|
Fix mis-naming from pylint cleanup
|
Python
|
apache-2.0
|
saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt
|
---
+++
@@ -9,5 +9,5 @@
return False
-def test():
+def ping():
return True
|
60497ba61c80863cd0414e39a9cd12b42b519897
|
chainer/training/extensions/value_observation.py
|
chainer/training/extensions/value_observation.py
|
from chainer.training import extension
import time
def observe_value(key, target_func):
"""Returns a trainer extension to continuously record a value.
Args:
key (str): Key of observation to record.
target_func (function): Function that returns the value to record.
It must take one argument: trainer object.
Returns:
The extension function.
"""
@extension.make_extension(
trigger=(1, 'epoch'), priority=extension.PRIORITY_WRITER)
def _observe_value(trainer):
trainer.observation[key] = target_func(trainer)
return _observe_value
def observe_time(key='time'):
"""Returns a trainer extension to record the elapsed time.
Args:
key (str): Key of observation to record.
Returns:
The extension function.
"""
start_time = time.time()
return observe_value(key, lambda _: time.time() - start_time)
def observe_lr(optimizer, key='lr'):
"""Returns a trainer extension to record the learning rate.
Args:
optimizer: Optimizer object whose learning rate is recorded.
key (str): Key of observation to record.
Returns:
The extension function.
"""
return observe_value(key, lambda _: optimizer.lr)
|
import time
from chainer.training import extension
def observe_value(key, target_func):
"""Returns a trainer extension to continuously record a value.
Args:
key (str): Key of observation to record.
target_func (function): Function that returns the value to record.
It must take one argument: trainer object.
Returns:
The extension function.
"""
@extension.make_extension(
trigger=(1, 'epoch'), priority=extension.PRIORITY_WRITER)
def _observe_value(trainer):
trainer.observation[key] = target_func(trainer)
return _observe_value
def observe_time(key='time'):
"""Returns a trainer extension to record the elapsed time.
Args:
key (str): Key of observation to record.
Returns:
The extension function.
"""
start_time = time.time()
return observe_value(key, lambda _: time.time() - start_time)
def observe_lr(optimizer, key='lr'):
"""Returns a trainer extension to record the learning rate.
Args:
optimizer: Optimizer object whose learning rate is recorded.
key (str): Key of observation to record.
Returns:
The extension function.
"""
return observe_value(key, lambda _: optimizer.lr)
|
Split system import and project import
|
Split system import and project import
|
Python
|
mit
|
cupy/cupy,chainer/chainer,keisuke-umezawa/chainer,wkentaro/chainer,okuta/chainer,ktnyt/chainer,wkentaro/chainer,jnishi/chainer,tkerola/chainer,keisuke-umezawa/chainer,jnishi/chainer,niboshi/chainer,delta2323/chainer,keisuke-umezawa/chainer,ktnyt/chainer,chainer/chainer,ysekky/chainer,chainer/chainer,keisuke-umezawa/chainer,niboshi/chainer,aonotas/chainer,rezoo/chainer,hvy/chainer,pfnet/chainer,okuta/chainer,jnishi/chainer,hvy/chainer,kiyukuta/chainer,cupy/cupy,niboshi/chainer,anaruse/chainer,hvy/chainer,chainer/chainer,niboshi/chainer,ronekko/chainer,jnishi/chainer,ktnyt/chainer,hvy/chainer,okuta/chainer,cupy/cupy,cupy/cupy,kashif/chainer,okuta/chainer,wkentaro/chainer,wkentaro/chainer,ktnyt/chainer
|
---
+++
@@ -1,5 +1,6 @@
+import time
+
from chainer.training import extension
-import time
def observe_value(key, target_func):
|
2b11762c418d1a985c3bd64d76abbf9b515bb8da
|
scanpointgenerator/compat.py
|
scanpointgenerator/compat.py
|
import os
try:
range_ = xrange
except NameError:
# For Python3
range_ = range
if os.name == 'java':
import scisoftpy as numpy
else:
import numpy
np = numpy
|
import os
try:
range_ = xrange
except NameError:
# For Python3
range_ = range
if os.name == 'java':
import numjy as numpy
else:
import numpy
np = numpy
|
Replace scisoftpy import with numjy
|
Replace scisoftpy import with numjy
|
Python
|
apache-2.0
|
dls-controls/scanpointgenerator
|
---
+++
@@ -8,7 +8,7 @@
if os.name == 'java':
- import scisoftpy as numpy
+ import numjy as numpy
else:
import numpy
|
1f22226575455349b3417cdae70a7a5b7181c3b3
|
tests/test_wikibugs2.py
|
tests/test_wikibugs2.py
|
import wikibugs
import configfetcher
import unittest
import os
p = os.path.split(__file__)[0]
class TestWikibugs(unittest.TestCase):
def setUp(self):
self.bugs = wikibugs.Wikibugs2(
configfetcher.ConfigFetcher()
)
def test_offline_scrape(self):
content = open(p + "/T87834", encoding="utf-8").read()
tags = self.bugs.get_tags(content)
self.assertSetEqual(set(tags), {
'§ Fundraising Sprint Devo',
'§ Fundraising Tech Backlog',
'Wikimedia-Fundraising',
'Wikimedia-Fundraising-CiviCRM',
})
self.assertSetEqual(set(next(iter(tags.values()))), {
'shade',
'disabled',
'uri',
'tagtype'
})
|
# encoding: utf-8
import wikibugs
import configfetcher
import unittest
import os
import requests
p = os.path.split(__file__)[0]
class TestWikibugs(unittest.TestCase):
def setUp(self):
self.bugs = wikibugs.Wikibugs2(
configfetcher.ConfigFetcher()
)
def run_scrape(self, content):
tags = self.bugs.get_tags(content)
self.assertSetEqual(set(tags), {
'§ Fundraising Sprint Devo',
'§ Fundraising Tech Backlog',
'Wikimedia-Fundraising',
'Wikimedia-Fundraising-CiviCRM',
})
self.assertSetEqual(set(next(iter(tags.values()))), {
'shade',
'disabled',
'uri',
'tagtype'
})
def test_offline_scrape(self):
content = open(p + "/T87834", encoding="utf-8").read()
self.run_scrape(content)
def test_online_scrape(self):
content = requests.get('https://phabricator.wikimedia.org/T87834').text
self.run_scrape(content)
|
Add online project scrape test
|
Add online project scrape test
Change-Id: Idc112f9397ac2f2fcfe51ca60259b257dff86e91
|
Python
|
mit
|
wikimedia/labs-tools-wikibugs2,wikimedia/labs-tools-wikibugs2
|
---
+++
@@ -1,7 +1,9 @@
+# encoding: utf-8
import wikibugs
import configfetcher
import unittest
import os
+import requests
p = os.path.split(__file__)[0]
@@ -12,8 +14,7 @@
configfetcher.ConfigFetcher()
)
- def test_offline_scrape(self):
- content = open(p + "/T87834", encoding="utf-8").read()
+ def run_scrape(self, content):
tags = self.bugs.get_tags(content)
self.assertSetEqual(set(tags), {
'§ Fundraising Sprint Devo',
@@ -27,3 +28,11 @@
'uri',
'tagtype'
})
+
+ def test_offline_scrape(self):
+ content = open(p + "/T87834", encoding="utf-8").read()
+ self.run_scrape(content)
+
+ def test_online_scrape(self):
+ content = requests.get('https://phabricator.wikimedia.org/T87834').text
+ self.run_scrape(content)
|
a774f3750b506c77f2c37d08b3d33dc66da2a12d
|
jsonmapping/util.py
|
jsonmapping/util.py
|
import os
import json
from jsonschema import Draft4Validator
def validate_mapping(mapping):
""" Validate a mapping configuration file against the relevant schema. """
file_path = os.path.join(os.path.dirname(__file__),
'schemas', 'mapping.json')
with open(file_path, 'rb') as fh:
validator = Draft4Validator(json.load(fh))
validator.validate(mapping)
return mapping
class RefScoped(object):
""" Objects which have a JSON schema-style scope. """
def __init__(self, resolver, scoped, scope=None, parent=None, name=None):
self.resolver = resolver
self._scoped = scoped
self._scope = scope or ''
self.name = name
self.parent = parent
@property
def id(self):
return self._scoped.get('id')
@property
def path(self):
if self.id is not None:
return self.id
if self.parent:
path = self.parent.path
if self.name:
if '#' not in path:
return path + '#/' + self.name
else:
return path + '/' + self.name
return path
@property
def scope(self):
if self.id:
return self.id
if self.parent:
return self.parent.scope
return self._scope
|
import os
import json
from jsonschema import Draft4Validator
def validate_mapping(mapping):
""" Validate a mapping configuration file against the relevant schema. """
file_path = os.path.join(os.path.dirname(__file__),
'schemas', 'mapping.json')
with open(file_path, 'r') as fh:
validator = Draft4Validator(json.load(fh))
validator.validate(mapping)
return mapping
class RefScoped(object):
""" Objects which have a JSON schema-style scope. """
def __init__(self, resolver, scoped, scope=None, parent=None, name=None):
self.resolver = resolver
self._scoped = scoped
self._scope = scope or ''
self.name = name
self.parent = parent
@property
def id(self):
return self._scoped.get('id')
@property
def path(self):
if self.id is not None:
return self.id
if self.parent:
path = self.parent.path
if self.name:
if '#' not in path:
return path + '#/' + self.name
else:
return path + '/' + self.name
return path
@property
def scope(self):
if self.id:
return self.id
if self.parent:
return self.parent.scope
return self._scope
|
Fix python 3 compat. BREAKS python 2.x
|
Fix python 3 compat. BREAKS python 2.x
|
Python
|
mit
|
Ramblurr/jsonmapping
|
---
+++
@@ -8,7 +8,7 @@
""" Validate a mapping configuration file against the relevant schema. """
file_path = os.path.join(os.path.dirname(__file__),
'schemas', 'mapping.json')
- with open(file_path, 'rb') as fh:
+ with open(file_path, 'r') as fh:
validator = Draft4Validator(json.load(fh))
validator.validate(mapping)
return mapping
|
64d17f591bfde49d3a7b5f49f987c1138eecebf8
|
tests/source/start_trace.py
|
tests/source/start_trace.py
|
"""Write some logs."""
import sys
import time
from mdk import start
mdk = start()
def main():
session = mdk.session()
session.info("process1", "hello")
time.sleep(1)
sys.stdout.write(session.inject())
sys.stdout.flush()
mdk.stop()
if __name__ == '__main__':
main()
|
"""Write some logs."""
import sys
import time
from mdk import start
mdk = start()
def main():
session = mdk.session()
session.info("process1", "hello")
time.sleep(5)
sys.stdout.write(session.inject())
sys.stdout.flush()
mdk.stop()
if __name__ == '__main__':
main()
|
Increase sleep in test to reduce timing sensitivity with prod servers.
|
Increase sleep in test to reduce timing sensitivity with prod servers.
|
Python
|
apache-2.0
|
datawire/mdk,datawire/mdk,datawire/mdk,datawire/mdk
|
---
+++
@@ -10,7 +10,7 @@
def main():
session = mdk.session()
session.info("process1", "hello")
- time.sleep(1)
+ time.sleep(5)
sys.stdout.write(session.inject())
sys.stdout.flush()
mdk.stop()
|
14ea472acfce8b5317a8c8c970db901501ea34c0
|
_tests/macro_testing/runner.py
|
_tests/macro_testing/runner.py
|
# -*- coding: utf-8 -*-
import os, os.path
import sys
import unittest
from macrotest import JSONSpecMacroTestCaseFactory
def JSONTestCaseLoader(tests_path, recursive=False):
"""
Load JSON specifications for Jinja2 macro test cases from the given
path and returns the resulting test classes.
This function will create a MacroTestCase subclass (using
JSONSpecMacrosTestCaseFactory) for each JSON file in the given path.
If `recursive` is True, it will also look in subdirectories. This is
not yet supported.
"""
json_files = [f for f in os.listdir(tests_path) if f.endswith('.json')]
for json_file in json_files:
# Create a camelcased name for the test. This is a minor thing, but I
# think it's nice.
name, extension = os.path.splitext(json_file)
class_name = ''.join(x for x in name.title() if x not in ' _-') + 'TestCase'
# Get the full path to the file and create a test class
json_file_path = os.path.join(tests_path, json_file)
test_class = JSONSpecMacroTestCaseFactory(class_name, json_file_path)
# Add the test class to globals() so that unittest.main() picks it up
globals()[class_name] = test_class
if __name__ == '__main__':
JSONTestCaseLoader('./tests/')
unittest.main()
|
# -*- coding: utf-8 -*-
import os, os.path
import sys
import unittest
from macrotest import JSONSpecMacroTestCaseFactory
def JSONTestCaseLoader(tests_path, recursive=False):
"""
Load JSON specifications for Jinja2 macro test cases from the given
path and returns the resulting test classes.
This function will create a MacroTestCase subclass (using
JSONSpecMacrosTestCaseFactory) for each JSON file in the given path.
If `recursive` is True, it will also look in subdirectories. This is
not yet supported.
"""
path = os.path.abspath(os.path.join(os.path.dirname( __file__ ), tests_path))
json_files = [f for f in os.listdir(path) if f.endswith('.json')]
for json_file in json_files:
# Create a camelcased name for the test. This is a minor thing, but I
# think it's nice.
name, extension = os.path.splitext(json_file)
class_name = ''.join(x for x in name.title() if x not in ' _-') + 'TestCase'
# Get the full path to the file and create a test class
json_file_path = os.path.join(path, json_file)
test_class = JSONSpecMacroTestCaseFactory(class_name, json_file_path)
# Add the test class to globals() so that unittest.main() picks it up
globals()[class_name] = test_class
if __name__ == '__main__':
JSONTestCaseLoader('./tests/')
unittest.main()
|
Make the paths not relative, so tests can be run from anywhere.
|
Make the paths not relative, so tests can be run from anywhere.
|
Python
|
cc0-1.0
|
kave/cfgov-refresh,kave/cfgov-refresh,kave/cfgov-refresh,kave/cfgov-refresh
|
---
+++
@@ -18,8 +18,9 @@
If `recursive` is True, it will also look in subdirectories. This is
not yet supported.
"""
+ path = os.path.abspath(os.path.join(os.path.dirname( __file__ ), tests_path))
- json_files = [f for f in os.listdir(tests_path) if f.endswith('.json')]
+ json_files = [f for f in os.listdir(path) if f.endswith('.json')]
for json_file in json_files:
# Create a camelcased name for the test. This is a minor thing, but I
# think it's nice.
@@ -27,7 +28,7 @@
class_name = ''.join(x for x in name.title() if x not in ' _-') + 'TestCase'
# Get the full path to the file and create a test class
- json_file_path = os.path.join(tests_path, json_file)
+ json_file_path = os.path.join(path, json_file)
test_class = JSONSpecMacroTestCaseFactory(class_name, json_file_path)
# Add the test class to globals() so that unittest.main() picks it up
|
90699f4fa6c1ce2b02e81a8fef9bfafd2175fa7f
|
kmapper/__init__.py
|
kmapper/__init__.py
|
from .kmapper import KeplerMapper
from .kmapper import cluster
from .kmapper import Cover
from .kmapper import GraphNerve
|
from .kmapper import KeplerMapper
from .kmapper import cluster
from .cover import Cover
from .nerve import GraphNerve
import pkg_resources
__version__ = pkg_resources.get_distribution('kmapper').version
|
Add __version__ variable to package
|
Add __version__ variable to package
|
Python
|
mit
|
MLWave/kepler-mapper,MLWave/kepler-mapper,MLWave/kepler-mapper
|
---
+++
@@ -1,4 +1,6 @@
from .kmapper import KeplerMapper
from .kmapper import cluster
-from .kmapper import Cover
-from .kmapper import GraphNerve
+from .cover import Cover
+from .nerve import GraphNerve
+import pkg_resources
+__version__ = pkg_resources.get_distribution('kmapper').version
|
884852eeb2dec07dccefc26595f097ec9ae8532b
|
forum/forms.py
|
forum/forms.py
|
from django.forms import ModelForm,Textarea
from .models import Post
class PostForm(ModelForm):
class Meta:
model = Post
fields = ('subject','body')
widgets = {
'body': Textarea(
attrs={
'data-provide':'markdown',
'data-hidden-buttons':'cmdHeading',
}),
}
|
from django.forms import ModelForm,Textarea,TextInput
from .models import Post
class PostForm(ModelForm):
class Meta:
model = Post
fields = ('subject','body')
widgets = {
'subject': TextInput(attrs={'autofocus':'autofocus'}),
'body': Textarea(
attrs={
'data-provide':'markdown',
'data-hidden-buttons':'cmdHeading',
}),
}
|
Add autofocus to subject field
|
Add autofocus to subject field
|
Python
|
mit
|
Kromey/fbxnano,Kromey/akwriters,Kromey/fbxnano,Kromey/akwriters,Kromey/fbxnano,Kromey/fbxnano,Kromey/akwriters,Kromey/akwriters
|
---
+++
@@ -1,4 +1,4 @@
-from django.forms import ModelForm,Textarea
+from django.forms import ModelForm,Textarea,TextInput
from .models import Post
@@ -9,6 +9,7 @@
model = Post
fields = ('subject','body')
widgets = {
+ 'subject': TextInput(attrs={'autofocus':'autofocus'}),
'body': Textarea(
attrs={
'data-provide':'markdown',
|
16553fd759e70d93824407f18cdea419703d85d4
|
gcn/metrics.py
|
gcn/metrics.py
|
import tensorflow as tf
def masked_softmax_cross_entropy(preds, labels, mask):
"""Softmax cross-entropy loss with masking."""
loss = tf.nn.softmax_cross_entropy_with_logits(logits=preds, lables=labels)
mask = tf.cast(mask, dtype=tf.float32)
mask /= tf.reduce_mean(mask)
loss *= mask
return tf.reduce_mean(loss)
def masked_accuracy(preds, labels, mask):
"""Accuracy with masking."""
correct_prediction = tf.equal(tf.argmax(preds, 1), tf.argmax(labels, 1))
accuracy_all = tf.cast(correct_prediction, tf.float32)
mask = tf.cast(mask, dtype=tf.float32)
mask /= tf.reduce_mean(mask)
accuracy_all *= mask
return tf.reduce_mean(accuracy_all)
|
import tensorflow as tf
def masked_softmax_cross_entropy(preds, labels, mask):
"""Softmax cross-entropy loss with masking."""
loss = tf.nn.softmax_cross_entropy_with_logits(logits=preds, labels=labels)
mask = tf.cast(mask, dtype=tf.float32)
mask /= tf.reduce_mean(mask)
loss *= mask
return tf.reduce_mean(loss)
def masked_accuracy(preds, labels, mask):
"""Accuracy with masking."""
correct_prediction = tf.equal(tf.argmax(preds, 1), tf.argmax(labels, 1))
accuracy_all = tf.cast(correct_prediction, tf.float32)
mask = tf.cast(mask, dtype=tf.float32)
mask /= tf.reduce_mean(mask)
accuracy_all *= mask
return tf.reduce_mean(accuracy_all)
|
Fix typo for named argument
|
Fix typo for named argument
|
Python
|
mit
|
tkipf/gcn
|
---
+++
@@ -3,7 +3,7 @@
def masked_softmax_cross_entropy(preds, labels, mask):
"""Softmax cross-entropy loss with masking."""
- loss = tf.nn.softmax_cross_entropy_with_logits(logits=preds, lables=labels)
+ loss = tf.nn.softmax_cross_entropy_with_logits(logits=preds, labels=labels)
mask = tf.cast(mask, dtype=tf.float32)
mask /= tf.reduce_mean(mask)
loss *= mask
|
1b7634e3a98919df5f2f4d54c57bb72dfbf308df
|
py3-test/tests.py
|
py3-test/tests.py
|
# -*- coding: utf-8 -*-
import nose.tools as nt
from asyncio import get_event_loop
from asyncio import sleep as async_sleep
from pyee import EventEmitter
def test_async_emit():
"""Test that event_emitters can handle wrapping coroutines
"""
ee = EventEmitter()
loop = get_event_loop()
class SenseWasCalled():
def __init__(self):
self.was_called = False
def am_calling(self):
self.was_called = True
def assert_was_called(self):
nt.assert_true(self.was_called)
sensor = SenseWasCalled()
@ee.on('event')
async def event_handler():
sensor.am_calling()
ee.emit('event')
loop.run_until_complete(async_sleep(1))
sensor.assert_was_called()
|
# -*- coding: utf-8 -*-
import nose.tools as nt
from asyncio import Future, gather, get_event_loop, sleep
from pyee import EventEmitter
def test_async_emit():
"""Test that event_emitters can handle wrapping coroutines
"""
loop = get_event_loop()
ee = EventEmitter(loop=loop)
future = Future()
@ee.on('event')
async def event_handler():
future.set_result(True)
async def create_timeout(loop=loop):
await sleep(1, loop=loop)
future.cancel()
timeout = create_timeout(loop=loop)
@future.add_done_callback
def _done(result):
nt.assert_true(result)
ee.emit('event')
loop.run_until_complete(gather(future, timeout))
|
Rewrite asyncio test to use futures
|
Rewrite asyncio test to use futures
|
Python
|
mit
|
jfhbrook/pyee
|
---
+++
@@ -1,8 +1,6 @@
# -*- coding: utf-8 -*-
-
import nose.tools as nt
-from asyncio import get_event_loop
-from asyncio import sleep as async_sleep
+from asyncio import Future, gather, get_event_loop, sleep
from pyee import EventEmitter
@@ -10,26 +8,25 @@
def test_async_emit():
"""Test that event_emitters can handle wrapping coroutines
"""
- ee = EventEmitter()
loop = get_event_loop()
+ ee = EventEmitter(loop=loop)
- class SenseWasCalled():
- def __init__(self):
- self.was_called = False
-
- def am_calling(self):
- self.was_called = True
-
- def assert_was_called(self):
- nt.assert_true(self.was_called)
-
- sensor = SenseWasCalled()
+ future = Future()
@ee.on('event')
async def event_handler():
- sensor.am_calling()
+ future.set_result(True)
+
+ async def create_timeout(loop=loop):
+ await sleep(1, loop=loop)
+ future.cancel()
+
+ timeout = create_timeout(loop=loop)
+
+ @future.add_done_callback
+ def _done(result):
+ nt.assert_true(result)
ee.emit('event')
- loop.run_until_complete(async_sleep(1))
- sensor.assert_was_called()
+ loop.run_until_complete(gather(future, timeout))
|
a8bb719061a68b5d322868768203476c4ee1e9b9
|
gnocchi/cli.py
|
gnocchi/cli.py
|
# Copyright (c) 2013 Mirantis Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
# implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from oslo.config import cfg
from gnocchi.indexer import sqlalchemy as sql_db
from gnocchi.rest import app
from gnocchi import service
def storage_dbsync():
service.prepare_service()
indexer = sql_db.SQLAlchemyIndexer(cfg.CONF)
indexer.upgrade()
def api():
service.prepare_service()
app.build_server()
|
# Copyright (c) 2013 Mirantis Inc.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
# implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from oslo.config import cfg
from gnocchi.indexer import sqlalchemy as sql_db
from gnocchi.rest import app
from gnocchi import service
def storage_dbsync():
service.prepare_service()
indexer = sql_db.SQLAlchemyIndexer(cfg.CONF)
indexer.connect()
indexer.upgrade()
def api():
service.prepare_service()
app.build_server()
|
Connect to database before upgrading it
|
Connect to database before upgrading it
This change ensure we are connected to the database before
we upgrade it.
Change-Id: Ia0be33892a99897ff294d004f4d935f3753e6200
|
Python
|
apache-2.0
|
idegtiarov/gnocchi-rep,leandroreox/gnocchi,sileht/gnocchi,idegtiarov/gnocchi-rep,gnocchixyz/gnocchi,sileht/gnocchi,idegtiarov/gnocchi-rep,gnocchixyz/gnocchi,leandroreox/gnocchi
|
---
+++
@@ -23,6 +23,7 @@
def storage_dbsync():
service.prepare_service()
indexer = sql_db.SQLAlchemyIndexer(cfg.CONF)
+ indexer.connect()
indexer.upgrade()
|
45fcbaa7515faaa2f812bd9ceb0800fbeb64f1e7
|
pytest_pycharm.py
|
pytest_pycharm.py
|
# -*- coding: utf-8 -*-
import threading
def pytest_exception_interact(node, call, report):
"""
Drop into PyCharm debugger, if available, on uncaught exceptions.
"""
try:
import pydevd
from pydevd import pydevd_tracing
except ImportError:
pass
else:
exctype, value, traceback = call.excinfo._excinfo
frames = []
while traceback:
frames.append(traceback.tb_frame)
traceback = traceback.tb_next
thread = threading.current_thread()
frames_by_id = dict([(id(frame), frame) for frame in frames])
frame = frames[-1]
thread.additionalInfo.exception = (exctype, value, traceback)
thread.additionalInfo.pydev_force_stop_at_exception = (frame, frames_by_id)
thread.additionalInfo.message = "test fail"
debugger = pydevd.debugger
debugger.force_post_mortem_stop += 1
pydevd_tracing.SetTrace(None)
debugger.handle_post_mortem_stop(thread.additionalInfo, thread)
return report
|
# -*- coding: utf-8 -*-
import threading
def pytest_exception_interact(node, call, report):
"""
Drop into PyCharm debugger, if available, on uncaught exceptions.
"""
try:
import pydevd
from pydevd import pydevd_tracing
except ImportError:
pass
else:
exctype, value, traceback = call.excinfo._excinfo
frames = []
while traceback:
frames.append(traceback.tb_frame)
traceback = traceback.tb_next
thread = threading.current_thread()
frames_by_id = dict([(id(frame), frame) for frame in frames])
frame = frames[-1]
thread.additionalInfo.exception = (exctype, value, traceback)
thread.additionalInfo.pydev_force_stop_at_exception = (frame, frames_by_id)
thread.additionalInfo.message = "test fail"
debugger = pydevd.debugger
if hasattr(debugger, "force_post_mortem_stop"):
debugger.force_post_mortem_stop += 1
pydevd_tracing.SetTrace(None)
debugger.handle_post_mortem_stop(thread.additionalInfo, thread)
return report
|
Fix for pycharm 4.0.0 changes
|
Fix for pycharm 4.0.0 changes
|
Python
|
bsd-3-clause
|
jlubcke/pytest-pycharm
|
---
+++
@@ -24,7 +24,8 @@
thread.additionalInfo.pydev_force_stop_at_exception = (frame, frames_by_id)
thread.additionalInfo.message = "test fail"
debugger = pydevd.debugger
- debugger.force_post_mortem_stop += 1
+ if hasattr(debugger, "force_post_mortem_stop"):
+ debugger.force_post_mortem_stop += 1
pydevd_tracing.SetTrace(None)
debugger.handle_post_mortem_stop(thread.additionalInfo, thread)
|
68fa1fd0df0f4048ac32e80698ebd872fb1f7fba
|
pytest_pycharm.py
|
pytest_pycharm.py
|
# -*- coding: utf-8 -*-
import threading
def pytest_exception_interact(node, call, report):
"""
Drop into PyCharm debugger, if available, on uncaught exceptions.
"""
try:
import pydevd
from pydevd import pydevd_tracing
except ImportError:
pass
else:
exctype, value, traceback = call.excinfo._excinfo
frames = []
while traceback:
frames.append(traceback.tb_frame)
traceback = traceback.tb_next
thread = threading.current_thread()
frames_by_id = dict([(id(frame), frame) for frame in frames])
frame = frames[-1]
exception = (exctype, value, traceback)
thread.additional_info.pydev_message = 'test fail'
debugger = pydevd.debugger
pydevd_tracing.SetTrace(None) # no tracing from here
debugger.handle_post_mortem_stop(thread, frame, frames_by_id, exception)
return report
|
# -*- coding: utf-8 -*-
import threading
def pytest_exception_interact(node, call, report):
"""
Drop into PyCharm debugger, if available, on uncaught exceptions.
"""
try:
import pydevd
from pydevd import pydevd_tracing
except ImportError:
pass
else:
exctype, value, traceback = call.excinfo._excinfo
frames = []
while traceback:
frames.append(traceback.tb_frame)
traceback = traceback.tb_next
thread = threading.current_thread()
frames_by_id = dict([(id(frame), frame) for frame in frames])
frame = frames[-1]
exception = (exctype, value, traceback)
thread.additional_info.pydev_message = 'test fail'
try:
debugger = pydevd.debugger
except AttributeError:
debugger = pydevd.get_global_debugger()
pydevd_tracing.SetTrace(None) # no tracing from here
debugger.handle_post_mortem_stop(thread, frame, frames_by_id, exception)
return report
|
Fix stopping on a breakpoint for PyCharm 2017.3
|
Fix stopping on a breakpoint for PyCharm 2017.3
|
Python
|
bsd-3-clause
|
jlubcke/pytest-pycharm
|
---
+++
@@ -22,7 +22,10 @@
frame = frames[-1]
exception = (exctype, value, traceback)
thread.additional_info.pydev_message = 'test fail'
- debugger = pydevd.debugger
+ try:
+ debugger = pydevd.debugger
+ except AttributeError:
+ debugger = pydevd.get_global_debugger()
pydevd_tracing.SetTrace(None) # no tracing from here
debugger.handle_post_mortem_stop(thread, frame, frames_by_id, exception)
|
82740c7956a2bae0baceedd658b9ad9352254ad0
|
nlppln/wfgenerator.py
|
nlppln/wfgenerator.py
|
from scriptcwl import WorkflowGenerator as WFGenerator
from .utils import CWL_PATH
class WorkflowGenerator(WFGenerator):
def __init__(self, working_dir=None, copy_steps=True):
WFGenerator.__init__(self, steps_dir=CWL_PATH, working_dir=working_dir,
copy_steps=copy_steps)
def save(self, fname, inline=True, relative=False, validate=True,
encoding='utf-8'):
"""Save workflow to file
For nlppln, the default is to save steps inline.
"""
super(WorkflowGenerator, self).save(fname,
inline=inline,
relative=relative,
validate=validate,
encoding=encoding)
|
from scriptcwl import WorkflowGenerator as WFGenerator
from .utils import CWL_PATH
class WorkflowGenerator(WFGenerator):
def __init__(self, working_dir=None):
WFGenerator.__init__(self, steps_dir=CWL_PATH, working_dir=working_dir)
def save(self, fname, validate=True, wd=True, inline=False, relative=False,
pack=False, encoding='utf-8'):
"""Save workflow to file
For nlppln, the default is to use a working directory (and save steps
using the ``wd`` option).
"""
super(WorkflowGenerator, self).save(fname,
validate=validate,
wd=wd,
inline=inline,
relative=relative,
pack=pack,
encoding=encoding)
|
Update to use newest (unreleased) scriptcwl options
|
Update to use newest (unreleased) scriptcwl options
|
Python
|
apache-2.0
|
WhatWorksWhenForWhom/nlppln,WhatWorksWhenForWhom/nlppln,WhatWorksWhenForWhom/nlppln
|
---
+++
@@ -4,18 +4,20 @@
class WorkflowGenerator(WFGenerator):
- def __init__(self, working_dir=None, copy_steps=True):
- WFGenerator.__init__(self, steps_dir=CWL_PATH, working_dir=working_dir,
- copy_steps=copy_steps)
+ def __init__(self, working_dir=None):
+ WFGenerator.__init__(self, steps_dir=CWL_PATH, working_dir=working_dir)
- def save(self, fname, inline=True, relative=False, validate=True,
- encoding='utf-8'):
+ def save(self, fname, validate=True, wd=True, inline=False, relative=False,
+ pack=False, encoding='utf-8'):
"""Save workflow to file
- For nlppln, the default is to save steps inline.
+ For nlppln, the default is to use a working directory (and save steps
+ using the ``wd`` option).
"""
super(WorkflowGenerator, self).save(fname,
+ validate=validate,
+ wd=wd,
inline=inline,
relative=relative,
- validate=validate,
+ pack=pack,
encoding=encoding)
|
b728470c61fbd742052e5befb4c27adbacef1a7e
|
pinax_theme_bootstrap/templatetags/pinax_theme_bootstrap_tags.py
|
pinax_theme_bootstrap/templatetags/pinax_theme_bootstrap_tags.py
|
from django import template
from django.contrib.messages.utils import get_level_tags
LEVEL_TAGS = get_level_tags()
register = template.Library()
@register.simple_tag()
def get_message_tags(message):
"""
Returns tags for a message
"""
level_name = LEVEL_TAGS[message.level]
if level_name == u"error":
level_name = u"danger"
level_tag = u"alert-{name}".format(name=level_name)
tags = [level_tag]
if message.extra_tags:
tags.append(message.extra_tags)
return u" ".join(tags)
|
from django import template
from django.contrib.messages.utils import get_level_tags
LEVEL_TAGS = get_level_tags()
register = template.Library()
@register.simple_tag()
def get_message_tags(message):
"""
Returns the message's level_tag prefixed with Bootstrap's "alert-" prefix
along with any tags included in message.extra_tags
Messages in Django >= 1.7 have a message.level_tag attr
"""
level_tag = LEVEL_TAGS[message.level]
if level_tag == u"error":
level_tag = u"danger"
alert_level_tag = u"alert-{tag}".format(tag=level_tag)
tags = [alert_level_tag]
if message.extra_tags:
tags.append(message.extra_tags)
return u" ".join(tags)
|
Use level_tag to be consistent with Django >= 1.7
|
Use level_tag to be consistent with Django >= 1.7
|
Python
|
mit
|
grahamu/pinax-theme-bootstrap,jacobwegner/pinax-theme-bootstrap,foraliving/foraliving,jacobwegner/pinax-theme-bootstrap,druss16/danslist,druss16/danslist,foraliving/foraliving,foraliving/foraliving,grahamu/pinax-theme-bootstrap,jacobwegner/pinax-theme-bootstrap,grahamu/pinax-theme-bootstrap,druss16/danslist
|
---
+++
@@ -10,15 +10,18 @@
@register.simple_tag()
def get_message_tags(message):
"""
- Returns tags for a message
+ Returns the message's level_tag prefixed with Bootstrap's "alert-" prefix
+ along with any tags included in message.extra_tags
+
+ Messages in Django >= 1.7 have a message.level_tag attr
"""
- level_name = LEVEL_TAGS[message.level]
- if level_name == u"error":
- level_name = u"danger"
+ level_tag = LEVEL_TAGS[message.level]
+ if level_tag == u"error":
+ level_tag = u"danger"
- level_tag = u"alert-{name}".format(name=level_name)
+ alert_level_tag = u"alert-{tag}".format(tag=level_tag)
- tags = [level_tag]
+ tags = [alert_level_tag]
if message.extra_tags:
tags.append(message.extra_tags)
|
fec7885d2632b887002f0071f4898faf52dd927c
|
chainerx/__init__.py
|
chainerx/__init__.py
|
import sys
if sys.version_info[0] < 3:
_available = False
else:
try:
from chainerx import _core
_available = True
except Exception:
_available = False
if _available:
from numpy import dtype, bool_, int8, int16, int32, int64, uint8, float32, float64 # NOQA
from chainerx._core import * # NOQA
from builtins import bool, int, float # NOQA
from chainerx.creation.from_data import asanyarray # NOQA
from chainerx.creation.from_data import fromfile # NOQA
from chainerx.creation.from_data import fromfunction # NOQA
from chainerx.creation.from_data import fromiter # NOQA
from chainerx.creation.from_data import fromstring # NOQA
from chainerx.creation.from_data import loadtxt # NOQA
_global_context = _core.Context()
_core.set_global_default_context(_global_context)
# Add workaround implementation for NumPy-compatible functions
from chainerx import _numpy_compat_workarounds
_numpy_compat_workarounds.populate()
else:
class ndarray(object):
pass # for type testing
def is_available():
return _available
|
import sys
if sys.version_info[0] < 3:
_available = False
else:
try:
from chainerx import _core
_available = True
except Exception:
_available = False
if _available:
from numpy import dtype, bool_, int8, int16, int32, int64, uint8, float32, float64 # NOQA
from chainerx._core import * # NOQA
from builtins import bool, int, float # NOQA
from chainerx.creation.from_data import asanyarray # NOQA
from chainerx.creation.from_data import fromfile # NOQA
from chainerx.creation.from_data import fromfunction # NOQA
from chainerx.creation.from_data import fromiter # NOQA
from chainerx.creation.from_data import fromstring # NOQA
from chainerx.creation.from_data import loadtxt # NOQA
_global_context = _core.Context()
_core.set_global_default_context(_global_context)
# Add workaround implementation for NumPy-compatible functions
from chainerx import _numpy_compat_workarounds
_numpy_compat_workarounds.populate()
else:
class ndarray(object):
"""Dummy class for type testing."""
def __init__(self, *args, **kwargs):
raise RuntimeError('chainerx is not available.')
def is_available():
return _available
|
Raise an error on dummy class init
|
Raise an error on dummy class init
|
Python
|
mit
|
okuta/chainer,jnishi/chainer,chainer/chainer,ktnyt/chainer,ktnyt/chainer,okuta/chainer,hvy/chainer,niboshi/chainer,ktnyt/chainer,chainer/chainer,ktnyt/chainer,hvy/chainer,wkentaro/chainer,jnishi/chainer,jnishi/chainer,okuta/chainer,wkentaro/chainer,keisuke-umezawa/chainer,niboshi/chainer,chainer/chainer,keisuke-umezawa/chainer,tkerola/chainer,hvy/chainer,jnishi/chainer,pfnet/chainer,wkentaro/chainer,keisuke-umezawa/chainer,niboshi/chainer,keisuke-umezawa/chainer,hvy/chainer,wkentaro/chainer,chainer/chainer,niboshi/chainer,okuta/chainer
|
---
+++
@@ -34,7 +34,11 @@
_numpy_compat_workarounds.populate()
else:
class ndarray(object):
- pass # for type testing
+
+ """Dummy class for type testing."""
+
+ def __init__(self, *args, **kwargs):
+ raise RuntimeError('chainerx is not available.')
def is_available():
|
f16c8f696a282da6c04de6b7530f1d0316eda88b
|
providers/edu/harvarddataverse/normalizer.py
|
providers/edu/harvarddataverse/normalizer.py
|
import arrow
import dateparser
from share.normalize import *
class Person(Parser):
given_name = ParseName(ctx).first
family_name = ParseName(ctx).last
additional_name = ParseName(ctx).middle
suffix = ParseName(ctx).suffix
class Contributor(Parser):
person = Delegate(Person, ctx)
cited_name = ctx
order_cited = ctx('index')
class Link(Parser):
url = ctx
type = RunPython('get_link_type', ctx)
def get_link_type(self, link):
if 'dx.doi.org' in link:
return 'doi'
elif 'dataverse.harvard.edu' in link:
return 'provider'
return 'misc'
class ThroughLinks(Parser):
link = Delegate(Link, ctx)
class CreativeWork(Parser):
title = ctx.name
description = ctx.description
contributors = Map(Delegate(Contributor), ctx.authors)
date_published = ParseDate(ctx.published_at)
links = Concat(
Delegate(ThroughLinks, ctx.url),
Delegate(ThroughLinks, ctx.image_url),
)
|
import arrow
import dateparser
from share.normalize import *
class Person(Parser):
given_name = ParseName(ctx).first
family_name = ParseName(ctx).last
additional_name = ParseName(ctx).middle
suffix = ParseName(ctx).suffix
class Contributor(Parser):
person = Delegate(Person, ctx)
cited_name = ctx
order_cited = ctx('index')
class Link(Parser):
url = ctx
type = RunPython('get_link_type', ctx)
def get_link_type(self, link):
if 'dx.doi.org' in link:
return 'doi'
elif 'dataverse.harvard.edu' in link:
return 'provider'
return 'misc'
class ThroughLinks(Parser):
link = Delegate(Link, ctx)
class CreativeWork(Parser):
title = ctx.name
description = Try(ctx.description)
contributors = Map(Delegate(Contributor), Try(ctx.authors))
date_published = ParseDate(ctx.published_at)
links = Concat(
Delegate(ThroughLinks, ctx.url),
Delegate(ThroughLinks, ctx.image_url),
)
|
Handle missing fields in dataverse
|
Handle missing fields in dataverse
|
Python
|
apache-2.0
|
CenterForOpenScience/SHARE,laurenbarker/SHARE,aaxelb/SHARE,aaxelb/SHARE,laurenbarker/SHARE,zamattiac/SHARE,zamattiac/SHARE,CenterForOpenScience/SHARE,laurenbarker/SHARE,CenterForOpenScience/SHARE,zamattiac/SHARE,aaxelb/SHARE
|
---
+++
@@ -36,8 +36,8 @@
class CreativeWork(Parser):
title = ctx.name
- description = ctx.description
- contributors = Map(Delegate(Contributor), ctx.authors)
+ description = Try(ctx.description)
+ contributors = Map(Delegate(Contributor), Try(ctx.authors))
date_published = ParseDate(ctx.published_at)
links = Concat(
Delegate(ThroughLinks, ctx.url),
|
3327c204f34a725a2d070beb24a7a5a66d414930
|
migrations/versions/538eeb160af6_.py
|
migrations/versions/538eeb160af6_.py
|
"""empty message
Revision ID: 538eeb160af6
Revises: 1727fb4309d8
Create Date: 2015-09-17 04:22:21.262285
"""
# revision identifiers, used by Alembic.
revision = '538eeb160af6'
down_revision = '1727fb4309d8'
from alembic import op
import sqlalchemy as sa
def upgrade():
### commands auto generated by Alembic - please adjust! ###
op.add_column('user', sa.Column('role', sa.String(length=30), nullable=True))
### end Alembic commands ###
def downgrade():
### commands auto generated by Alembic - please adjust! ###
op.drop_column('user', 'role')
### end Alembic commands ###
|
"""empty message
Revision ID: 538eeb160af6
Revises: 1727fb4309d8
Create Date: 2015-09-17 04:22:21.262285
"""
# revision identifiers, used by Alembic.
revision = '538eeb160af6'
down_revision = '6b9d673d8e30'
from alembic import op
import sqlalchemy as sa
def upgrade():
### commands auto generated by Alembic - please adjust! ###
op.add_column('user', sa.Column('role', sa.String(length=30), nullable=True))
### end Alembic commands ###
def downgrade():
### commands auto generated by Alembic - please adjust! ###
op.drop_column('user', 'role')
### end Alembic commands ###
|
Update alembic order for merging
|
Update alembic order for merging
|
Python
|
apache-2.0
|
bunjiboys/security_monkey,stackArmor/security_monkey,markofu/security_monkey,bunjiboys/security_monkey,bunjiboys/security_monkey,markofu/security_monkey,markofu/security_monkey,Netflix/security_monkey,stackArmor/security_monkey,Netflix/security_monkey,Netflix/security_monkey,stackArmor/security_monkey,Netflix/security_monkey,bunjiboys/security_monkey,Netflix/security_monkey,bunjiboys/security_monkey,stackArmor/security_monkey,stackArmor/security_monkey,markofu/security_monkey,markofu/security_monkey
|
---
+++
@@ -8,7 +8,7 @@
# revision identifiers, used by Alembic.
revision = '538eeb160af6'
-down_revision = '1727fb4309d8'
+down_revision = '6b9d673d8e30'
from alembic import op
import sqlalchemy as sa
|
aae0a19e6b03c9b42272f403cc1e5e7c943ea8ab
|
setup.py
|
setup.py
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from __future__ import unicode_literals
from setuptools import setup, find_packages
import os
def read_file(filename):
"""Read a file into a string"""
path = os.path.abspath(os.path.dirname(__file__))
filepath = os.path.join(path, filename)
try:
return open(filepath).read()
except IOError:
return ''
setup(
name='python-hcalendar',
version=__import__('hcalendar').__version__,
author='Marc Hoersken',
author_email='info@marc-hoersken.de',
packages=find_packages(exclude=['unittests']),
include_package_data=True,
url='https://github.com/mback2k/python-hcalendar',
license='MIT',
description=' '.join(__import__('hcalendar').__doc__.splitlines()).strip(),
install_requires=['isodate>=0.5.0', 'beautifulsoup4>=4.3.2'],
classifiers=[
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python',
'Topic :: Text Processing :: Markup :: HTML',
'Topic :: Software Development :: Libraries :: Python Modules',
'Development Status :: 4 - Beta',
'Operating System :: OS Independent',
],
long_description=read_file('README.md'),
)
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
from __future__ import absolute_import
from __future__ import division
from __future__ import print_function
from __future__ import unicode_literals
from setuptools import setup, find_packages
import os
def read_file(filename):
"""Read a file into a string"""
path = os.path.abspath(os.path.dirname(__file__))
filepath = os.path.join(path, filename)
try:
return open(filepath).read()
except IOError:
return ''
setup(
name='python-hcalendar',
version=__import__('hcalendar').__version__,
author='Marc Hoersken',
author_email='info@marc-hoersken.de',
packages=find_packages(exclude=['unittests']),
include_package_data=True,
url='https://github.com/mback2k/python-hcalendar',
license='MIT',
description=' '.join(__import__('hcalendar').__doc__.splitlines()).strip(),
install_requires=['isodate>=0.5.0', 'beautifulsoup4>=4.3.2'],
classifiers=[
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python',
'Topic :: Text Processing :: Markup :: HTML',
'Topic :: Software Development :: Libraries :: Python Modules',
'Development Status :: 4 - Beta',
'Operating System :: OS Independent',
],
long_description=read_file('README.md'),
long_description_content_type='text/markdown',
)
|
Set long description content type to markdown
|
Set long description content type to markdown
|
Python
|
mit
|
mback2k/python-hcalendar
|
---
+++
@@ -38,4 +38,5 @@
'Operating System :: OS Independent',
],
long_description=read_file('README.md'),
+ long_description_content_type='text/markdown',
)
|
fea8e206f3ef2755710a585b6aa76ce5c148a817
|
setup.py
|
setup.py
|
#! /usr/bin/env python
from setuptools import setup, find_packages
setup(
name='armet',
version='0.3.0-pre',
description='Clean and modern framework for creating RESTful APIs.',
author='Concordus Applications',
author_email='support@concordusapps.com',
url='http://github.com/armet/python-armet',
package_dir={'armet': 'src/armet'},
packages=find_packages('src'),
install_requires=(
'six', # Python 2 and 3 normalization layer
'python-mimeparse' # For parsing accept and content-type headers
),
extras_require={
'test': (
'nose',
'yanc'
)
}
)
|
#! /usr/bin/env python
from setuptools import setup, find_packages
setup(
name='armet',
version='0.3.0-pre',
description='Clean and modern framework for creating RESTful APIs.',
author='Concordus Applications',
author_email='support@concordusapps.com',
url='http://github.com/armet/python-armet',
package_dir={'armet': 'src/armet'},
packages=find_packages('src'),
install_requires=(
'six', # Python 2 and 3 normalization layer
'python-mimeparse' # For parsing accept and content-type headers
),
extras_require={
'test': (
'nose',
'yanc',
'httplib2'
)
}
)
|
Add httplib2 as a test dependency.
|
Add httplib2 as a test dependency.
|
Python
|
mit
|
armet/python-armet
|
---
+++
@@ -12,13 +12,14 @@
package_dir={'armet': 'src/armet'},
packages=find_packages('src'),
install_requires=(
- 'six', # Python 2 and 3 normalization layer
- 'python-mimeparse' # For parsing accept and content-type headers
+ 'six', # Python 2 and 3 normalization layer
+ 'python-mimeparse' # For parsing accept and content-type headers
),
extras_require={
'test': (
'nose',
- 'yanc'
+ 'yanc',
+ 'httplib2'
)
}
)
|
20117a8c35ef47842a98bff92fc98773508ebbe0
|
setup.py
|
setup.py
|
"""
Usage instructions:
- If you are installing: `python setup.py install`
- If you are developing: `python setup.py sdist bdist --format=zip bdist_wheel --universal`
"""
try:
import pypandoc
long_description = pypandoc.convert('README.md', 'rst')
except ImportError:
try:
long_description = open('README.md').read()
except:
try:
long_description = open('README.rst').read()
except:
print('Failed to convert README to rst.')
long_description = ''
from setuptools import setup
setup(
name='keyboard',
version='0.7.1',
author='BoppreH',
author_email='boppreh@gmail.com',
packages=['keyboard'],
url='https://github.com/boppreh/keyboard',
license='MIT',
description='Hook and simulate keyboard events on Windows and Linux',
keywords = 'keyboard hook simulate hotkey',
long_description=long_description,
classifiers=[
'Development Status :: 3 - Alpha',
'License :: OSI Approved :: MIT License',
'Natural Language :: English',
'Operating System :: Microsoft :: Windows',
'Operating System :: Unix',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 3',
'Topic :: Software Development :: Libraries :: Python Modules',
'Topic :: Utilities',
],
)
|
"""
Usage instructions:
- If you are installing: `python setup.py install`
- If you are developing: `python setup.py sdist bdist --format=zip bdist_wheel --universal`
"""
try:
long_description = open('README.rst').read()
except ImportError:
import pypandoc
long_description = pypandoc.convert('README.md', 'rst')
from setuptools import setup
setup(
name='keyboard',
version='0.7.1',
author='BoppreH',
author_email='boppreh@gmail.com',
packages=['keyboard'],
url='https://github.com/boppreh/keyboard',
license='MIT',
description='Hook and simulate keyboard events on Windows and Linux',
keywords = 'keyboard hook simulate hotkey',
long_description=long_description,
classifiers=[
'Development Status :: 3 - Alpha',
'License :: OSI Approved :: MIT License',
'Natural Language :: English',
'Operating System :: Microsoft :: Windows',
'Operating System :: Unix',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 3',
'Topic :: Software Development :: Libraries :: Python Modules',
'Topic :: Utilities',
],
)
|
Make README publishing more conservative
|
Make README publishing more conservative
|
Python
|
mit
|
boppreh/keyboard,glitchassassin/keyboard
|
---
+++
@@ -6,17 +6,10 @@
"""
try:
+ long_description = open('README.rst').read()
+except ImportError:
import pypandoc
long_description = pypandoc.convert('README.md', 'rst')
-except ImportError:
- try:
- long_description = open('README.md').read()
- except:
- try:
- long_description = open('README.rst').read()
- except:
- print('Failed to convert README to rst.')
- long_description = ''
from setuptools import setup
|
d2d822a9fb60bbc8ded7f9e3c70d91cf25f794b2
|
src/volunteers/models.py
|
src/volunteers/models.py
|
from django.db import models
from django.utils.translation import ugettext_lazy as _
from django.core.validators import MinValueValidator
class Volunteer(models.Model):
first_name = models.CharField(_('First name'), max_length=100)
last_name = models.CharField(_('Last name'), max_length=100)
age = models.PositiveIntegerField(_('Age'))
phone = models.CharField(_('Phone'), max_length=100)
email = models.EmailField(_('E-mail'), unique=True)
is_group = models.BooleanField(_('Is group representative'), default=False)
group_name = models.CharField(_('Group/organization name'), max_length=100,
blank=True)
participant_count = models.PositiveIntegerField(_('Participant count'),
default=1, validators=[MinValueValidator(1)])
class Meta:
verbose_name = _('Volunteer')
verbose_name_plural = _('Volunteers')
@property
def name(self):
template = u'{first_name} {last_name}'
if self.is_group:
template += u' (grupp, {participant_count} osalejat)'
return template.format(**self.__dict__)
def __unicode__(self):
return self.name
|
from django.db import models
from django.utils.translation import ugettext_lazy as _
from django.core.validators import MinValueValidator
class Volunteer(models.Model):
first_name = models.CharField(_('First name'), max_length=100)
last_name = models.CharField(_('Last name'), max_length=100)
age = models.PositiveIntegerField(_('Age'))
phone = models.CharField(_('Phone'), max_length=100)
email = models.EmailField(_('E-mail'), unique=True)
is_group = models.BooleanField(_('Is group representative'), default=False)
group_name = models.CharField(_('Group/organization name'), max_length=100,
blank=True)
participant_count = models.PositiveIntegerField(_('Participant count'),
default=1, validators=[MinValueValidator(1)])
class Meta:
verbose_name = _('Volunteer')
verbose_name_plural = _('Volunteers')
@property
def name(self):
template = u'{first_name} {last_name}'
if self.is_group:
template += u' ({group_name} grupp, {participant_count} osalejat)'
return template.format(**self.__dict__)
def __unicode__(self):
return self.name
|
Add group name to volunteer string representation
|
Add group name to volunteer string representation
|
Python
|
mit
|
mrts/foodbank-campaign,mrts/foodbank-campaign,mrts/foodbank-campaign,mrts/foodbank-campaign
|
---
+++
@@ -23,7 +23,7 @@
def name(self):
template = u'{first_name} {last_name}'
if self.is_group:
- template += u' (grupp, {participant_count} osalejat)'
+ template += u' ({group_name} grupp, {participant_count} osalejat)'
return template.format(**self.__dict__)
def __unicode__(self):
|
e57e13bde61a233b18504ab1617c6ecabad20fc3
|
setup.py
|
setup.py
|
#!/usr/bin/env python
from setuptools import setup
from setuptools import find_packages
import re
def find_version():
return re.search(r"^__version__ = '(.*)'$",
open('cantools/version.py', 'r').read(),
re.MULTILINE).group(1)
setup(name='cantools',
version=find_version(),
description='CAN BUS tools.',
long_description=open('README.rst', 'r').read(),
author='Erik Moqvist',
author_email='erik.moqvist@gmail.com',
license='MIT',
classifiers=[
'License :: OSI Approved :: MIT License',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3 :: Only',
'Programming Language :: Python :: 3.6',
'Programming Language :: Python :: 3.7',
'Programming Language :: Python :: 3.8',
'Programming Language :: Python :: 3.9',
],
keywords=['can', 'can bus', 'dbc', 'kcd', 'automotive'],
url='https://github.com/eerimoq/cantools',
packages=find_packages(exclude=['tests']),
python_requires='>=3.6',
install_requires=[
'bitstruct>=6.0.0',
'python-can>=2.2.0',
'textparser>=0.21.1',
'diskcache',
'argparse_addons',
],
test_suite="tests",
entry_points = {
'console_scripts': ['cantools=cantools.__init__:_main']
})
|
#!/usr/bin/env python
from setuptools import setup
from setuptools import find_packages
import re
def find_version():
return re.search(r"^__version__ = '(.*)'$",
open('cantools/version.py', 'r').read(),
re.MULTILINE).group(1)
setup(name='cantools',
version=find_version(),
description='CAN BUS tools.',
long_description=open('README.rst', 'r').read(),
author='Erik Moqvist',
author_email='erik.moqvist@gmail.com',
license='MIT',
classifiers=[
'License :: OSI Approved :: MIT License',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3 :: Only',
],
keywords=['can', 'can bus', 'dbc', 'kcd', 'automotive'],
url='https://github.com/eerimoq/cantools',
packages=find_packages(exclude=['tests']),
python_requires='>=3.6',
install_requires=[
'bitstruct>=6.0.0',
'python-can>=2.2.0',
'textparser>=0.21.1',
'diskcache',
'argparse_addons',
],
test_suite="tests",
entry_points = {
'console_scripts': ['cantools=cantools.__init__:_main']
})
|
Remove per patch version classifiers
|
Remove per patch version classifiers
|
Python
|
mit
|
eerimoq/cantools,cantools/cantools
|
---
+++
@@ -22,10 +22,6 @@
'License :: OSI Approved :: MIT License',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3 :: Only',
- 'Programming Language :: Python :: 3.6',
- 'Programming Language :: Python :: 3.7',
- 'Programming Language :: Python :: 3.8',
- 'Programming Language :: Python :: 3.9',
],
keywords=['can', 'can bus', 'dbc', 'kcd', 'automotive'],
url='https://github.com/eerimoq/cantools',
|
621968ca6bd2fb986d83ccc8d4c0d5a83ce3aadb
|
setup.py
|
setup.py
|
from setuptools import setup
import generate
generate.run()
setup(
name='nnpy',
version='0.1',
url='https://github.com/nanomsg/nnpy',
license='MIT',
author='Dirkjan Ochtman',
author_email='dirkjan@ochtman.nl',
description='cffi-based Python bindings for nanomsg',
long_description=open('README.rst').read(),
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
],
packages=['nnpy'],
package_data={'nnpy': ['*.h']},
install_requires=['cffi'],
)
|
from setuptools import setup
import generate
generate.run()
setup(
name='nnpy',
version='0.1',
url='https://github.com/nanomsg/nnpy',
license='MIT',
author='Dirkjan Ochtman',
author_email='dirkjan@ochtman.nl',
description='cffi-based Python bindings for nanomsg',
long_description=open('README.rst').read(),
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: Implementation :: CPython',
],
packages=['nnpy'],
package_data={'nnpy': ['*.h']},
install_requires=['cffi'],
)
|
Add tested python version in classifiers
|
Add tested python version in classifiers
|
Python
|
mit
|
nanomsg/nnpy
|
---
+++
@@ -19,6 +19,11 @@
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
+ 'Programming Language :: Python :: 2.7',
+ 'Programming Language :: Python :: 3',
+ 'Programming Language :: Python :: 3.3',
+ 'Programming Language :: Python :: 3.4',
+ 'Programming Language :: Python :: Implementation :: CPython',
],
packages=['nnpy'],
package_data={'nnpy': ['*.h']},
|
80f59dd5b99605d212243641d01fcd635ddd0d93
|
setup.py
|
setup.py
|
#! /usr/bin/env python
from setuptools import find_packages, setup
setup(name='oemof.db',
# Unfortunately we can't use a `__version__` attribute on `oemof.db` as
# we can't import that module here. It depends on packages which might
# not be available prior to installation.
version='0.0.6dev',
description='The oemof database extension',
namespace_package = ['oemof'],
packages=find_packages(),
package_dir={'oemof': 'oemof'},
install_requires=['sqlalchemy >= 1.0',
'keyring >= 4.0',
'shapely',
'psycopg2',
'keyrings.alt',
'pandas >=0.19.1, <=0.19.1'])
|
#! /usr/bin/env python
from setuptools import find_packages, setup
setup(name='oemof.db',
# Unfortunately we can't use a `__version__` attribute on `oemof.db` as
# we can't import that module here. It depends on packages which might
# not be available prior to installation.
version='0.0.6dev',
description='The oemof database extension',
namespace_package = ['oemof'],
packages=find_packages(),
package_dir={'oemof': 'oemof'},
install_requires=['sqlalchemy >= 1.0',
'keyring >= 4.0',
'shapely',
'psycopg2',
'keyrings.alt',
'pandas >=0.19.1'])
|
Drop upper version bound on pandas
|
Drop upper version bound on pandas
There's really no reason for having it.
|
Python
|
mit
|
oemof/oemof.db
|
---
+++
@@ -16,4 +16,4 @@
'shapely',
'psycopg2',
'keyrings.alt',
- 'pandas >=0.19.1, <=0.19.1'])
+ 'pandas >=0.19.1'])
|
e1111ad6e8802b3c90df55e05eb695d6db9005e4
|
import_script/create_users.py
|
import_script/create_users.py
|
#!/usr/bin/python
import django.contrib.auth.models as auth_models
import django.contrib.contenttypes as contenttypes
def main():
# Read only user:
# auth_models.User.objects.create_user('cube', 'toolkit_admin_readonly@localhost', '***REMOVED***')
# Read/write user:
user_rw = auth_models.User.objects.create_user('admin', 'toolkit_admin@localhost', '***REMOVED***')
# Create dummy ContentType:
ct = contenttypes.models.ContentType.objects.get_or_create(
model='',
app_label='toolkit'
)[0]
# Create 'write' permission:
write_permission = auth_models.Permission.objects.get_or_create(
name='Write access to all toolkit content',
content_type=ct,
codename='write'
)[0]
# Give "admin" user the write permission:
user_rw.user_permissions.add(write_permission)
if __name__ == "__main__":
main()
|
#!/usr/bin/python
import django.contrib.auth.models as auth_models
import django.contrib.contenttypes as contenttypes
def get_password():
print "*" * 80
password = raw_input("Please enter string to use as admin password: ")
check_password = None
while check_password != password:
print
check_password = raw_input("Please re-enter for confirmation: ")
return password
def main():
# Read only user:
# auth_models.User.objects.create_user('cube', 'toolkit_admin_readonly@localhost', '********')
# Read/write user:
cube_password = get_password()
user_rw = auth_models.User.objects.create_user('admin', 'toolkit_admin@localhost', cube_password)
# Create dummy ContentType:
ct = contenttypes.models.ContentType.objects.get_or_create(
model='',
app_label='toolkit'
)[0]
# Create 'write' permission:
write_permission = auth_models.Permission.objects.get_or_create(
name='Write access to all toolkit content',
content_type=ct,
codename='write'
)[0]
# Give "admin" user the write permission:
user_rw.user_permissions.add(write_permission)
if __name__ == "__main__":
main()
|
Remove cube credentials from import script
|
Remove cube credentials from import script
|
Python
|
agpl-3.0
|
BenMotz/cubetoolkit,BenMotz/cubetoolkit,BenMotz/cubetoolkit,BenMotz/cubetoolkit
|
---
+++
@@ -3,12 +3,24 @@
import django.contrib.auth.models as auth_models
import django.contrib.contenttypes as contenttypes
+def get_password():
+ print "*" * 80
+ password = raw_input("Please enter string to use as admin password: ")
+ check_password = None
+
+ while check_password != password:
+ print
+ check_password = raw_input("Please re-enter for confirmation: ")
+
+ return password
+
def main():
# Read only user:
- # auth_models.User.objects.create_user('cube', 'toolkit_admin_readonly@localhost', '***REMOVED***')
+ # auth_models.User.objects.create_user('cube', 'toolkit_admin_readonly@localhost', '********')
# Read/write user:
- user_rw = auth_models.User.objects.create_user('admin', 'toolkit_admin@localhost', '***REMOVED***')
+ cube_password = get_password()
+ user_rw = auth_models.User.objects.create_user('admin', 'toolkit_admin@localhost', cube_password)
# Create dummy ContentType:
ct = contenttypes.models.ContentType.objects.get_or_create(
model='',
|
e6457c384eaa13eff82217ef4eb15f580efd8121
|
setup.py
|
setup.py
|
import re
from setuptools import setup
init_py = open('wikipediabase/__init__.py').read()
metadata = dict(re.findall("__([a-z]+)__ = '([^']+)'", init_py))
metadata['doc'] = re.findall('"""(.+)"""', init_py)[0]
setup(
name='wikipediabase',
version=metadata['version'],
description=metadata['doc'],
author=metadata['author'],
author_email=metadata['email'],
url=metadata['url'],
packages=[
'wikipediabase',
'wikipediabase.resolvers',
'wikipediabase.adhoc',
'tests',
],
include_package_data=True,
install_requires=[
'edn_format',
'docopt',
'flake8 < 3.0.0',
'unittest2 < 1.0.0',
'overlay-parse',
'lxml',
'sqlitedict',
'requests',
'beautifulsoup4',
'redis',
'redis',
'hiredis',
],
dependency_links=[
'git+https://github.com/fakedrake/overlay_parse#egg=overlay-parse',
],
tests_require=[
'nose>=1.0',
'sqlitedict',
],
entry_points={
'console_scripts': [
'wikipediabase = wikipediabase.cli:main',
],
},
test_suite='nose.collector',
license=open('LICENSE').read(),
)
|
import re
from setuptools import setup
init_py = open('wikipediabase/__init__.py').read()
metadata = dict(re.findall("__([a-z]+)__ = '([^']+)'", init_py))
metadata['doc'] = re.findall('"""(.+)"""', init_py)[0]
setup(
name='wikipediabase',
version=metadata['version'],
description=metadata['doc'],
author=metadata['author'],
author_email=metadata['email'],
url=metadata['url'],
packages=[
'wikipediabase',
'wikipediabase.resolvers',
'wikipediabase.adhoc',
'tests',
],
include_package_data=True,
install_requires=[
'edn_format',
'docopt',
'flake8 < 3.0.0',
'unittest2 < 1.0.0',
'overlay-parse',
'lxml',
'sqlitedict',
'requests',
'beautifulsoup4',
'redis',
'redis',
'hiredis',
'unidecode',
],
dependency_links=[
'git+https://github.com/fakedrake/overlay_parse#egg=overlay-parse',
],
tests_require=[
'nose>=1.0',
'sqlitedict',
],
entry_points={
'console_scripts': [
'wikipediabase = wikipediabase.cli:main',
],
},
test_suite='nose.collector',
license=open('LICENSE').read(),
)
|
Add unidecode as a dependency
|
Add unidecode as a dependency
|
Python
|
apache-2.0
|
fakedrake/WikipediaBase
|
---
+++
@@ -32,6 +32,7 @@
'redis',
'redis',
'hiredis',
+ 'unidecode',
],
dependency_links=[
'git+https://github.com/fakedrake/overlay_parse#egg=overlay-parse',
|
1a37d53df02942cb27794303e68fe041f73f5135
|
setup.py
|
setup.py
|
import os
from setuptools import setup
from setuptools import find_packages
setup(
name='MobOff',
version='0.1',
py_modules=['moboff'],
packages=find_packages(),
description = 'Download youtube music and send to devices',
author = 'Parth Verma',
author_email = 'vermaparth97@gmail.com',
url = "https://github.com/parth-vader/MobOff",
license = "MIT License",
install_requires=[
'pushbullet.py',
'youtube_dl',
'Click',
],
entry_points='''
[console_scripts]
moboff=mainscript:cli
''',
)
os.system('chmod a+x mainscript.py')
os.system('export PATH=mainscript.py:$PATH')
|
import os
from setuptools import setup
from setuptools import find_packages
setup(
name='MobOff',
version='0.2',
py_modules=['moboff'],
packages=find_packages(),
description = 'Download youtube music and send to devices',
author = 'Parth Verma',
author_email = 'vermaparth97@gmail.com',
url = "https://github.com/parth-vader/MobOff",
download_url = 'https://codeload.github.com/Parth-Vader/MobOff/tar.gz/0.2'
license = "MIT License",
install_requires=[
'pushbullet.py',
'youtube_dl',
'Click',
],
entry_points='''
[console_scripts]
moboff=mainscript:cli
''',
)
os.system('chmod a+x mainscript.py')
os.system('export PATH=mainscript.py:$PATH')
|
Tag updated and download link added
|
Tag updated and download link added
|
Python
|
mit
|
Parth-Vader/MobOff
|
---
+++
@@ -5,13 +5,14 @@
setup(
name='MobOff',
- version='0.1',
+ version='0.2',
py_modules=['moboff'],
packages=find_packages(),
description = 'Download youtube music and send to devices',
author = 'Parth Verma',
author_email = 'vermaparth97@gmail.com',
url = "https://github.com/parth-vader/MobOff",
+ download_url = 'https://codeload.github.com/Parth-Vader/MobOff/tar.gz/0.2'
license = "MIT License",
install_requires=[
'pushbullet.py',
|
c89abd6a285225313c91ba03c0fd8ab2cfed399d
|
setup.py
|
setup.py
|
#!/usr/bin/env python
import os
import urllib
import zipfile
script_path = os.path.dirname(os.path.realpath(__file__))
packer_archive_path = script_path + "/packer.zip"
bin_path = script_path + "/bin"
if not os.path.isfile(bin_path + "/packer"):
if not os.path.exists(bin_path):
os.makedirs(bin_path)
try:
urllib.urlretrieve("https://dl.bintray.com/mitchellh/packer/packer_0.8.6_linux_amd64.zip", packer_archive_path)
with zipfile.ZipFile(packer_archive_path, "r") as packer_archive:
packer_archive.extractall(path=bin_path)
finally:
os.remove(packer_archive_path)
for root, subdirectories, files in os.walk(bin_path):
for f in files:
os.chmod("%s/%s" % (root, f), 0755)
|
#!/usr/bin/env python
import os
import urllib
import zipfile
script_path = os.path.dirname(os.path.realpath(__file__))
packer_archive_path = script_path + "/packer.zip"
bin_path = script_path + "/bin"
if not os.path.isfile(bin_path + "/packer"):
if not os.path.exists(bin_path):
os.makedirs(bin_path)
try:
urllib.urlretrieve("https://dl.bintray.com/mitchellh/packer/packer_0.8.6_linux_amd64.zip", packer_archive_path)
with zipfile.ZipFile(packer_archive_path, "r") as packer_archive:
packer_archive.extractall(path=bin_path)
finally:
os.remove(packer_archive_path)
for root, subdirectories, files in os.walk(bin_path):
for f in files:
os.chmod(root + "/" + f, 755)
|
Fix false positive octal syntax warning
|
Fix false positive octal syntax warning
|
Python
|
unlicense
|
dharmab/centos-vagrant
|
---
+++
@@ -20,5 +20,5 @@
for root, subdirectories, files in os.walk(bin_path):
for f in files:
- os.chmod("%s/%s" % (root, f), 0755)
+ os.chmod(root + "/" + f, 755)
|
110f62562a4802f99c93a53ea09e1e3deb2b3714
|
setup.py
|
setup.py
|
from setuptools import setup, find_packages
import os
version = '0.3.10'
here = os.path.abspath(os.path.dirname(__file__))
try:
README = open(os.path.join(here, 'README.txt')).read()
CHANGES = open(os.path.join(here, 'docs/HISTORY.txt')).read()
except IOError:
README = CHANGES = ''
setup(name='tgext.admin',
version=version,
description="Admin Controller add-on for basic TG identity model.",
long_description=README + "\n" +
CHANGES,
# Get more strings from http://www.python.org/pypi?%3Aaction=list_classifiers
classifiers=[
"Programming Language :: Python",
"Topic :: Software Development :: Libraries :: Python Modules",
],
keywords='TG2, TG, sprox, Rest, internet, adminn',
author='Christopher Perkins',
author_email='chris@percious.com',
url='tgtools.googlecode.com',
license='MIT',
packages=find_packages(exclude=['ez_setup']),
namespace_packages=['tgext'],
include_package_data=True,
zip_safe=True,
install_requires=[
'setuptools',
'sprox>=0.6.4',
'tgext.crud>=0.3.7',
# -*- Extra requirements: -*-
],
entry_points="""
# -*- Entry points: -*-
""",
)
|
from setuptools import setup, find_packages
import os
version = '0.3.11'
here = os.path.abspath(os.path.dirname(__file__))
try:
README = open(os.path.join(here, 'README.txt')).read()
CHANGES = open(os.path.join(here, 'docs/HISTORY.txt')).read()
except IOError:
README = CHANGES = ''
setup(name='tgext.admin',
version=version,
description="Admin Controller add-on for basic TG identity model.",
long_description=README + "\n" +
CHANGES,
# Get more strings from http://www.python.org/pypi?%3Aaction=list_classifiers
classifiers=[
"Programming Language :: Python",
"Topic :: Software Development :: Libraries :: Python Modules",
],
keywords='TG2, TG, sprox, Rest, internet, adminn',
author='Christopher Perkins',
author_email='chris@percious.com',
url='tgtools.googlecode.com',
license='MIT',
packages=find_packages(exclude=['ez_setup']),
namespace_packages=['tgext'],
include_package_data=True,
zip_safe=True,
install_requires=[
'setuptools',
'sprox>=0.6.4',
'tgext.crud>=0.3.7',
# -*- Extra requirements: -*-
],
entry_points="""
# -*- Entry points: -*-
""",
)
|
Increase to version 0.3.11 due to TG-dev requiring it for ming support
|
Increase to version 0.3.11 due to TG-dev requiring it for ming support
|
Python
|
mit
|
pedersen/tgtools.tgext-admin,pedersen/tgtools.tgext-admin
|
---
+++
@@ -1,7 +1,7 @@
from setuptools import setup, find_packages
import os
-version = '0.3.10'
+version = '0.3.11'
here = os.path.abspath(os.path.dirname(__file__))
try:
README = open(os.path.join(here, 'README.txt')).read()
|
23f2306617a4e4bceecd20190c328b2b3418abc4
|
setup.py
|
setup.py
|
#! /usr/bin/python
"""Setuptools-based setup script for datreant.
For a basic installation just type the command::
python setup.py install
"""
from setuptools import setup
setup(name='datreant',
version='0.5.1',
author='David Dotson',
author_email='dotsdl@gmail.com',
packages=['datreant', 'datreant.tests'],
scripts=[],
license='BSD',
long_description=open('README.rst').read(),
install_requires=['pandas', 'tables', 'h5py', 'scandir']
)
|
#! /usr/bin/python
"""Setuptools-based setup script for datreant.
For a basic installation just type the command::
python setup.py install
"""
from setuptools import setup
setup(name='datreant',
version='0.5.1',
author='David Dotson',
author_email='dotsdl@gmail.com',
packages=['datreant', 'datreant.tests'],
scripts=[],
license='BSD',
long_description=open('README.rst').read(),
install_requires=[
'numpy',
'pandas',
'tables',
'h5py',
'scandir',
'PyYAML'
]
)
|
Add PyYAML & numpy dependency
|
Add PyYAML & numpy dependency
I'm adding numpy too because we import it directly.
|
Python
|
bsd-3-clause
|
datreant/datreant,dotsdl/datreant,datreant/datreant.core,datreant/datreant.core,datreant/datreant,datreant/datreant.data
|
---
+++
@@ -11,11 +11,18 @@
setup(name='datreant',
version='0.5.1',
- author='David Dotson',
+ author='David Dotson',
author_email='dotsdl@gmail.com',
packages=['datreant', 'datreant.tests'],
scripts=[],
license='BSD',
long_description=open('README.rst').read(),
- install_requires=['pandas', 'tables', 'h5py', 'scandir']
- )
+ install_requires=[
+ 'numpy',
+ 'pandas',
+ 'tables',
+ 'h5py',
+ 'scandir',
+ 'PyYAML'
+ ]
+ )
|
95e15792b50f28c5c40dcad17de2fd4535345bc8
|
setup.py
|
setup.py
|
from setuptools import setup
from os import path
here = path.abspath(path.dirname(__file__))
with open(path.join(here, 'README.rst')) as readme_file:
long_description = readme_file.read()
setup(
name='pyfds',
description='Modular field simulation tool using finite differences.',
long_description=long_description,
url='http://emt.uni-paderborn.de',
author='Leander Claes',
author_email='claes@emt.uni-paderborn.de',
license='Proprietary',
# Automatically generate version number from git tags
use_scm_version=True,
packages=[
'pyfds'
],
# Runtime dependencies
install_requires=[
'numpy',
'scipy',
'matplotlib'
],
# Setup/build dependencies; setuptools_scm required for git-based versioning
setup_requires=['setuptools_scm'],
# For a list of valid classifiers, see
# See https://pypi.python.org/pypi?%3Aaction=list_classifiers for full list.
classifiers=[
'Development Status :: 2 - Pre-Alpha',
'Intended Audience :: Science/Research',
'License :: Other/Proprietary License',
'Programming Language :: Python :: 3',
'Topic :: Scientific/Engineering',
],
)
|
from setuptools import setup
from os import path
here = path.abspath(path.dirname(__file__))
with open(path.join(here, 'README.rst')) as readme_file:
long_description = readme_file.read()
setup(
name='pyfds',
description='Modular field simulation tool using finite differences.',
long_description=long_description,
url='http://emt.uni-paderborn.de',
author='Leander Claes',
author_email='claes@emt.uni-paderborn.de',
license='Proprietary',
# Automatically generate version number from git tags
use_scm_version=True,
packages=[
'pyfds'
],
# Runtime dependencies
install_requires=[
'numpy',
'scipy',
'matplotlib',
'siprefix'
],
# Setup/build dependencies; setuptools_scm required for git-based versioning
setup_requires=['setuptools_scm'],
# For a list of valid classifiers, see
# See https://pypi.python.org/pypi?%3Aaction=list_classifiers for full list.
classifiers=[
'Development Status :: 2 - Pre-Alpha',
'Intended Audience :: Science/Research',
'License :: Other/Proprietary License',
'Programming Language :: Python :: 3',
'Topic :: Scientific/Engineering',
],
)
|
Add siprefix to runtime dependencies.
|
Add siprefix to runtime dependencies.
|
Python
|
bsd-3-clause
|
emtpb/pyfds
|
---
+++
@@ -27,7 +27,8 @@
install_requires=[
'numpy',
'scipy',
- 'matplotlib'
+ 'matplotlib',
+ 'siprefix'
],
# Setup/build dependencies; setuptools_scm required for git-based versioning
|
a25f36c76301f67beab29dc2a273d819c49cd8eb
|
setup.py
|
setup.py
|
import os
import sys
try:
from setuptools import setup
except:
from distutils.core import setup
from distutils.command.build_py import build_py
path, script = os.path.split(sys.argv[0])
os.chdir(os.path.abspath(path))
VERSION="0.0.3"
install_requires = []
setup(name='imgix',
cmdclass={'build_py': build_py},
version=VERSION,
description='Python client library for imgix.',
author='imgix',
author_email='support@imgix.com',
url='http://www.imgix.com/',
packages=['imgix'],
install_requires=install_requires
)
|
import os
import sys
try:
from setuptools import setup
except:
from distutils.core import setup
from distutils.command.build_py import build_py
path, script = os.path.split(sys.argv[0])
os.chdir(os.path.abspath(path))
VERSION="0.0.3"
install_requires = []
setup(name='imgix',
cmdclass={'build_py': build_py},
version=VERSION,
description='Python client library for imgix.',
author='imgix',
author_email='support@imgix.com',
url='https://github.com/imgix/imgix-python',
packages=['imgix'],
install_requires=install_requires
)
|
Change the URL of the package
|
Change the URL of the package
|
Python
|
bsd-2-clause
|
imgix/imgix-python
|
---
+++
@@ -19,7 +19,7 @@
description='Python client library for imgix.',
author='imgix',
author_email='support@imgix.com',
- url='http://www.imgix.com/',
+ url='https://github.com/imgix/imgix-python',
packages=['imgix'],
install_requires=install_requires
)
|
b0e44fcca09a2d62ea0dc217d1538e03d48e2558
|
setup.py
|
setup.py
|
from setuptools import setup, find_packages
from buildcmds.addon import addon
setup(
name='io_scene_previz',
# Versions should comply with PEP440. For a discussion on single-sourcing
# the version across setup.py and the project code, see
# https://packaging.python.org/en/latest/single_source_version.html
version='0.0.7',
description='Blender Previz addon',
url='https://app.previz.co',
author='Previz',
author_email='info@previz.co',
license='MIT',
# See https://pypi.python.org/pypi?%3Aaction=list_classifiers
classifiers=[
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'Topic :: Multimedia :: Graphics :: 3D Modeling',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python :: 3',
],
keywords='previz development 3d scene exporter',
packages=find_packages(exclude=['tests']),
install_requires=['previz'],
extras_require={},
package_data={},
data_files=[],
cmdclass={
'addon': addon
}
)
|
from setuptools import setup, find_packages
from buildcmds.addon import bdist_blender_addon
setup(
name='io_scene_previz',
# Versions should comply with PEP440. For a discussion on single-sourcing
# the version across setup.py and the project code, see
# https://packaging.python.org/en/latest/single_source_version.html
version='0.0.7',
description='Blender Previz addon',
url='https://app.previz.co',
author='Previz',
author_email='info@previz.co',
license='MIT',
# See https://pypi.python.org/pypi?%3Aaction=list_classifiers
classifiers=[
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'Topic :: Multimedia :: Graphics :: 3D Modeling',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python :: 3',
],
keywords='previz development 3d scene exporter',
packages=find_packages(exclude=['buildcmds', 'tests']),
install_requires=['previz'],
extras_require={},
package_data={},
data_files=[],
cmdclass={
'bdist_blender_addon': bdist_blender_addon
}
)
|
Rename command addon to bdist_blender_addon
|
Rename command addon to bdist_blender_addon
|
Python
|
mit
|
Previz-app/io_scene_dnb_previz,Previz-app/io_scene_previz,Previz-app/io_scene_previz,Previz-app/io_scene_dnb_previz
|
---
+++
@@ -1,6 +1,6 @@
from setuptools import setup, find_packages
-from buildcmds.addon import addon
+from buildcmds.addon import bdist_blender_addon
setup(
name='io_scene_previz',
@@ -25,12 +25,12 @@
],
keywords='previz development 3d scene exporter',
- packages=find_packages(exclude=['tests']),
+ packages=find_packages(exclude=['buildcmds', 'tests']),
install_requires=['previz'],
extras_require={},
package_data={},
data_files=[],
cmdclass={
- 'addon': addon
+ 'bdist_blender_addon': bdist_blender_addon
}
)
|
3a3c1491cf185899a5e5b6288ae0a3542b536dee
|
setup.py
|
setup.py
|
from distutils.core import setup
setup(
name="pocketlint",
description="Pocket-lint a composite linter and style checker.",
version="0.5.7",
maintainer="Curtis C. Hovey",
maintainer_email="sinzui.is@verizon.net",
url="https://launchpad.net/pocket-lint",
packages=[
'pocketlint', 'pocketlint/contrib', 'pocketlint/contrib/pyflakes'],
scripts=['scripts/pocketlint'],
)
|
from distutils.core import setup
setup(
name="pocketlint",
description="Pocket-lint a composite linter and style checker.",
version="0.5.8",
maintainer="Curtis C. Hovey",
maintainer_email="sinzui.is@verizon.net",
url="https://launchpad.net/pocket-lint",
packages=[
'pocketlint', 'pocketlint/contrib', 'pocketlint/contrib/pyflakes'],
scripts=['scripts/pocketlint'],
)
|
Increment version to suport Python 2.6.
|
Increment version to suport Python 2.6.
|
Python
|
mit
|
chevah/pocket-lint,chevah/pocket-lint
|
---
+++
@@ -3,7 +3,7 @@
setup(
name="pocketlint",
description="Pocket-lint a composite linter and style checker.",
- version="0.5.7",
+ version="0.5.8",
maintainer="Curtis C. Hovey",
maintainer_email="sinzui.is@verizon.net",
url="https://launchpad.net/pocket-lint",
|
f52c77b746ca98ed06df293ba142a5ecf2577a62
|
setup.py
|
setup.py
|
import sys
from setuptools import setup
# Be verbose about Python < 3.4 being deprecated.
if sys.version_info < (3, 4):
print('\n' * 3 + '*' * 64)
print('lastcast requires Python 3.4+, and might be broken if run with\n'
'this version of Python.')
print('*' * 64 + '\n' * 3)
setup(
name='lastcast',
version='1.0.0',
description='Scrobble music to last.fm from Chromecast.',
author='Erik Price',
url='https://github.com/erik/lastcast',
packages=['lastcast'],
entry_points={
'console_scripts': [
'lastcast = lastcast:main',
],
},
license='MIT',
classifiers=[
"Programming Language :: Python",
"Programming Language :: Python :: 3",
"Operating System :: OS Independent",
],
install_requires=[
'PyChromecast==2.0.0',
'click==6.7',
'pylast==1.7.0',
'toml==0.9.4',
]
)
|
import sys
from setuptools import setup
# Be verbose about Python < 3.5 being deprecated.
if sys.version_info < (3, 5):
print('\n' * 3 + '*' * 64)
print('lastcast requires Python 3.5+, and might be broken if run with\n'
'this version of Python.')
print('*' * 64 + '\n' * 3)
setup(
name='lastcast',
version='1.0.1',
description='Scrobble music to last.fm from Chromecast.',
author='Erik Price',
url='https://github.com/erik/lastcast',
packages=['lastcast'],
entry_points={
'console_scripts': [
'lastcast = lastcast:main',
],
},
license='MIT',
classifiers=[
"Programming Language :: Python",
"Programming Language :: Python :: 3",
"Operating System :: OS Independent",
],
install_requires=[
'PyChromecast==2.0.0',
'click==6.7',
'pylast==1.7.0',
'toml==0.9.4',
]
)
|
Support is actually py >= 35
|
Support is actually py >= 35
|
Python
|
mit
|
erik/lastcast
|
---
+++
@@ -2,17 +2,17 @@
from setuptools import setup
-# Be verbose about Python < 3.4 being deprecated.
-if sys.version_info < (3, 4):
+# Be verbose about Python < 3.5 being deprecated.
+if sys.version_info < (3, 5):
print('\n' * 3 + '*' * 64)
- print('lastcast requires Python 3.4+, and might be broken if run with\n'
+ print('lastcast requires Python 3.5+, and might be broken if run with\n'
'this version of Python.')
print('*' * 64 + '\n' * 3)
setup(
name='lastcast',
- version='1.0.0',
+ version='1.0.1',
description='Scrobble music to last.fm from Chromecast.',
author='Erik Price',
url='https://github.com/erik/lastcast',
|
ab55fab78743650eb9bb219957d7305114e4f8ca
|
setup.py
|
setup.py
|
#!/usr/bin/env python
from setuptools import find_packages, Command
setup_params = dict(
name='bugimporters',
version=0.1,
author='Various contributers to the OpenHatch project, Berry Phillips',
author_email='all@openhatch.org, berryphillips@gmail.com',
packages=find_packages(),
description='Bug importers for the OpenHatch project',
install_requires=[
'gdata',
'lxml',
'pyopenssl',
'unicodecsv',
'feedparser',
'twisted',
'python-dateutil',
'decorator',
],
)
if __name__ == '__main__':
from setuptools import setup
setup(**setup_params)
|
#!/usr/bin/env python
from setuptools import find_packages, Command
setup_params = dict(
name='bugimporters',
version=0.1,
author='Various contributers to the OpenHatch project, Berry Phillips',
author_email='all@openhatch.org, berryphillips@gmail.com',
packages=find_packages(),
description='Bug importers for the OpenHatch project',
install_requires=[
'gdata',
'lxml',
'pyopenssl',
'unicodecsv',
'feedparser',
'twisted',
'python-dateutil',
'decorator',
'scrapy',
],
)
if __name__ == '__main__':
from setuptools import setup
setup(**setup_params)
|
Add a dependency on Scrapy
|
Add a dependency on Scrapy
|
Python
|
agpl-3.0
|
openhatch/oh-bugimporters,openhatch/oh-bugimporters,openhatch/oh-bugimporters
|
---
+++
@@ -18,6 +18,7 @@
'twisted',
'python-dateutil',
'decorator',
+ 'scrapy',
],
)
|
3d888afa88326c97246947141c357509c2f72bbc
|
setup.py
|
setup.py
|
from distutils.core import setup
setup(
name='firebase-token-generator',
version='1.2',
author='Greg Soltis',
author_email='greg@firebase.com',
py_modules=['firebase_token_generator'],
license='LICENSE',
url='https://github.com/firebase/firebase-token-generator-python',
description='A utility to generate signed Firebase Authentication Tokens',
long_description=open('README.md').read()
)
|
from distutils.core import setup
setup(
name='firebase-token-generator',
version='1.3',
author='Greg Soltis',
author_email='greg@firebase.com',
zip_safe=False,
py_modules=['firebase_token_generator'],
license='LICENSE',
url='https://github.com/firebase/firebase-token-generator-python',
description='A utility to generate signed Firebase Authentication Tokens',
long_description=open('README.md').read()
)
|
Set zip_safe=False. Bump version to 1.3.
|
Set zip_safe=False. Bump version to 1.3.
|
Python
|
mit
|
googlearchive/firebase-token-generator-python
|
---
+++
@@ -2,9 +2,10 @@
setup(
name='firebase-token-generator',
- version='1.2',
+ version='1.3',
author='Greg Soltis',
author_email='greg@firebase.com',
+ zip_safe=False,
py_modules=['firebase_token_generator'],
license='LICENSE',
url='https://github.com/firebase/firebase-token-generator-python',
|
927e2bc4b8774cdffd753f99f63086baf437a509
|
setup.py
|
setup.py
|
from pip.download import PipSession
from pip.req import parse_requirements
from setuptools import setup
setup(
name='aws-portknock',
version='0.1',
py_modules=['aws_portknock'],
description='Port knocking for AWS security groups',
author='Michel Alexandre Salim',
author_email='michel@michel-slm.name',
url='https://github.com/michel-slm/aws-portknock',
download_url='https://github.com/michel-slm/aws-portknock/tarball/0.1',
keywords=['aws'],
classifiers=[],
install_requires=[str(r.req)
for r in parse_requirements(
'requirements.txt',
session=PipSession())],
entry_points='''
[console_scripts]
aws-portknock=aws_portknock:cli
''',
)
|
import os
from pip.download import PipSession
from pip.req import parse_requirements
from setuptools import setup
BASE_DIR = os.path.dirname(os.path.realpath(__file__))
reqs_file = os.path.join(BASE_DIR, 'requirements.txt')
install_reqs = parse_requirements(reqs_file, session=PipSession())
setup(
name='aws-portknock',
version='0.1',
py_modules=['aws_portknock'],
description='Port knocking for AWS security groups',
author='Michel Alexandre Salim',
author_email='michel@michel-slm.name',
url='https://github.com/michel-slm/aws-portknock',
download_url='https://github.com/michel-slm/aws-portknock/tarball/0.1',
keywords=['aws'],
classifiers=[],
install_requires=[str(r.req)
for r in install_reqs],
entry_points='''
[console_scripts]
aws-portknock=aws_portknock:cli
''',
)
|
Add path to requirements.txt so installation from pip succeeds
|
Add path to requirements.txt so installation from pip succeeds
cf http://lorenamesa.com/packaging-my-first-python-egg.html
|
Python
|
mpl-2.0
|
michel-slm/aws-portknock
|
---
+++
@@ -1,6 +1,11 @@
+import os
from pip.download import PipSession
from pip.req import parse_requirements
from setuptools import setup
+
+BASE_DIR = os.path.dirname(os.path.realpath(__file__))
+reqs_file = os.path.join(BASE_DIR, 'requirements.txt')
+install_reqs = parse_requirements(reqs_file, session=PipSession())
setup(
name='aws-portknock',
@@ -14,9 +19,7 @@
keywords=['aws'],
classifiers=[],
install_requires=[str(r.req)
- for r in parse_requirements(
- 'requirements.txt',
- session=PipSession())],
+ for r in install_reqs],
entry_points='''
[console_scripts]
aws-portknock=aws_portknock:cli
|
03875be96ba09a252a05ba0ea84a3558c6428011
|
setup.py
|
setup.py
|
#!/usr/bin/env python
import os
from setuptools import setup, find_packages
reqs_file = os.path.join(os.path.dirname(os.path.realpath(__file__))
, "requirements.txt")
reqs = None
with open(reqs_file) as f:
reqs = f.readlines()
setup(
version='0.1.0',
name='mediachain-client',
description='mediachain reader command line interface',
author='Mediachain Labs',
packages=find_packages('.'),
entry_points={
'console_scripts': [
'mediachain = mediachain.cli.main:main'
]
},
url='http://mediachain.io',
install_requires=reqs,
)
|
#!/usr/bin/env python
import os, sys
from setuptools import setup, find_packages
from setuptools.command.install import install as _install
reqs_file = os.path.join(os.path.dirname(os.path.realpath(__file__))
, "requirements.txt")
reqs = None
with open(reqs_file) as f:
reqs = f.readlines()
def _pre_install(dir):
from subprocess import check_call
check_call(['scripts/build_grpc.sh'],
cwd=dir)
class install(_install):
def run(self):
self.execute(_pre_install, [os.path.dirname(__file__)],
msg="Generating protobuf")
_install.run(self)
setup(
version='0.1.0',
name='mediachain-client',
description='mediachain reader command line interface',
author='Mediachain Labs',
packages=find_packages('.'),
entry_points={
'console_scripts': [
'mediachain = mediachain.cli.main:main'
]
},
url='http://mediachain.io',
install_requires=reqs,
cmdclass={'install': install},
)
|
Add pre-install protobuf gen script
|
Add pre-install protobuf gen script
|
Python
|
mit
|
mediachain/mediachain-client,mediachain/mediachain-client
|
---
+++
@@ -1,6 +1,7 @@
#!/usr/bin/env python
-import os
+import os, sys
from setuptools import setup, find_packages
+from setuptools.command.install import install as _install
reqs_file = os.path.join(os.path.dirname(os.path.realpath(__file__))
, "requirements.txt")
@@ -8,6 +9,17 @@
reqs = None
with open(reqs_file) as f:
reqs = f.readlines()
+
+def _pre_install(dir):
+ from subprocess import check_call
+ check_call(['scripts/build_grpc.sh'],
+ cwd=dir)
+
+class install(_install):
+ def run(self):
+ self.execute(_pre_install, [os.path.dirname(__file__)],
+ msg="Generating protobuf")
+ _install.run(self)
setup(
version='0.1.0',
@@ -22,4 +34,5 @@
},
url='http://mediachain.io',
install_requires=reqs,
+ cmdclass={'install': install},
)
|
453bd468d9de609999ef814f24111923177374fe
|
setup.py
|
setup.py
|
from setuptools import find_packages
from setuptools import setup
from pip import req
_install_requirements = req.parse_requirements('requirements.txt')
setup(
name='jetway',
version=open('VERSION').read().strip(),
description=(
'Client library for the Jetway static site staging service.'
),
url='https://github.com/grow/jetway-client',
license='MIT',
author='Grow SDK Authors',
author_email='hello@grow.io',
include_package_data=True,
install_requires=[str(ir.req) for ir in _install_requirements],
packages=find_packages(),
keywords=[
'grow',
'cms',
'static site generator',
's3',
'google cloud storage',
'content management'
],
classifiers=[
'Development Status :: 2 - Pre-Alpha',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Natural Language :: English',
'Operating System :: OS Independent',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
'Topic :: Internet :: WWW/HTTP',
'Topic :: Software Development :: Libraries :: Python Modules',
])
|
import uuid
from setuptools import find_packages
from setuptools import setup
from pip import req
_install_requirements = req.parse_requirements(
'requirements.txt', session=uuid.uuid1())
setup(
name='jetway',
version=open('VERSION').read().strip(),
description=(
'Client library for the Jetway static site staging service.'
),
url='https://github.com/grow/jetway-client',
license='MIT',
author='Grow SDK Authors',
author_email='hello@grow.io',
include_package_data=True,
install_requires=[str(ir.req) for ir in _install_requirements],
packages=find_packages(),
keywords=[
'grow',
'cms',
'static site generator',
's3',
'google cloud storage',
'content management'
],
classifiers=[
'Development Status :: 2 - Pre-Alpha',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Natural Language :: English',
'Operating System :: OS Independent',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
'Topic :: Internet :: WWW/HTTP',
'Topic :: Software Development :: Libraries :: Python Modules',
])
|
Fix pip parse_requirements for v0.6+.
|
Fix pip parse_requirements for v0.6+.
|
Python
|
mit
|
grow/webreview-client
|
---
+++
@@ -1,8 +1,10 @@
+import uuid
from setuptools import find_packages
from setuptools import setup
from pip import req
-_install_requirements = req.parse_requirements('requirements.txt')
+_install_requirements = req.parse_requirements(
+ 'requirements.txt', session=uuid.uuid1())
setup(
|
ee2d27eca45768a07a562405cf4431cb8d2b09bf
|
setup.py
|
setup.py
|
from distutils.core import setup
setup(name='pyresttest',
version='0.1',
description='Python Rest Testing',
maintainer='Naveen Malik',
maintainer_email='jewzaam@gmail.com',
url='https://github.com/svanoort/pyresttest',
py_modules=['resttest','pycurl_benchmark','test_resttest'],
license='Apache License, Version 2.0'
)
|
from distutils.core import setup
setup(name='pyresttest',
version='0.1',
description='Python Rest Testing',
maintainer='Sam Van Oort',
maintainer_email='acetonespam@gmail.com',
url='https://github.com/svanoort/pyresttest',
py_modules=['resttest','test_resttest'],
license='Apache License, Version 2.0',
requires=['argparse','yaml','pycurl']
)
|
Set maintainer and add dependencies to distutils config
|
Set maintainer and add dependencies to distutils config
|
Python
|
apache-2.0
|
sunyanhui/pyresttest,satish-suradkar/pyresttest,suvarnaraju/pyresttest,wirewit/pyresttest,netjunki/pyresttest,MorrisJobke/pyresttest,wirewit/pyresttest,suvarnaraju/pyresttest,svanoort/pyresttest,alazaro/pyresttest,sunyanhui/pyresttest,TimYi/pyresttest,MorrisJobke/pyresttest,holdenweb/pyresttest,TimYi/pyresttest,alazaro/pyresttest,janusnic/pyresttest,janusnic/pyresttest,holdenweb/pyresttest,netjunki/pyresttest,svanoort/pyresttest,satish-suradkar/pyresttest
|
---
+++
@@ -3,9 +3,10 @@
setup(name='pyresttest',
version='0.1',
description='Python Rest Testing',
- maintainer='Naveen Malik',
- maintainer_email='jewzaam@gmail.com',
+ maintainer='Sam Van Oort',
+ maintainer_email='acetonespam@gmail.com',
url='https://github.com/svanoort/pyresttest',
- py_modules=['resttest','pycurl_benchmark','test_resttest'],
- license='Apache License, Version 2.0'
+ py_modules=['resttest','test_resttest'],
+ license='Apache License, Version 2.0',
+ requires=['argparse','yaml','pycurl']
)
|
b5898dc6d16868cd7908872916fa6099ad85049f
|
setup.py
|
setup.py
|
#!/usr/bin/env python
from setuptools import setup, find_packages
setup(
name='prompt_toolkit',
author='Jonathan Slenders',
version='0.25',
license='LICENSE.txt',
url='https://github.com/jonathanslenders/python-prompt-toolkit',
description='Library for building powerful interactive command lines in Python',
long_description='',
packages=find_packages('.'),
install_requires = [
'docopt',
'jedi>=0.8.1-final0',
'pygments',
'six>=1.8.0',
'wcwidth',
],
entry_points={
'console_scripts': [
'ptpython = prompt_toolkit.contrib.entry_points.ptpython:run',
'ptipython = prompt_toolkit.contrib.entry_points.ptipython:run',
]
},
extras_require = {
'ptipython': ['ipython'] # For ptipython, we need to have IPython
}
)
|
#!/usr/bin/env python
from setuptools import setup, find_packages
setup(
name='prompt_toolkit',
author='Jonathan Slenders',
version='0.25',
license='LICENSE.txt',
url='https://github.com/jonathanslenders/python-prompt-toolkit',
description='Library for building powerful interactive command lines in Python',
long_description='',
packages=find_packages('.'),
install_requires = [
'docopt',
'jedi>=0.8.1',
'pygments',
'six>=1.8.0',
'wcwidth',
],
entry_points={
'console_scripts': [
'ptpython = prompt_toolkit.contrib.entry_points.ptpython:run',
'ptipython = prompt_toolkit.contrib.entry_points.ptipython:run',
]
},
extras_require = {
'ptipython': ['ipython'] # For ptipython, we need to have IPython
}
)
|
Use again jedi==0.8.1. (This was pushed to pypi.)
|
Use again jedi==0.8.1. (This was pushed to pypi.)
|
Python
|
bsd-3-clause
|
melund/python-prompt-toolkit,ALSchwalm/python-prompt-toolkit,niklasf/python-prompt-toolkit,amjith/python-prompt-toolkit,jonathanslenders/python-prompt-toolkit,jaseg/python-prompt-toolkit,ddalex/python-prompt-toolkit
|
---
+++
@@ -14,7 +14,7 @@
packages=find_packages('.'),
install_requires = [
'docopt',
- 'jedi>=0.8.1-final0',
+ 'jedi>=0.8.1',
'pygments',
'six>=1.8.0',
'wcwidth',
|
2ac5d587c057d877db217b0303a06e0382bd8a2d
|
setup.py
|
setup.py
|
#!/usr/bin/python3
# SPDX-License-Identifier: LGPL-2.1+
import sys
from setuptools import setup
if sys.version_info < (3, 6):
sys.exit("Sorry, we need at least Python 3.6.")
setup(
name="mkosi",
version="5",
description="Create legacy-free OS images",
url="https://github.com/systemd/mkosi",
maintainer="mkosi contributors",
maintainer_email="systemd-devel@lists.freedesktop.org",
license="LGPLv2+",
scripts=["mkosi"],
)
|
#!/usr/bin/python3
# SPDX-License-Identifier: LGPL-2.1+
import sys
from setuptools import setup, Command
class BuildManpage(Command):
description = ('builds the manpage')
user_options = []
def initialize_options(self):
pass
def finalize_options(self):
pass
def run(self):
self.spawn(['pandoc', '-t', 'man', '-o', 'mkosi.1', 'mkosi.md'])
if sys.version_info < (3, 6):
sys.exit("Sorry, we need at least Python 3.6.")
setup(
name="mkosi",
version="5",
description="Create legacy-free OS images",
url="https://github.com/systemd/mkosi",
maintainer="mkosi contributors",
maintainer_email="systemd-devel@lists.freedesktop.org",
license="LGPLv2+",
scripts=["mkosi"],
cmdclass = { "man": BuildManpage }
)
|
Add custom command to build the manpage
|
Add custom command to build the manpage
Remembering the exact pandoc incantation is annoying
|
Python
|
lgpl-2.1
|
systemd/mkosi,systemd/mkosi
|
---
+++
@@ -3,7 +3,19 @@
import sys
-from setuptools import setup
+from setuptools import setup, Command
+
+class BuildManpage(Command):
+ description = ('builds the manpage')
+ user_options = []
+
+ def initialize_options(self):
+ pass
+ def finalize_options(self):
+ pass
+
+ def run(self):
+ self.spawn(['pandoc', '-t', 'man', '-o', 'mkosi.1', 'mkosi.md'])
if sys.version_info < (3, 6):
sys.exit("Sorry, we need at least Python 3.6.")
@@ -18,4 +30,5 @@
maintainer_email="systemd-devel@lists.freedesktop.org",
license="LGPLv2+",
scripts=["mkosi"],
+ cmdclass = { "man": BuildManpage }
)
|
8fea58292e41352b0b58947f4182dd32ff4f225d
|
opps/fields/models.py
|
opps/fields/models.py
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
from django.db import models
from django.utils.translation import ugettext_lazy as _
from opps.boxes.models import OPPS_APPS
FIELD_TYPE = (
('checkbox', _('CheckBox')),
('radio', _('Radio')),
('text', _('Text')),
('textarea', _('TextArea')),
)
class Field(models.Model):
name = models.CharField(_('Name'), max_length=100)
slug = models.SlugField(_('Slug'), max_length=255)
application = models.CharField(_('Application'),
max_length=255,
choices=OPPS_APPS,
db_index=True)
type = models.CharField(_("Type"), max_length=15,
choices=FIELD_TYPE,
db_index=True)
def __unicode__(self):
return u"{} - {}".format(self.application, self.name)
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
from django.db import models
from django.utils.translation import ugettext_lazy as _
from opps.boxes.models import OPPS_APPS
FIELD_TYPE = (
('checkbox', _('CheckBox')),
('radio', _('Radio')),
('text', _('Text')),
('textarea', _('TextArea')),
)
class Field(models.Model):
name = models.CharField(_('Name'), max_length=100)
slug = models.SlugField(_('Slug'), max_length=255)
application = models.CharField(_('Application'),
max_length=255,
choices=OPPS_APPS,
db_index=True)
type = models.CharField(_("Type"), max_length=15,
choices=FIELD_TYPE,
db_index=True)
def __unicode__(self):
return u"{} - {}".format(self.application, self.name)
class Option(models.Model):
field = models.ForeignKey('fields.Field')
name = models.CharField(_('Name'), max_length=100)
slug = models.SlugField(_('Slug'), max_length=140)
value = models.CharField(_('Value'), max_length=255)
def __unicode__(self):
return u"{} - {}".format(self.field.slug, self.name)
class FieldOption(models.Model):
field = models.ForeignKey('fields.Field')
option = models.ForeignKey('fields.Option')
order = models.PositiveIntegerField(_(u'Order'), default=0)
def __unicode__(self):
return u"{} - {}".format(self.field.slug, self.option.slug)
class Meta:
ordering = ['-order']
|
Add new model option to add field options if exist (radio/checkbox)
|
Add new model option to add field options if exist (radio/checkbox)
|
Python
|
mit
|
williamroot/opps,jeanmask/opps,jeanmask/opps,williamroot/opps,YACOWS/opps,williamroot/opps,opps/opps,opps/opps,YACOWS/opps,jeanmask/opps,jeanmask/opps,YACOWS/opps,opps/opps,williamroot/opps,YACOWS/opps,opps/opps
|
---
+++
@@ -27,3 +27,26 @@
def __unicode__(self):
return u"{} - {}".format(self.application, self.name)
+
+
+class Option(models.Model):
+ field = models.ForeignKey('fields.Field')
+ name = models.CharField(_('Name'), max_length=100)
+ slug = models.SlugField(_('Slug'), max_length=140)
+ value = models.CharField(_('Value'), max_length=255)
+
+ def __unicode__(self):
+ return u"{} - {}".format(self.field.slug, self.name)
+
+
+
+class FieldOption(models.Model):
+ field = models.ForeignKey('fields.Field')
+ option = models.ForeignKey('fields.Option')
+ order = models.PositiveIntegerField(_(u'Order'), default=0)
+
+ def __unicode__(self):
+ return u"{} - {}".format(self.field.slug, self.option.slug)
+
+ class Meta:
+ ordering = ['-order']
|
4f730cb70fc4e65b568732fdcbb7f0209870bfca
|
osf_tests/settings.py
|
osf_tests/settings.py
|
# Use API defaults. This allows these settings to work with API tests
from api.base.settings.defaults import * # noqa
DEBUG_PROPAGATE_EXCEPTIONS = True
#DATABASES = {
# 'default': {
# 'CONN_MAX_AGE': 0,
# 'ENGINE': 'osf.db.backends.postgresql',
# 'HOST': '',
# 'NAME': 'osf-models-test',
# 'PASSWORD': '',
# 'PORT': '',
# 'USER': '',
# 'ATOMIC_REQUESTS': True,
# }
#}
SITE_ID = 1
# SECRET_KEY = 'not very secret in tests'
PASSWORD_HASHERS = (
'django.contrib.auth.hashers.MD5PasswordHasher',
'django.contrib.auth.hashers.SHA1PasswordHasher',
)
REST_FRAMEWORK['ALLOWED_VERSIONS'] = (
'2.0',
'2.1',
'2.2',
'2.3',
'3.0',
'3.0.1',
)
|
# Use API defaults. This allows these settings to work with API tests
from api.base.settings.defaults import * # noqa
DEBUG_PROPAGATE_EXCEPTIONS = True
#DATABASES = {
# 'default': {
# 'CONN_MAX_AGE': 0,
# 'ENGINE': 'osf.db.backends.postgresql',
# 'HOST': '',
# 'NAME': 'osf-models-test',
# 'PASSWORD': '',
# 'PORT': '',
# 'USER': '',
# 'ATOMIC_REQUESTS': True,
# }
#}
SITE_ID = 1
# SECRET_KEY = 'not very secret in tests'
PASSWORD_HASHERS = (
'django.contrib.auth.hashers.MD5PasswordHasher',
'django.contrib.auth.hashers.SHA1PasswordHasher',
)
REST_FRAMEWORK['ALLOWED_VERSIONS'] = (
'2.0',
'2.0.1',
'2.1',
'2.2',
'2.3',
'3.0',
'3.0.1',
)
|
Add in API testing version
|
Add in API testing version
Signed-off-by: Chris Wisecarver <5fccdd17c1f7bcc7e393d2cb5e2fad37705ca69f@cos.io>
|
Python
|
apache-2.0
|
cwisecarver/osf.io,chennan47/osf.io,mfraezz/osf.io,Johnetordoff/osf.io,cslzchen/osf.io,monikagrabowska/osf.io,baylee-d/osf.io,baylee-d/osf.io,crcresearch/osf.io,saradbowman/osf.io,monikagrabowska/osf.io,mattclark/osf.io,CenterForOpenScience/osf.io,Nesiehr/osf.io,brianjgeiger/osf.io,CenterForOpenScience/osf.io,Nesiehr/osf.io,felliott/osf.io,erinspace/osf.io,felliott/osf.io,acshi/osf.io,brianjgeiger/osf.io,laurenrevere/osf.io,baylee-d/osf.io,felliott/osf.io,TomBaxter/osf.io,mfraezz/osf.io,chrisseto/osf.io,monikagrabowska/osf.io,caseyrollins/osf.io,CenterForOpenScience/osf.io,caneruguz/osf.io,chennan47/osf.io,leb2dg/osf.io,icereval/osf.io,crcresearch/osf.io,cslzchen/osf.io,chrisseto/osf.io,Johnetordoff/osf.io,erinspace/osf.io,TomBaxter/osf.io,adlius/osf.io,laurenrevere/osf.io,caseyrollins/osf.io,leb2dg/osf.io,cwisecarver/osf.io,acshi/osf.io,acshi/osf.io,chennan47/osf.io,hmoco/osf.io,adlius/osf.io,laurenrevere/osf.io,binoculars/osf.io,cslzchen/osf.io,hmoco/osf.io,mfraezz/osf.io,cwisecarver/osf.io,brianjgeiger/osf.io,sloria/osf.io,icereval/osf.io,cwisecarver/osf.io,adlius/osf.io,HalcyonChimera/osf.io,hmoco/osf.io,mfraezz/osf.io,aaxelb/osf.io,Johnetordoff/osf.io,caseyrollins/osf.io,CenterForOpenScience/osf.io,monikagrabowska/osf.io,caneruguz/osf.io,mattclark/osf.io,sloria/osf.io,hmoco/osf.io,HalcyonChimera/osf.io,brianjgeiger/osf.io,monikagrabowska/osf.io,caneruguz/osf.io,felliott/osf.io,aaxelb/osf.io,pattisdr/osf.io,HalcyonChimera/osf.io,sloria/osf.io,saradbowman/osf.io,acshi/osf.io,Nesiehr/osf.io,caneruguz/osf.io,pattisdr/osf.io,Johnetordoff/osf.io,chrisseto/osf.io,leb2dg/osf.io,binoculars/osf.io,Nesiehr/osf.io,pattisdr/osf.io,acshi/osf.io,aaxelb/osf.io,icereval/osf.io,HalcyonChimera/osf.io,TomBaxter/osf.io,mattclark/osf.io,adlius/osf.io,leb2dg/osf.io,cslzchen/osf.io,binoculars/osf.io,chrisseto/osf.io,erinspace/osf.io,crcresearch/osf.io,aaxelb/osf.io
|
---
+++
@@ -23,6 +23,7 @@
REST_FRAMEWORK['ALLOWED_VERSIONS'] = (
'2.0',
+ '2.0.1',
'2.1',
'2.2',
'2.3',
|
6f83fb7dd071786dc01a015addbdb541e7eaf7db
|
meinberlin/apps/documents/migrations/0002_rename_document_to_chapter.py
|
meinberlin/apps/documents/migrations/0002_rename_document_to_chapter.py
|
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import migrations
from django.db import models
class Migration(migrations.Migration):
dependencies = [
('meinberlin_documents', '0001_initial'),
]
operations = [
migrations.RenameModel(
old_name='Document',
new_name='Chapter',
),
migrations.RenameField(
model_name='paragraph',
old_name='document',
new_name='chapter',
),
]
|
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import migrations
from django.db import models
class Migration(migrations.Migration):
atomic=False
dependencies = [
('meinberlin_documents', '0001_initial'),
]
operations = [
migrations.RenameModel(
old_name='Document',
new_name='Chapter',
),
migrations.RenameField(
model_name='paragraph',
old_name='document',
new_name='chapter',
),
]
|
Work around a migration issue in sqlite
|
apps/documents: Work around a migration issue in sqlite
|
Python
|
agpl-3.0
|
liqd/a4-meinberlin,liqd/a4-meinberlin,liqd/a4-meinberlin,liqd/a4-meinberlin
|
---
+++
@@ -6,6 +6,7 @@
class Migration(migrations.Migration):
+ atomic=False
dependencies = [
('meinberlin_documents', '0001_initial'),
|
408ec05722cacd1e8574d9af206bffcf319c23b2
|
pipes/iam/__main__.py
|
pipes/iam/__main__.py
|
"""Create IAM Instance Profiles, Roles, Users, and Groups."""
import argparse
import logging
from .create_iam import create_iam_resources
LOG = logging.getLogger(__name__)
def main():
"""Command to create IAM Instance Profiles, Roles, Users, and Groups."""
logging.basicConfig()
parser = argparse.ArgumentParser(description=main.__doc__)
parser.add_argument('-d',
'--debug',
action='store_const',
const=logging.DEBUG,
default=logging.INFO,
help='Set DEBUG output')
parser.add_argument('-e',
'--env',
choices=('dev', 'stage', 'prod'),
default='dev',
help='Deploy environment')
parser.add_argument('-a',
'--app',
default='testapp',
help='Spinnaker Application name')
args = parser.parse_args()
LOG.setLevel(args.debug)
logging.getLogger(__package__).setLevel(args.debug)
vars(args).pop('debug')
assert create_iam_resources(env=args.env, app=args.app)
if __name__ == '__main__':
main()
|
"""Create IAM Instance Profiles, Roles, Users, and Groups."""
import argparse
import logging
from .create_iam import create_iam_resources
LOG = logging.getLogger(__name__)
def main():
"""Command to create IAM Instance Profiles, Roles, Users, and Groups."""
logging.basicConfig()
parser = argparse.ArgumentParser(description=main.__doc__)
parser.add_argument('-d',
'--debug',
action='store_const',
const=logging.DEBUG,
default=logging.INFO,
help='Set DEBUG output')
parser.add_argument('-e',
'--env',
choices=('build', 'dev', 'stage', 'prod'),
default='dev',
help='Deploy environment')
parser.add_argument('-a',
'--app',
default='testapp',
help='Spinnaker Application name')
args = parser.parse_args()
LOG.setLevel(args.debug)
logging.getLogger(__package__).setLevel(args.debug)
vars(args).pop('debug')
assert create_iam_resources(env=args.env, app=args.app)
if __name__ == '__main__':
main()
|
Add build as available environment
|
Add build as available environment
|
Python
|
apache-2.0
|
gogoair/foremast,gogoair/foremast
|
---
+++
@@ -20,7 +20,7 @@
help='Set DEBUG output')
parser.add_argument('-e',
'--env',
- choices=('dev', 'stage', 'prod'),
+ choices=('build', 'dev', 'stage', 'prod'),
default='dev',
help='Deploy environment')
parser.add_argument('-a',
|
3d2f19ff097cf144efd9135c52e4d584193f9ddb
|
tohu/v7/custom_generator/tohu_items_class.py
|
tohu/v7/custom_generator/tohu_items_class.py
|
import attr
__all__ = ["make_tohu_items_class"]
def make_tohu_items_class(clsname, field_names):
"""
Parameters
----------
clsname: string
Name of the class to be created.
field_names: list of strings
Names of the field attributes of the class to be created.
"""
item_cls = attr.make_class(clsname, {name: attr.ib() for name in field_names}, repr=True, cmp=True, frozen=True)
func_eq_orig = item_cls.__eq__
def func_eq_new(self, other):
"""
Custom __eq__() method which also allows comparisons with
tuples and dictionaries. This is mostly for convenience
during testing.
"""
if isinstance(other, self.__class__):
return func_eq_orig(self, other)
else:
if isinstance(other, tuple):
return attr.astuple(self) == other
elif isinstance(other, dict):
return attr.asdict(self) == other
else:
raise TypeError(
f"Tohu items have types that cannot be compared: "
"{self.__class__.__name__}, {other.__class__.__name__}"
)
item_cls.__eq__ = func_eq_new
item_cls.field_names = field_names
item_cls.as_dict = lambda self: attr.asdict(self)
item_cls.as_tuple = lambda self: attr.astuple(self)
return item_cls
|
import attr
__all__ = ["make_tohu_items_class"]
def make_tohu_items_class(clsname, field_names):
"""
Parameters
----------
clsname: string
Name of the class to be created.
field_names: list of strings
Names of the field attributes of the class to be created.
"""
item_cls = attr.make_class(clsname, {name: attr.ib() for name in field_names}, repr=True, cmp=True, frozen=True)
func_eq_orig = item_cls.__eq__
def func_eq_new(self, other):
"""
Custom __eq__() method which also allows comparisons with
tuples and dictionaries. This is mostly for convenience
during testing.
"""
if isinstance(other, self.__class__):
return func_eq_orig(self, other)
else:
if isinstance(other, tuple):
return attr.astuple(self) == other
elif isinstance(other, dict):
return attr.asdict(self) == other
else:
raise TypeError(
f"Tohu items have types that cannot be compared: "
"{self.__class__.__name__}, {other.__class__.__name__}"
)
item_cls.__eq__ = func_eq_new
item_cls.field_names = field_names
item_cls.as_dict = lambda self: attr.asdict(self)
item_cls.as_tuple = lambda self: attr.astuple(self)
item_cls.is_unset = False
return item_cls
|
Add attribute 'is_unset' so that the interface is consistent with MissingTohuItemsCls
|
Add attribute 'is_unset' so that the interface is consistent with MissingTohuItemsCls
|
Python
|
mit
|
maxalbert/tohu
|
---
+++
@@ -40,4 +40,5 @@
item_cls.field_names = field_names
item_cls.as_dict = lambda self: attr.asdict(self)
item_cls.as_tuple = lambda self: attr.astuple(self)
+ item_cls.is_unset = False
return item_cls
|
445b80562e038bc3749930d44e00eda55edaa180
|
ci_scripts/buildLinuxWheels.py
|
ci_scripts/buildLinuxWheels.py
|
from subprocess import call, check_output
import sys
import os
isPython3 = sys.version_info.major == 3
# https://stackoverflow.com/a/3357357
command = 'git log --format=%B -n 1'.split()
out = check_output(command)
if b'build wheels' not in out.lower() or not isPython3:
exit(0)
path = os.path.abspath(argv[1])
call('pip install cibuildwheel==0.7.0'.split())
call('cibuildwheel --output-dir {}'.format(path).split())
from dropboxUpload import uploadAll
uploadAll(path)
|
from subprocess import call, check_output
import sys
import os
isPython3 = sys.version_info.major == 3
# https://stackoverflow.com/a/3357357
command = 'git log --format=%B -n 1'.split()
out = check_output(command)
if b'build wheels' not in out.lower() or not isPython3:
exit(0)
path = os.path.abspath(sys.argv[1])
call('pip install cibuildwheel==0.7.0'.split())
call('cibuildwheel --output-dir {}'.format(sys.argv[1]).split())
from dropboxUpload import uploadAll
uploadAll(path)
|
Fix build wheels and upload 4.
|
Fix build wheels and upload 4.
|
Python
|
bsd-3-clause
|
jr-garcia/AssimpCy,jr-garcia/AssimpCy
|
---
+++
@@ -11,9 +11,9 @@
if b'build wheels' not in out.lower() or not isPython3:
exit(0)
-path = os.path.abspath(argv[1])
+path = os.path.abspath(sys.argv[1])
call('pip install cibuildwheel==0.7.0'.split())
-call('cibuildwheel --output-dir {}'.format(path).split())
+call('cibuildwheel --output-dir {}'.format(sys.argv[1]).split())
from dropboxUpload import uploadAll
|
bec498894ebd612058f905e1cfe4f5230df91841
|
client/examples/cycle-cards.py
|
client/examples/cycle-cards.py
|
#!/bin/python
import removinator
import subprocess
# This example cycles through each card slot in the Removinator. Any
# slots that have a card present will then have the certificates on the
# card printed out using the pkcs15-tool utility, which is provided by
# the OpenSC project.
#
# Examples of parsing the Removinator status output and enabling debug
# output from the firmware are also provided.
print('--- Connecting to Removinator ---')
ctl = removinator.Removinator()
print('--- Cycling through cards ---')
for card in range(1, 9):
try:
ctl.insert_card(card)
print('Inserted card {0}'.format(card))
print('{0}'.format(subprocess.check_output(['pkcs15-tool',
'--list-certificates'])
.rstrip()))
except removinator.SlotError:
print('Card {0} is not inserted'.format(card))
print('--- Checking Removinator status ---')
status = ctl.get_status()
print('Current card: {0}'.format(status['current']))
for card in status['present']:
print('Card {0} is present'.format(card))
print('--- Debug output for re-insertion of current card ---')
ctl.set_debug(True)
ctl.insert_card(status['current'])
print('{0}'.format(ctl.last_response.rstrip()))
ctl.set_debug(False)
print('--- Remove current card ---')
ctl.remove_card()
|
#!/bin/python
import removinator
import subprocess
# This example cycles through each card slot in the Removinator. Any
# slots that have a card present will then have the certificates on the
# card printed out using the pkcs15-tool utility, which is provided by
# the OpenSC project.
#
# Examples of parsing the Removinator status output and enabling debug
# output from the firmware are also provided.
print('--- Connecting to Removinator ---')
ctl = removinator.Removinator()
print('--- Cycling through cards ---')
for card in range(1, 9):
try:
ctl.insert_card(card)
print('Inserted card {0}'.format(card))
print('{0}'.format(subprocess.check_output(['pkcs15-tool',
'--list-certificates'])
.rstrip()))
except removinator.SlotError:
print('Card {0} is not inserted'.format(card))
print('--- Checking Removinator status ---')
status = ctl.get_status()
print('Current card: {0}'.format(status['current']))
for card in status['present']:
print('Card {0} is present'.format(card))
print('--- Debug output for re-insertion of current card ---')
ctl.set_debug(True)
ctl.insert_card(status['current'])
print('{0}'.format(ctl.last_response.rstrip()))
ctl.set_debug(False)
print('--- Remove current card ---')
ctl.remove_card()
|
Correct PEP8 indendation issue in client example
|
Correct PEP8 indendation issue in client example
|
Python
|
apache-2.0
|
nkinder/smart-card-removinator
|
---
+++
@@ -21,7 +21,7 @@
print('Inserted card {0}'.format(card))
print('{0}'.format(subprocess.check_output(['pkcs15-tool',
'--list-certificates'])
- .rstrip()))
+ .rstrip()))
except removinator.SlotError:
print('Card {0} is not inserted'.format(card))
|
a10407bf4d9dd404d734985717aa7bcebfa0981d
|
api/digital_ocean.py
|
api/digital_ocean.py
|
"""
@fileoverview Digital Ocean API
@author David Parlevliet
@version 20130315
@preserve Copyright 2013 David Parlevliet.
Digital Ocean API
=================
Class to get the server details via the Digital Ocean API.
"""
import urllib2
import json
class Api():
group_name = "Digital Ocean"
client_key = None
api_key = None
servers = {}
def __init__(self, **kwargs):
for key in kwargs:
setattr(self, key, kwargs[key])
def grab_servers(self):
DROPLETS_URL = 'https%s/droplets/?client_id=%s&api_key=%s' % \
('://api.digitalocean.com',
self.client_key,
self.api_key)
droplets = urllib2.urlopen(DROPLETS_URL)
try:
data = json.loads(droplets.read())
except:
raise Exception("Fatal error: No droplets found")
for droplet in data['droplets']:
if droplet['status'] == 'active':
name = droplet['name']
if name not in self.servers:
self.servers[name] = []
self.servers[name].append(droplet['ip_address'])
def get_servers(self, name):
return self.servers[name] if name in self.servers else None
|
"""
@fileoverview Digital Ocean API
@author David Parlevliet
@version 20130315
@preserve Copyright 2013 David Parlevliet.
Digital Ocean API
=================
Class to get the server details via the Digital Ocean API.
"""
import urllib2
import json
class Api():
group_name = "Digital Ocean"
client_key = None
api_key = None
servers = {}
def __init__(self, **kwargs):
for key in kwargs:
setattr(self, key, kwargs[key])
def grab_servers(self):
DROPLETS_URL = 'https%s/droplets/?client_id=%s&api_key=%s' % \
('://api.digitalocean.com',
self.client_key,
self.api_key)
try:
droplets = urllib2.urlopen(DROPLETS_URL)
except urllib2.URLError:
raise Exception("Fatal error: Unable to connect to API")
try:
data = json.loads(droplets.read())
except:
raise Exception("Fatal error: No droplets found")
for droplet in data['droplets']:
if droplet['status'] == 'active':
name = droplet['name']
if name not in self.servers:
self.servers[name] = []
self.servers[name].append(droplet['ip_address'])
def get_servers(self, name):
return self.servers[name] if name in self.servers else None
|
Return a helpful exception if API is uncontactable
|
Return a helpful exception if API is uncontactable
|
Python
|
mit
|
dparlevliet/elastic-firewall,dparlevliet/elastic-firewall,dparlevliet/elastic-firewall
|
---
+++
@@ -27,7 +27,11 @@
self.client_key,
self.api_key)
- droplets = urllib2.urlopen(DROPLETS_URL)
+ try:
+ droplets = urllib2.urlopen(DROPLETS_URL)
+ except urllib2.URLError:
+ raise Exception("Fatal error: Unable to connect to API")
+
try:
data = json.loads(droplets.read())
except:
|
7ff6c9d85eef03c225b511f39bbb07796b47659f
|
datapipe/history.py
|
datapipe/history.py
|
class History:
def __init__(self):
self.conn = sqlite3.connect('.history.db')
|
import sqlite3
class History:
def __init__(self, path):
self.conn = sqlite3.connect(path)
|
Make database filepath configurable on History
|
Make database filepath configurable on History
|
Python
|
mit
|
ibab/datapipe
|
---
+++
@@ -1,5 +1,6 @@
+import sqlite3
class History:
- def __init__(self):
- self.conn = sqlite3.connect('.history.db')
+ def __init__(self, path):
+ self.conn = sqlite3.connect(path)
|
fd4539942dafe622d3f7a7d183db3d69f95a00c4
|
shop/urls/cart.py
|
shop/urls/cart.py
|
from django.conf.urls.defaults import url, patterns
from shop.views.cart import CartDetails, CartItemDetail
urlpatterns = patterns('',
url(r'^delete/$', CartDetails.as_view(action='delete'), # DELETE
name='cart_delete'),
url('^item/$', CartDetails.as_view(action='post'), # POST
name='cart_item_add'),
url(r'^$', CartDetails.as_view(), name='cart'), # GET
url(r'^update/$', CartDetails.as_view(action='put'),
name='cart_update'),
# CartItems
url('^item/(?P<id>[0-9A-Za-z-_.//]+)$', CartItemDetail.as_view(),
name='cart_item'),
url('^item/(?P<id>[0-9A-Za-z-_.//]+)/delete$',
CartItemDetail.as_view(action='delete'),
name='cart_item_delete'),
)
|
from django.conf.urls.defaults import url, patterns
from shop.views.cart import CartDetails, CartItemDetail
urlpatterns = patterns('',
url(r'^delete/$', CartDetails.as_view(action='delete'), # DELETE
name='cart_delete'),
url('^item/$', CartDetails.as_view(action='post'), # POST
name='cart_item_add'),
url(r'^$', CartDetails.as_view(), name='cart'), # GET
url(r'^update/$', CartDetails.as_view(action='put'),
name='cart_update'),
# CartItems
url('^item/(?P<id>[0-9]+)$', CartItemDetail.as_view(),
name='cart_item'),
url('^item/(?P<id>[0-9]+)/delete$',
CartItemDetail.as_view(action='delete'),
name='cart_item_delete'),
)
|
Make sure that ID will not match the first CartItems rule EVERY time ("//" was in regex).
|
Make sure that ID will not match the first CartItems rule EVERY time ("//" was in regex).
|
Python
|
bsd-3-clause
|
schacki/django-shop,khchine5/django-shop,khchine5/django-shop,dwx9/test,febsn/django-shop,DavideyLee/django-shop,awesto/django-shop,jrief/django-shop,dwx9/test,thenewguy/django-shop,thenewguy/django-shop,bmihelac/django-shop,pjdelport/django-shop,creimers/django-shop,creimers/django-shop,jrief/django-shop,bmihelac/django-shop,awesto/django-shop,awesto/django-shop,febsn/django-shop,febsn/django-shop,nimbis/django-shop,khchine5/django-shop,pjdelport/django-shop,rfleschenberg/django-shop,rfleschenberg/django-shop,dwx9/test,rfleschenberg/django-shop,fusionbox/django-shop,chriscauley/django-shop,jrief/django-shop,divio/django-shop,creimers/django-shop,DavideyLee/django-shop,pjdelport/django-shop,schacki/django-shop,schacki/django-shop,nimbis/django-shop,atheiste/django-shop,nimbis/django-shop,katomaso/django-shop,fusionbox/django-shop,chriscauley/django-shop,atheiste/django-shop,chriscauley/django-shop,jrutila/django-shop,jrutila/django-shop,khchine5/django-shop,schacki/django-shop,nimbis/django-shop,jrutila/django-shop,divio/django-shop,divio/django-shop,katomaso/django-shop,katomaso/django-shop,rfleschenberg/django-shop,atheiste/django-shop,jrief/django-shop
|
---
+++
@@ -13,9 +13,9 @@
name='cart_update'),
# CartItems
- url('^item/(?P<id>[0-9A-Za-z-_.//]+)$', CartItemDetail.as_view(),
+ url('^item/(?P<id>[0-9]+)$', CartItemDetail.as_view(),
name='cart_item'),
- url('^item/(?P<id>[0-9A-Za-z-_.//]+)/delete$',
+ url('^item/(?P<id>[0-9]+)/delete$',
CartItemDetail.as_view(action='delete'),
name='cart_item_delete'),
)
|
10948cd88d51383e13af0a116703984752092c6a
|
jenkinsapi_tests/systests/test_jenkins_matrix.py
|
jenkinsapi_tests/systests/test_jenkins_matrix.py
|
'''
System tests for `jenkinsapi.jenkins` module.
'''
import re
import time
import unittest
from jenkinsapi_tests.systests.base import BaseSystemTest
from jenkinsapi_tests.systests.job_configs import MATRIX_JOB
from jenkinsapi_tests.test_utils.random_strings import random_string
class TestMatrixJob(BaseSystemTest):
def test_invoke_matrix_job(self):
job_name = 'create_%s' % random_string()
job = self.jenkins.create_job(job_name, MATRIX_JOB)
job.invoke(block=True)
b = job.get_last_build()
while b.is_running():
time.sleep(1)
s = set()
for r in b.get_matrix_runs():
self.assertEquals(r.get_number(), b.get_number())
self.assertEquals(r.get_upstream_build(), b)
m = re.search(u'\xbb (.*) #\\d+$', r.name)
self.assertIsNotNone(m)
s.add(m.group(1))
# This is a bad test, it simply verifies that this function does
# not crash on a build from a matrix job.
self.assertFalse(b.get_master_job_name())
self.assertEqual(s, set(['one', 'two', 'three']))
if __name__ == '__main__':
unittest.main()
|
'''
System tests for `jenkinsapi.jenkins` module.
'''
import re
import time
import unittest
from jenkinsapi_tests.systests.base import BaseSystemTest
from jenkinsapi_tests.systests.job_configs import MATRIX_JOB
from jenkinsapi_tests.test_utils.random_strings import random_string
class TestMatrixJob(BaseSystemTest):
def test_invoke_matrix_job(self):
job_name = 'create_%s' % random_string()
job = self.jenkins.create_job(job_name, MATRIX_JOB)
job.invoke(block=True)
build = job.get_last_build()
while build.is_running():
time.sleep(1)
set_of_groups = set()
for run in build.get_matrix_runs():
self.assertEquals(run.get_number(), build.get_number())
self.assertEquals(run.get_upstream_build(), build)
match_result = re.search(u'\xbb (.*) #\\d+$', run.name)
self.assertIsNotNone(match_result)
set_of_groups.add(match_result.group(1))
build.get_master_job_name()
# This is a bad test, it simply verifies that this function does
# not crash on a build from a matrix job.
self.assertFalse(build.get_master_job_name())
self.assertEqual(set_of_groups, set(['one', 'two', 'three']))
if __name__ == '__main__':
unittest.main()
|
Tidy up this test - still quite bad & useless.
|
Tidy up this test - still quite bad & useless.
|
Python
|
mit
|
imsardine/jenkinsapi,salimfadhley/jenkinsapi,JohnLZeller/jenkinsapi,JohnLZeller/jenkinsapi,aerickson/jenkinsapi,domenkozar/jenkinsapi,zaro0508/jenkinsapi,imsardine/jenkinsapi,zaro0508/jenkinsapi,jduan/jenkinsapi,mistermocha/jenkinsapi,domenkozar/jenkinsapi,salimfadhley/jenkinsapi,zaro0508/jenkinsapi,mistermocha/jenkinsapi,aerickson/jenkinsapi,jduan/jenkinsapi,JohnLZeller/jenkinsapi,mistermocha/jenkinsapi,imsardine/jenkinsapi
|
---
+++
@@ -17,24 +17,25 @@
job = self.jenkins.create_job(job_name, MATRIX_JOB)
job.invoke(block=True)
- b = job.get_last_build()
+ build = job.get_last_build()
- while b.is_running():
+ while build.is_running():
time.sleep(1)
- s = set()
- for r in b.get_matrix_runs():
- self.assertEquals(r.get_number(), b.get_number())
- self.assertEquals(r.get_upstream_build(), b)
- m = re.search(u'\xbb (.*) #\\d+$', r.name)
- self.assertIsNotNone(m)
- s.add(m.group(1))
+ set_of_groups = set()
+ for run in build.get_matrix_runs():
+ self.assertEquals(run.get_number(), build.get_number())
+ self.assertEquals(run.get_upstream_build(), build)
+ match_result = re.search(u'\xbb (.*) #\\d+$', run.name)
+ self.assertIsNotNone(match_result)
+ set_of_groups.add(match_result.group(1))
+ build.get_master_job_name()
# This is a bad test, it simply verifies that this function does
# not crash on a build from a matrix job.
- self.assertFalse(b.get_master_job_name())
+ self.assertFalse(build.get_master_job_name())
- self.assertEqual(s, set(['one', 'two', 'three']))
+ self.assertEqual(set_of_groups, set(['one', 'two', 'three']))
if __name__ == '__main__':
unittest.main()
|
238ba8cec34ec02dc521f25ef1ada6e230194c32
|
kitsune/kbadge/migrations/0002_auto_20181023_1319.py
|
kitsune/kbadge/migrations/0002_auto_20181023_1319.py
|
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import migrations
class Migration(migrations.Migration):
dependencies = [
('kbadge', '0001_initial'),
]
operations = [
migrations.RunSQL(
"UPDATE badger_badge SET image = CONCAT('uploads/', image)"
)
]
|
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import migrations
class Migration(migrations.Migration):
dependencies = [
('kbadge', '0001_initial'),
]
operations = [
migrations.RunSQL(
"UPDATE badger_badge SET image = CONCAT('uploads/', image) WHERE image NOT LIKE 'uploads/%'"
)
]
|
Add WHERE clause to SQL data migration.
|
Add WHERE clause to SQL data migration.
|
Python
|
bsd-3-clause
|
mozilla/kitsune,anushbmx/kitsune,anushbmx/kitsune,anushbmx/kitsune,mozilla/kitsune,mozilla/kitsune,mozilla/kitsune,anushbmx/kitsune
|
---
+++
@@ -12,6 +12,6 @@
operations = [
migrations.RunSQL(
- "UPDATE badger_badge SET image = CONCAT('uploads/', image)"
+ "UPDATE badger_badge SET image = CONCAT('uploads/', image) WHERE image NOT LIKE 'uploads/%'"
)
]
|
4651d3b5666fe3ddf3bd92b31ee6ffe4a72ce94e
|
core/api/__init__.py
|
core/api/__init__.py
|
import os
from flask import Flask, jsonify
from flask_pymongo import PyMongo, BSONObjectIdConverter
from werkzeug.exceptions import HTTPException, default_exceptions
from core.api import settings
def create_app(environment=None):
app = Flask('veritrans')
app.url_map.converters['ObjectId'] = BSONObjectIdConverter
# Config app for environment
if not environment:
environment = os.environ.get('BACKEND_ENVIRONMENT', 'Prod')
app.config.from_object('core.api.settings.%s' % environment)
# convert exceptions to JSON
def make_json_error(ex):
response = jsonify(
message=str(ex)
)
response.status_code = (ex.code
if isinstance(ex, HTTPException)
else 500)
return response
for code in default_exceptions.items():
app.error_handler_spec[None][code] = make_json_error
from core.api.views.endpoints import api
app.register_module(api)
return app
class API(object):
app = None
mongo_client = None
@staticmethod
def init():
env = os.environ.get('SITE_NAME', 'Dev')
API.app = create_app(env)
API.mongo_client = PyMongo(API.app)
|
import os
from flask import Flask, jsonify
from flask_pymongo import PyMongo, BSONObjectIdConverter
from werkzeug.exceptions import HTTPException, default_exceptions
from core.api import settings
def create_app(environment=None):
app = Flask('veritrans')
app.url_map.converters['ObjectId'] = BSONObjectIdConverter
# Config app for environment
if not environment:
environment = os.environ.get('BACKEND_ENVIRONMENT', 'Prod')
app.config.from_object('core.api.settings.%s' % environment)
# convert exceptions to JSON
def make_json_error(ex):
response = jsonify(
message=str(ex)
)
response.status_code = (ex.code
if isinstance(ex, HTTPException)
else 500)
return response
for code in default_exceptions.items():
app.error_handler_spec[None][code] = make_json_error
from core.api.views.endpoints import api
app.register_module(api)
return app
class API(object):
app = None
mongo_client = None
@staticmethod
def init():
API.app = create_app()
API.mongo_client = PyMongo(API.app)
|
Use Production config unless specified
|
Use Production config unless specified
|
Python
|
mit
|
onyb/veritrans-payment-portals
|
---
+++
@@ -42,6 +42,5 @@
@staticmethod
def init():
- env = os.environ.get('SITE_NAME', 'Dev')
- API.app = create_app(env)
+ API.app = create_app()
API.mongo_client = PyMongo(API.app)
|
ced218643784838d68961a926cc0dd18c3a3f01f
|
skald/geometry.py
|
skald/geometry.py
|
# -*- coding: utf-8 -*-
from collections import namedtuple
Size = namedtuple("Size", ["width", "height"])
Rectangle = namedtuple("Rectangle", ["x0", "y0", "x1", "y1"])
class Point(namedtuple("Point", ["x", "y"])):
"""Point in a two-dimensional space.
Named tuple implementation that allows for addition and subtraction.
"""
__slots__ = ()
def __add__(self, other):
x = self.x + other.x
y = self.y + other.y
return Point(x, y)
def __sub__(self, other):
x = self.x - other.x
y = self.y - other.y
return Point(x, y)
class Box(namedtuple("Box", ["point", "size"])):
__slots__ = ()
@property
def rectangle(self):
return Rectangle(
x0=self.point.x,
y0=self.point.y,
x1=self.point.x+self.size.width,
y1=self.point.y+self.size.height
)
|
# -*- coding: utf-8 -*-
from collections import namedtuple
Size = namedtuple("Size", ["width", "height"])
class Rectangle(namedtuple("Rectangle", ["x0", "y0", "x1", "y1"])):
def __contains__(self, other):
"""Check if this rectangle and `other` overlaps eachother.
Essentially this is a bit of a hack to be able to write
`rect1 in rect2`.
"""
if self.x0 < other.x0 and self.x1 > other.x1 and \
self.y0 < other.y0 and self.y1 > other.y1:
return True
return False
class Point(namedtuple("Point", ["x", "y"])):
"""Point in a two-dimensional space.
Named tuple implementation that allows for addition and subtraction.
"""
__slots__ = ()
def __add__(self, other):
x = self.x + other.x
y = self.y + other.y
return Point(x, y)
def __sub__(self, other):
x = self.x - other.x
y = self.y - other.y
return Point(x, y)
class Box(namedtuple("Box", ["point", "size"])):
__slots__ = ()
@property
def rectangle(self):
return Rectangle(
x0=self.point.x,
y0=self.point.y,
x1=self.point.x+self.size.width,
y1=self.point.y+self.size.height
)
|
Add intersection test for rectangles
|
Add intersection test for rectangles
|
Python
|
mit
|
bjornarg/skald,bjornarg/skald
|
---
+++
@@ -2,7 +2,19 @@
from collections import namedtuple
Size = namedtuple("Size", ["width", "height"])
-Rectangle = namedtuple("Rectangle", ["x0", "y0", "x1", "y1"])
+
+class Rectangle(namedtuple("Rectangle", ["x0", "y0", "x1", "y1"])):
+ def __contains__(self, other):
+ """Check if this rectangle and `other` overlaps eachother.
+
+ Essentially this is a bit of a hack to be able to write
+ `rect1 in rect2`.
+ """
+ if self.x0 < other.x0 and self.x1 > other.x1 and \
+ self.y0 < other.y0 and self.y1 > other.y1:
+ return True
+ return False
+
class Point(namedtuple("Point", ["x", "y"])):
"""Point in a two-dimensional space.
|
8f03f51c89aeea44943f9cb0b39330e676ae0089
|
utils.py
|
utils.py
|
import vx
from contextlib import contextmanager
from functools import partial
import sys
from io import StringIO
def _expose(f=None, name=None):
if f is None:
return partial(_expose, name=name)
if name is None:
name = f.__name__.lstrip('_')
if getattr(vx, name, None) is not None:
raise AttributeError("Cannot expose duplicate name: '{}'".format(name))
setattr(vx, name, f)
return f
vx.expose = _expose
@vx.expose
def _repeat(c, times=4):
for _ in range(times):
c()
@vx.expose
@contextmanager
def _cursor_wander(command=None, window=None):
if window is None:
window = vx.window.focused_window
y, x = vx.get_linecol_window(window)
if command is not None:
command()
yp, xp = vx.get_linecol_window(window)
yield (yp, xp)
vx.set_linecol_window(window, y, x)
@contextmanager
def stdoutIO(stdout=None):
old = sys.stdout
if stdout is None:
stdout = StringIO()
sys.stdout = stdout
yield stdout
sys.stdout = old
|
import vx
from contextlib import contextmanager
from functools import partial
import sys
from io import StringIO
def _expose(f=None, name=None):
if f is None:
return partial(_expose, name=name)
if name is None:
name = f.__name__.lstrip('_')
if getattr(vx, name, None) is not None:
raise AttributeError("Cannot expose duplicate name: '{}'".format(name))
setattr(vx, name, f)
return f
vx.expose = _expose
@vx.expose
def _repeat(c, times=4):
res = []
for _ in range(times):
res.append(c())
return res
@vx.expose
@contextmanager
def _cursor_wander(command=None, window=None):
if window is None:
window = vx.window.focused_window
y, x = vx.get_linecol_window(window)
if command is not None:
command()
yp, xp = vx.get_linecol_window(window)
yield (yp, xp)
vx.set_linecol_window(window, y, x)
@contextmanager
def stdoutIO(stdout=None):
old = sys.stdout
if stdout is None:
stdout = StringIO()
sys.stdout = stdout
yield stdout
sys.stdout = old
|
Change repeat command to return a list of the results of the repeated commands
|
Change repeat command to return a list of the results of the repeated commands
|
Python
|
mit
|
philipdexter/vx,philipdexter/vx
|
---
+++
@@ -19,8 +19,10 @@
@vx.expose
def _repeat(c, times=4):
+ res = []
for _ in range(times):
- c()
+ res.append(c())
+ return res
@vx.expose
@contextmanager
|
822e6123cc598b4f6a0eafedfb2f0d0cbfba5f37
|
currencies/migrations/0003_auto_20151216_1906.py
|
currencies/migrations/0003_auto_20151216_1906.py
|
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import migrations
from extra_countries.models import ExtraCountry
def add_currencies_with_countries(apps, schema_editor):
# We can't import the model directly as it may be a newer
# version than this migration expects. We use the historical version.
Currency = apps.get_model("currencies", "Currency")
for extra_country in ExtraCountry.objects.all():
print("seeding currency for county: %s" % extra_country.country.name)
# trying to find a currency with the same code first
try:
currency = Currency.objects.get(code=extra_country.country.currency)
except Currency.DoesNotExist: # no such currency yet
currency = Currency(code=extra_country.country.currency,
name=extra_country.country.currency_name)
currency.save()
currency.countries.add(extra_country.pk)
def reverse_data(apps, schema_editor):
Currency = apps.get_model("currencies", "Currency")
Currency.objects.all().delete()
class Migration(migrations.Migration):
dependencies = [
('currencies', '0002_currency_countries'),
]
operations = [
migrations.RunPython(add_currencies_with_countries, reverse_data)
]
|
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import migrations
from extra_countries.models import ExtraCountry
def add_currencies_with_countries(apps, schema_editor):
# We can't import the model directly as it may be a newer
# version than this migration expects. We use the historical version.
Currency = apps.get_model("currencies", "Currency")
for extra_country in ExtraCountry.objects.all():
print("seeding currency for county: %s" % extra_country.country.name)
# trying to find a currency with the same code first
try:
currency = Currency.objects.get(code=extra_country.country.currency)
except Currency.DoesNotExist: # no such currency yet
currency = Currency(code=extra_country.country.currency,
name=extra_country.country.currency_name)
if (str(extra_country.country.currency) == '') or (str(extra_country.country.currency_name) == ''):
pass
else:
currency.save()
currency.countries.add(extra_country.pk)
def reverse_data(apps, schema_editor):
Currency = apps.get_model("currencies", "Currency")
Currency.objects.all().delete()
class Migration(migrations.Migration):
dependencies = [
('currencies', '0002_currency_countries'),
]
operations = [
migrations.RunPython(add_currencies_with_countries, reverse_data)
]
|
Fix currencies seeding, so it won't have empty currencies
|
Fix currencies seeding, so it won't have empty currencies
|
Python
|
mit
|
openspending/cosmopolitan,kiote/cosmopolitan
|
---
+++
@@ -19,8 +19,11 @@
except Currency.DoesNotExist: # no such currency yet
currency = Currency(code=extra_country.country.currency,
name=extra_country.country.currency_name)
- currency.save()
- currency.countries.add(extra_country.pk)
+ if (str(extra_country.country.currency) == '') or (str(extra_country.country.currency_name) == ''):
+ pass
+ else:
+ currency.save()
+ currency.countries.add(extra_country.pk)
def reverse_data(apps, schema_editor):
Currency = apps.get_model("currencies", "Currency")
|
041b271baa7ae0bbd20c30ac4f70b42fda267e93
|
mozillians/groups/__init__.py
|
mozillians/groups/__init__.py
|
from django.apps import AppConfig
CIS_GROUPS = [
'cis_whitelist',
'nda'
]
default_app_config = 'mozillians.groups.GroupConfig'
class GroupConfig(AppConfig):
name = 'mozillians.groups'
|
from django.apps import AppConfig
CIS_GROUPS = [
'cis_whitelist',
'nda',
'open-innovation-reps-council'
]
default_app_config = 'mozillians.groups.GroupConfig'
class GroupConfig(AppConfig):
name = 'mozillians.groups'
|
Add a group in the whitelist.
|
Add a group in the whitelist.
|
Python
|
bsd-3-clause
|
mozilla/mozillians,akatsoulas/mozillians,mozilla/mozillians,johngian/mozillians,mozilla/mozillians,mozilla/mozillians,akatsoulas/mozillians,akatsoulas/mozillians,johngian/mozillians,johngian/mozillians,johngian/mozillians,akatsoulas/mozillians
|
---
+++
@@ -3,7 +3,8 @@
CIS_GROUPS = [
'cis_whitelist',
- 'nda'
+ 'nda',
+ 'open-innovation-reps-council'
]
|
199c9bae8e2ad42ee1c8699c678dd56d6074b2de
|
main/models.py
|
main/models.py
|
from django.db import models
from django.contrib.auth.models import User
import string, random
from django import forms
# Create your models here.
def _generate_default_hashtag():
return "".join(random.choice(string.lowercase) for i in range(3))
class Wall(models.Model):
hashtag = models.CharField(max_length=20, help_text='Twitter hashtag to tweet to', default=_generate_default_hashtag())
user = models.ForeignKey(User, editable=False)
sms_keyword = models.CharField(max_length=20)
def __unicode__(self):
return self.name
TRAFFIC_SOURCE = (
('BG', 'Blog'),
('FR', 'Friend'),
('OT', 'Other',)
)
|
from django.db import models
from django.contrib.auth.models import User
import string, random
from django import forms
# Create your models here.
def _generate_default_hashtag():
return "".join(random.choice(string.lowercase) for i in range(3))
class Wall(models.Model):
hashtag = models.CharField(max_length=20, help_text='Twitter hashtag to tweet to', default=_generate_default_hashtag())
user = models.ForeignKey(User, editable=False)
sms_keyword = models.CharField(max_length=20)
def __unicode__(self):
return self.sms_keyword
TRAFFIC_SOURCE = (
('BG', 'Blog'),
('FR', 'Friend'),
('OT', 'Other',)
)
|
Return sms_keyword as wall name
|
Return sms_keyword as wall name
|
Python
|
mit
|
Aaron1011/texting_wall
|
---
+++
@@ -10,10 +10,12 @@
user = models.ForeignKey(User, editable=False)
sms_keyword = models.CharField(max_length=20)
def __unicode__(self):
- return self.name
-
+ return self.sms_keyword
+
TRAFFIC_SOURCE = (
('BG', 'Blog'),
('FR', 'Friend'),
('OT', 'Other',)
)
+
+
|
523ee893118413caa45e66869e9380f5e52f3402
|
src/parser/python/print_libpython.py
|
src/parser/python/print_libpython.py
|
# Print embeddable python library, as well as other libs it requires.
# Will prefer static linkage unless invoked with "shared" argument.
# JGG, 8/4/04
import sys, distutils.sysconfig
static_link = 1
nargs = len(sys.argv)
if nargs == 2 and sys.argv[1] == "shared":
static_link = 0
# Note that this adds libraries we've certainly already linked to.
libs = distutils.sysconfig.get_config_var("LIBS")
libs += " " + distutils.sysconfig.get_config_var("SYSLIBS")
if static_link:
prefix = distutils.sysconfig.get_config_var("LIBPL")
pythonlib = distutils.sysconfig.get_config_var("BLDLIBRARY")
if len(pythonlib) > 0:
print prefix + '/' + pythonlib, libs
sys.exit(0)
# else try shared linkage
linkshared = distutils.sysconfig.get_config_vars("LINKFORSHARED")[0]
# FIXME: Will this sanity test work for all platforms??
# NB: sys.platform can help us if we need to test for platform
if linkshared.find("ython") != -1:
print linkshared, libs
sys.exit(0)
print >> sys.stderr, "***ERROR: Can't find a python to embed."
sys.exit(1)
|
# Print embeddable python library, as well as other libs it requires.
# Will prefer static linkage unless invoked with "shared" argument.
# JGG, 8/4/04
import sys, distutils.sysconfig
static_link = 1
nargs = len(sys.argv)
if nargs == 2 and sys.argv[1] == "shared":
static_link = 0
# Note that this adds libraries we've certainly already linked to.
libs = distutils.sysconfig.get_config_var("LIBS")
libs += " " + distutils.sysconfig.get_config_var("SYSLIBS")
if static_link:
prefix = distutils.sysconfig.get_config_var("LIBPL")
pythonlib = distutils.sysconfig.get_config_var("BLDLIBRARY")
if len(pythonlib) > 0:
import os.path
plib = prefix + '/' + pythonlib
# Must see if file exists, because it doesn't in Jaguar!
if os.path.exists(plib):
print plib, libs
sys.exit(0)
# else try shared linkage
linkshared = distutils.sysconfig.get_config_vars("LINKFORSHARED")[0]
# FIXME: Will this sanity test work for all platforms??
# NB: sys.platform can help us if we need to test for platform
if linkshared.find("ython") != -1:
print linkshared, libs
sys.exit(0)
print >> sys.stderr, "***ERROR: Can't find a python to embed."
sys.exit(1)
|
Test file existence for brain-dead Jaguar.
|
Test file existence for brain-dead Jaguar.
|
Python
|
apache-2.0
|
RTcmix/RTcmix,RTcmix/RTcmix,RTcmix/RTcmix,RTcmix/RTcmix,RTcmix/RTcmix,RTcmix/RTcmix
|
---
+++
@@ -17,8 +17,12 @@
prefix = distutils.sysconfig.get_config_var("LIBPL")
pythonlib = distutils.sysconfig.get_config_var("BLDLIBRARY")
if len(pythonlib) > 0:
- print prefix + '/' + pythonlib, libs
- sys.exit(0)
+ import os.path
+ plib = prefix + '/' + pythonlib
+ # Must see if file exists, because it doesn't in Jaguar!
+ if os.path.exists(plib):
+ print plib, libs
+ sys.exit(0)
# else try shared linkage
linkshared = distutils.sysconfig.get_config_vars("LINKFORSHARED")[0]
|
b6ec3ba9efae7b6b291391b0333e80f2e9fc6fa0
|
src/waldur_mastermind/invoices/migrations/0053_invoiceitem_uuid.py
|
src/waldur_mastermind/invoices/migrations/0053_invoiceitem_uuid.py
|
import uuid
from django.db import migrations
import waldur_core.core.fields
def gen_uuid(apps, schema_editor):
InvoiceItem = apps.get_model('invoices', 'InvoiceItem')
for row in InvoiceItem.objects.all():
row.uuid = uuid.uuid4().hex
row.save(update_fields=['uuid'])
class Migration(migrations.Migration):
dependencies = [
('invoices', '0052_delete_servicedowntime'),
]
operations = [
migrations.AddField(
model_name='invoiceitem',
name='uuid',
field=waldur_core.core.fields.UUIDField(null=True),
),
migrations.RunPython(gen_uuid, elidable=True),
migrations.AlterField(
model_name='invoiceitem',
name='uuid',
field=waldur_core.core.fields.UUIDField(),
),
]
|
import uuid
from django.db import migrations, models
import waldur_core.core.fields
def gen_uuid(apps, schema_editor):
InvoiceItem = apps.get_model('invoices', 'InvoiceItem')
for row in InvoiceItem.objects.all():
row.uuid = uuid.uuid4().hex
row.save(update_fields=['uuid'])
class Migration(migrations.Migration):
dependencies = [
('invoices', '0052_delete_servicedowntime'),
]
operations = [
migrations.AddField(
model_name='invoiceitem', name='uuid', field=models.UUIDField(null=True),
),
migrations.RunPython(gen_uuid, elidable=True),
migrations.AlterField(
model_name='invoiceitem',
name='uuid',
field=waldur_core.core.fields.UUIDField(),
),
]
|
Fix database migration script for UUID field in invoice item model.
|
Fix database migration script for UUID field in invoice item model.
|
Python
|
mit
|
opennode/waldur-mastermind,opennode/waldur-mastermind,opennode/waldur-mastermind,opennode/waldur-mastermind
|
---
+++
@@ -1,6 +1,6 @@
import uuid
-from django.db import migrations
+from django.db import migrations, models
import waldur_core.core.fields
@@ -20,9 +20,7 @@
operations = [
migrations.AddField(
- model_name='invoiceitem',
- name='uuid',
- field=waldur_core.core.fields.UUIDField(null=True),
+ model_name='invoiceitem', name='uuid', field=models.UUIDField(null=True),
),
migrations.RunPython(gen_uuid, elidable=True),
migrations.AlterField(
|
9b19d366c7e1cf41ffc6af4eaed789995ddc5cc2
|
byceps/blueprints/core_admin/views.py
|
byceps/blueprints/core_admin/views.py
|
"""
byceps.blueprints.core_admin.views
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
:Copyright: 2006-2018 Jochen Kupperschmidt
:License: Modified BSD, see LICENSE for details.
"""
from ...services.brand import service as brand_service
from ...util.framework.blueprint import create_blueprint
from ..authorization.registry import permission_registry
from .authorization import AdminPermission
blueprint = create_blueprint('core_admin', __name__)
permission_registry.register_enum(AdminPermission)
@blueprint.app_context_processor
def inject_brands():
brands = brand_service.get_brands()
return {
'all_brands': brands,
}
|
"""
byceps.blueprints.core_admin.views
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
:Copyright: 2006-2018 Jochen Kupperschmidt
:License: Modified BSD, see LICENSE for details.
"""
from ...services.brand import service as brand_service
from ...util.framework.blueprint import create_blueprint
from ..authorization.registry import permission_registry
from .authorization import AdminPermission
blueprint = create_blueprint('core_admin', __name__)
permission_registry.register_enum(AdminPermission)
@blueprint.app_context_processor
def inject_template_variables():
brands = brand_service.get_brands()
return {
'all_brands': brands,
}
|
Generalize name of function to inject admin template variables
|
Generalize name of function to inject admin template variables
|
Python
|
bsd-3-clause
|
homeworkprod/byceps,m-ober/byceps,m-ober/byceps,homeworkprod/byceps,homeworkprod/byceps,m-ober/byceps
|
---
+++
@@ -21,7 +21,7 @@
@blueprint.app_context_processor
-def inject_brands():
+def inject_template_variables():
brands = brand_service.get_brands()
return {
|
f3bb89a58375ac91efb1fa6f3426c0a7e7dd85d5
|
tests/fixtures/source.py
|
tests/fixtures/source.py
|
#!/usr/bin/env python
class Person:
def __init__(self, first_name, last_name):
self.first_name = first_name
self.last_name = last_name
def fullname(self):
return "%s %s" % (self.first_name, self.last_name)
def not_called(self):
print("Shouldn't be called")
person = Person("Marty", "McFly")
person.fullname()
|
#!/usr/bin/env python
class Person:
def __init__(self, first_name, last_name):
self.first_name = first_name
self.last_name = last_name
def fullname(self):
return "%s %s" % (self.first_name, self.last_name)
def not_called(self):
print("Shouldn't be called")
person = Person("Marty", "McFly")
person.fullname()
|
Add BOM signature to test fixture
|
Add BOM signature to test fixture
|
Python
|
mit
|
codeclimate/python-test-reporter,codeclimate/python-test-reporter
|
---
+++
@@ -1,4 +1,4 @@
-#!/usr/bin/env python
+#!/usr/bin/env python
class Person:
|
6ac683ca1905fbf17dd63c1264609e770439fa7f
|
test/integration/targets/module_utils/library/test_env_override.py
|
test/integration/targets/module_utils/library/test_env_override.py
|
#!/usr/bin/python
from ansible.module_utils.basic import AnsibleModule
from ansible.module_utils.json_utils import data
from ansible.module_utils.mork import data as mork_data
results = {"json_utils": data, "mork": mork_data}
AnsibleModule(argument_spec=dict()).exit_json(**results)
|
#!/usr/bin/python
# Most of these names are only available via PluginLoader so pylint doesn't
# know they exist
# pylint: disable=no-name-in-module
from ansible.module_utils.basic import AnsibleModule
from ansible.module_utils.json_utils import data
from ansible.module_utils.mork import data as mork_data
results = {"json_utils": data, "mork": mork_data}
AnsibleModule(argument_spec=dict()).exit_json(**results)
|
Disable pylint check for names existing in modules for test data
|
Disable pylint check for names existing in modules for test data
This test data imports from modules which are only available via
PluginLoader for this test case. So pylint doesn't know anything about
them
|
Python
|
mit
|
thaim/ansible,thaim/ansible
|
---
+++
@@ -1,4 +1,7 @@
#!/usr/bin/python
+# Most of these names are only available via PluginLoader so pylint doesn't
+# know they exist
+# pylint: disable=no-name-in-module
from ansible.module_utils.basic import AnsibleModule
from ansible.module_utils.json_utils import data
from ansible.module_utils.mork import data as mork_data
|
d4acff165712281cb7205cca1c39e7eaf1fdd23b
|
base/broadcast.py
|
base/broadcast.py
|
from abc import ABCMeta, abstractmethod
from enum import Enum
class Broadcast(metaclass=ABCMeta):
"""
An interface for defining a broadcast protocol.
The 'propose' and 'decide' methods need to be defined
"""
class MessageType(Enum):
SEND = 1
ECHO = 2
READY = 3
def __init__(self, node_number, faulty_nodes):
self.N = node_number
self.f = faulty_nodes
@abstractmethod
def broadcast(self, message):
pass
@abstractmethod
def broadcast_listener(self):
pass
@abstractmethod
def deliver(self, sender, message):
pass
|
Add abstract Broadcast protocol class
|
Add abstract Broadcast protocol class
|
Python
|
mit
|
koevskinikola/ByzantineRandomizedConsensus
|
---
+++
@@ -0,0 +1,30 @@
+from abc import ABCMeta, abstractmethod
+from enum import Enum
+
+
+class Broadcast(metaclass=ABCMeta):
+ """
+ An interface for defining a broadcast protocol.
+ The 'propose' and 'decide' methods need to be defined
+ """
+
+ class MessageType(Enum):
+ SEND = 1
+ ECHO = 2
+ READY = 3
+
+ def __init__(self, node_number, faulty_nodes):
+ self.N = node_number
+ self.f = faulty_nodes
+
+ @abstractmethod
+ def broadcast(self, message):
+ pass
+
+ @abstractmethod
+ def broadcast_listener(self):
+ pass
+
+ @abstractmethod
+ def deliver(self, sender, message):
+ pass
|
|
7872abf00b24a504fccba576b13ecdd140e0135f
|
pybb/read_tracking.py
|
pybb/read_tracking.py
|
def update_read_tracking(topic, user):
tracking = user.readtracking
#if last_read > last_read - don't check topics
if tracking.last_read and tracking.last_read > (topic.last_post.updated or
topic.last_post.created):
return
if isinstance(tracking.topics, dict):
#clear topics if len > 5Kb and set last_read to current time
if len(tracking.topics) > 5120:
tracking.topics = None
tracking.last_read = datetime.now()
tracking.save()
#update topics if exist new post or does't exist in dict
if topic.last_post.pk > tracking.topics.get(str(topic.pk), 0):
tracking.topics.setdefault(str(topic.pk), topic.last_post.pk)
tracking.save()
else:
#initialize topic tracking dict
tracking.topics = {topic.pk: topic.last_post.pk}
tracking.save()
|
def update_read_tracking(topic, user):
tracking = user.readtracking
#if last_read > last_read - don't check topics
if tracking.last_read and tracking.last_read > (topic.last_post.updated or
topic.last_post.created):
return
if isinstance(tracking.topics, dict):
#clear topics if len > 5Kb and set last_read to current time
if len(tracking.topics) > 5120:
tracking.topics = None
tracking.last_read = datetime.now()
tracking.save()
#update topics if new post exists or cache entry is empty
if topic.last_post.pk > tracking.topics.get(str(topic.pk), 0):
tracking.topics[str(topic.pk)] = topic.last_post.pk
tracking.save()
else:
#initialize topic tracking dict
tracking.topics = {topic.pk: topic.last_post.pk}
tracking.save()
|
Fix bug in read tracking system
|
Fix bug in read tracking system
|
Python
|
bsd-2-clause
|
ttyS15/pybbm,onecue/pybbm,katsko/pybbm,katsko/pybbm,wengole/pybbm,wengole/pybbm,webu/pybbm,acamposruiz/quecoins,springmerchant/pybbm,NEERAJIITKGP/pybbm,webu/pybbm,concentricsky/pybbm,skolsuper/pybbm,hovel/pybbm,NEERAJIITKGP/pybbm,hovel/pybbm,webu/pybbm,artfinder/pybbm,onecue/pybbm,katsko/pybbm,ttyS15/pybbm,wengole/pybbm,jonsimington/pybbm,jonsimington/pybbm,ttyS15/pybbm,skolsuper/pybbm,just-work/pybbm,zekone/dj_pybb,onecue/pybbm,hovel/pybbm,just-work/pybbm,acamposruiz/quecoins,acamposruiz/quecoins,springmerchant/pybbm,springmerchant/pybbm,skolsuper/pybbm,artfinder/pybbm,zekone/dj_pybb,concentricsky/pybbm,DylannCordel/pybbm,just-work/pybbm,zekone/dj_pybb,DylannCordel/pybbm,jonsimington/pybbm,artfinder/pybbm,concentricsky/pybbm,DylannCordel/pybbm,NEERAJIITKGP/pybbm
|
---
+++
@@ -12,9 +12,9 @@
tracking.topics = None
tracking.last_read = datetime.now()
tracking.save()
- #update topics if exist new post or does't exist in dict
+ #update topics if new post exists or cache entry is empty
if topic.last_post.pk > tracking.topics.get(str(topic.pk), 0):
- tracking.topics.setdefault(str(topic.pk), topic.last_post.pk)
+ tracking.topics[str(topic.pk)] = topic.last_post.pk
tracking.save()
else:
#initialize topic tracking dict
|
346ffdb3e3836e2931f838a6dd929a325da0d5e6
|
tests/test_arithmetic.py
|
tests/test_arithmetic.py
|
from intervals import Interval
class TestArithmeticOperators(object):
def test_add_operator(self):
assert Interval(1, 2) + Interval(1, 2) == Interval(2, 4)
def test_sub_operator(self):
assert Interval(1, 3) - Interval(1, 2) == Interval(-1, 2)
def test_isub_operator(self):
range_ = Interval(1, 3)
range_ -= Interval(1, 2)
assert range_ == Interval(-1, 2)
def test_iadd_operator(self):
range_ = Interval(1, 2)
range_ += Interval(1, 2)
assert range_ == Interval(2, 4)
|
from pytest import mark
from intervals import Interval
class TestArithmeticOperators(object):
def test_add_operator(self):
assert Interval(1, 2) + Interval(1, 2) == Interval(2, 4)
def test_sub_operator(self):
assert Interval(1, 3) - Interval(1, 2) == Interval(-1, 2)
def test_isub_operator(self):
range_ = Interval(1, 3)
range_ -= Interval(1, 2)
assert range_ == Interval(-1, 2)
def test_iadd_operator(self):
range_ = Interval(1, 2)
range_ += Interval(1, 2)
assert range_ == Interval(2, 4)
@mark.parametrize(('first', 'second', 'intersection'), (
('[1, 5]', '[2, 9]', '[2, 5]'),
('[3, 4]', '[3, 9]', '[3, 4]'),
('(3, 6]', '[2, 6)', '(3, 6)')
))
def test_intersection(self, first, second, intersection):
Interval(first) & Interval(second) == Interval(intersection)
|
Add some tests for intersection
|
Add some tests for intersection
|
Python
|
bsd-3-clause
|
kvesteri/intervals
|
---
+++
@@ -1,3 +1,4 @@
+from pytest import mark
from intervals import Interval
@@ -17,3 +18,11 @@
range_ = Interval(1, 2)
range_ += Interval(1, 2)
assert range_ == Interval(2, 4)
+
+ @mark.parametrize(('first', 'second', 'intersection'), (
+ ('[1, 5]', '[2, 9]', '[2, 5]'),
+ ('[3, 4]', '[3, 9]', '[3, 4]'),
+ ('(3, 6]', '[2, 6)', '(3, 6)')
+ ))
+ def test_intersection(self, first, second, intersection):
+ Interval(first) & Interval(second) == Interval(intersection)
|
2a0c8bdb74404556483cd60e7cb2f8e25559e500
|
pylearn2/scripts/tutorials/grbm_smd/test_grbm_smd.py
|
pylearn2/scripts/tutorials/grbm_smd/test_grbm_smd.py
|
import pylearn2
from pylearn2.utils.serial import load_train_file
import os
from pylearn2.testing import no_debug_mode
from theano import config
@no_debug_mode
def test_train_example():
""" tests that the grbm_smd example script runs correctly """
assert config.mode != "DEBUG_MODE"
path = pylearn2.__path__[0]
train_example_path = os.path.join(path, 'scripts', 'tutorials', 'grbm_smd')
train_yaml_path = os.path.join(train_example_path, 'cifar_grbm_smd.yaml')
train_object = load_train_file(train_yaml_path)
#make the termination criterion really lax so the test won't run for long
train_object.algorithm.termination_criterion.prop_decrease = 0.5
train_object.algorithm.termination_criterion.N = 1
train_object.main_loop()
if __name__ == '__main__':
test_train_example()
|
import pylearn2
from pylearn2.utils.serial import load_train_file
import os
from pylearn2.testing import no_debug_mode
from theano import config
@no_debug_mode
def test_train_example():
""" tests that the grbm_smd example script runs correctly """
assert config.mode != "DEBUG_MODE"
path = pylearn2.__path__[0]
train_example_path = os.path.join(path, 'scripts', 'tutorials', 'grbm_smd')
cwd = os.getcwd()
try:
os.chdir(train_example_path)
train_yaml_path = os.path.join(train_example_path, 'cifar_grbm_smd.yaml')
train_object = load_train_file(train_yaml_path)
#make the termination criterion really lax so the test won't run for long
train_object.algorithm.termination_criterion.prop_decrease = 0.5
train_object.algorithm.termination_criterion.N = 1
train_object.main_loop()
finally:
os.chdir(cwd)
if __name__ == '__main__':
test_train_example()
|
Fix failing test through loading of examples from $PWD.
|
Fix failing test through loading of examples from $PWD.
|
Python
|
bsd-3-clause
|
aalmah/pylearn2,TNick/pylearn2,shiquanwang/pylearn2,alexjc/pylearn2,fyffyt/pylearn2,bartvm/pylearn2,pkainz/pylearn2,lancezlin/pylearn2,pkainz/pylearn2,CIFASIS/pylearn2,daemonmaker/pylearn2,woozzu/pylearn2,ddboline/pylearn2,caidongyun/pylearn2,shiquanwang/pylearn2,mclaughlin6464/pylearn2,chrish42/pylearn,theoryno3/pylearn2,ddboline/pylearn2,Refefer/pylearn2,abergeron/pylearn2,skearnes/pylearn2,aalmah/pylearn2,lisa-lab/pylearn2,chrish42/pylearn,alexjc/pylearn2,JesseLivezey/plankton,alexjc/pylearn2,TNick/pylearn2,woozzu/pylearn2,fulmicoton/pylearn2,aalmah/pylearn2,matrogers/pylearn2,mclaughlin6464/pylearn2,fishcorn/pylearn2,hyqneuron/pylearn2-maxsom,junbochen/pylearn2,shiquanwang/pylearn2,CIFASIS/pylearn2,matrogers/pylearn2,bartvm/pylearn2,theoryno3/pylearn2,skearnes/pylearn2,fyffyt/pylearn2,se4u/pylearn2,pombredanne/pylearn2,fulmicoton/pylearn2,nouiz/pylearn2,ddboline/pylearn2,lisa-lab/pylearn2,fyffyt/pylearn2,skearnes/pylearn2,jamessergeant/pylearn2,junbochen/pylearn2,alexjc/pylearn2,w1kke/pylearn2,w1kke/pylearn2,daemonmaker/pylearn2,sandeepkbhat/pylearn2,daemonmaker/pylearn2,lancezlin/pylearn2,fishcorn/pylearn2,bartvm/pylearn2,TNick/pylearn2,sandeepkbhat/pylearn2,JesseLivezey/pylearn2,ashhher3/pylearn2,lancezlin/pylearn2,sandeepkbhat/pylearn2,matrogers/pylearn2,shiquanwang/pylearn2,JesseLivezey/plankton,kastnerkyle/pylearn2,TNick/pylearn2,CIFASIS/pylearn2,woozzu/pylearn2,theoryno3/pylearn2,KennethPierce/pylearnk,fulmicoton/pylearn2,jamessergeant/pylearn2,junbochen/pylearn2,daemonmaker/pylearn2,chrish42/pylearn,hantek/pylearn2,kastnerkyle/pylearn2,nouiz/pylearn2,mkraemer67/pylearn2,lisa-lab/pylearn2,lamblin/pylearn2,w1kke/pylearn2,Refefer/pylearn2,kose-y/pylearn2,w1kke/pylearn2,ddboline/pylearn2,goodfeli/pylearn2,jeremyfix/pylearn2,sandeepkbhat/pylearn2,hantek/pylearn2,kose-y/pylearn2,KennethPierce/pylearnk,se4u/pylearn2,JesseLivezey/pylearn2,lancezlin/pylearn2,ashhher3/pylearn2,goodfeli/pylearn2,lisa-lab/pylearn2,lunyang/pylearn2,theoryno3/pylearn2,hyqneuron/pylearn2-maxsom,ashhher3/pylearn2,se4u/pylearn2,JesseLivezey/pylearn2,lunyang/pylearn2,hantek/pylearn2,kose-y/pylearn2,fishcorn/pylearn2,goodfeli/pylearn2,matrogers/pylearn2,cosmoharrigan/pylearn2,lamblin/pylearn2,cosmoharrigan/pylearn2,pombredanne/pylearn2,hyqneuron/pylearn2-maxsom,mkraemer67/pylearn2,fyffyt/pylearn2,nouiz/pylearn2,msingh172/pylearn2,JesseLivezey/plankton,lunyang/pylearn2,ashhher3/pylearn2,Refefer/pylearn2,caidongyun/pylearn2,pombredanne/pylearn2,CIFASIS/pylearn2,chrish42/pylearn,caidongyun/pylearn2,abergeron/pylearn2,msingh172/pylearn2,kastnerkyle/pylearn2,mkraemer67/pylearn2,mclaughlin6464/pylearn2,skearnes/pylearn2,hantek/pylearn2,pombredanne/pylearn2,msingh172/pylearn2,JesseLivezey/plankton,kastnerkyle/pylearn2,junbochen/pylearn2,KennethPierce/pylearnk,lamblin/pylearn2,bartvm/pylearn2,fulmicoton/pylearn2,pkainz/pylearn2,KennethPierce/pylearnk,pkainz/pylearn2,fishcorn/pylearn2,Refefer/pylearn2,jeremyfix/pylearn2,nouiz/pylearn2,cosmoharrigan/pylearn2,jeremyfix/pylearn2,caidongyun/pylearn2,msingh172/pylearn2,mkraemer67/pylearn2,jamessergeant/pylearn2,hyqneuron/pylearn2-maxsom,mclaughlin6464/pylearn2,cosmoharrigan/pylearn2,jamessergeant/pylearn2,goodfeli/pylearn2,JesseLivezey/pylearn2,woozzu/pylearn2,abergeron/pylearn2,se4u/pylearn2,aalmah/pylearn2,lamblin/pylearn2,kose-y/pylearn2,jeremyfix/pylearn2,abergeron/pylearn2,lunyang/pylearn2
|
---
+++
@@ -11,14 +11,19 @@
assert config.mode != "DEBUG_MODE"
path = pylearn2.__path__[0]
train_example_path = os.path.join(path, 'scripts', 'tutorials', 'grbm_smd')
- train_yaml_path = os.path.join(train_example_path, 'cifar_grbm_smd.yaml')
- train_object = load_train_file(train_yaml_path)
+ cwd = os.getcwd()
+ try:
+ os.chdir(train_example_path)
+ train_yaml_path = os.path.join(train_example_path, 'cifar_grbm_smd.yaml')
+ train_object = load_train_file(train_yaml_path)
- #make the termination criterion really lax so the test won't run for long
- train_object.algorithm.termination_criterion.prop_decrease = 0.5
- train_object.algorithm.termination_criterion.N = 1
+ #make the termination criterion really lax so the test won't run for long
+ train_object.algorithm.termination_criterion.prop_decrease = 0.5
+ train_object.algorithm.termination_criterion.N = 1
- train_object.main_loop()
+ train_object.main_loop()
+ finally:
+ os.chdir(cwd)
if __name__ == '__main__':
test_train_example()
|
4cfd8771b91c7c2b9f28ca4b9776e9770683093b
|
frigg/builds/admin.py
|
frigg/builds/admin.py
|
# -*- coding: utf8 -*-
from django.contrib import admin
from .models import Build, BuildResult, Project
class BuildResultInline(admin.StackedInline):
model = BuildResult
readonly_fields = ('result_log', 'succeeded', 'return_code')
extra = 0
max_num = 0
class BuildInline(admin.TabularInline):
model = Build
readonly_fields = ('build_number', 'branch', 'color', 'pull_request_id', 'sha')
extra = 0
max_num = 0
@admin.register(Project)
class ProjectAdmin(admin.ModelAdmin):
list_display = ('__str__', 'git_repository', 'average_time', 'last_build_number')
inlines = [BuildInline]
list_filter = ['owner']
@admin.register(Build)
class BuildAdmin(admin.ModelAdmin):
list_display = ('build_number', 'project', 'branch', 'pull_request_id', 'sha', 'color')
inlines = [BuildResultInline]
list_filter = ['project']
@admin.register(BuildResult)
class BuildResultAdmin(admin.ModelAdmin):
list_display = ('__str__', 'succeeded', 'return_code', 'coverage')
|
# -*- coding: utf8 -*-
from django.contrib import admin
from django.template.defaultfilters import pluralize
from .models import Build, BuildResult, Project
class BuildResultInline(admin.StackedInline):
model = BuildResult
readonly_fields = ('result_log', 'succeeded', 'return_code')
extra = 0
max_num = 0
class BuildInline(admin.TabularInline):
model = Build
readonly_fields = ('build_number', 'branch', 'color', 'pull_request_id', 'sha')
extra = 0
max_num = 0
@admin.register(Project)
class ProjectAdmin(admin.ModelAdmin):
list_display = ('__str__', 'git_repository', 'average_time', 'last_build_number')
inlines = [BuildInline]
list_filter = ['owner']
@admin.register(Build)
class BuildAdmin(admin.ModelAdmin):
list_display = ('build_number', 'project', 'branch', 'pull_request_id', 'sha', 'color')
inlines = [BuildResultInline]
list_filter = ['project']
actions = ['restart_build']
def restart_build(self, request, queryset):
for build in queryset:
build.start()
self.message_user(
request,
'{} build{} was restarted'.format(len(queryset), pluralize(len(queryset)))
)
restart_build.short_description = 'Restart selected builds'
@admin.register(BuildResult)
class BuildResultAdmin(admin.ModelAdmin):
list_display = ('__str__', 'succeeded', 'return_code', 'coverage')
|
Add restart_build action to BuildAdmin
|
Add restart_build action to BuildAdmin
|
Python
|
mit
|
frigg/frigg-hq,frigg/frigg-hq,frigg/frigg-hq
|
---
+++
@@ -1,5 +1,6 @@
# -*- coding: utf8 -*-
from django.contrib import admin
+from django.template.defaultfilters import pluralize
from .models import Build, BuildResult, Project
@@ -30,6 +31,18 @@
list_display = ('build_number', 'project', 'branch', 'pull_request_id', 'sha', 'color')
inlines = [BuildResultInline]
list_filter = ['project']
+ actions = ['restart_build']
+
+ def restart_build(self, request, queryset):
+ for build in queryset:
+ build.start()
+
+ self.message_user(
+ request,
+ '{} build{} was restarted'.format(len(queryset), pluralize(len(queryset)))
+ )
+
+ restart_build.short_description = 'Restart selected builds'
@admin.register(BuildResult)
|
b1b1392d2f268a5c74fd21c826a3ea6387567cab
|
froide/bounce/apps.py
|
froide/bounce/apps.py
|
from django.apps import AppConfig
from django.utils.translation import ugettext_lazy as _
class BounceConfig(AppConfig):
name = 'froide.bounce'
verbose_name = _('Bounce')
def ready(self):
from froide.account import account_canceled
account_canceled.connect(cancel_user)
def cancel_user(sender, user=None, **kwargs):
from .models import Bounce
if user is None:
return
Bounce.objects.filter(user=user).delete()
|
import json
from django.apps import AppConfig
from django.utils.translation import ugettext_lazy as _
class BounceConfig(AppConfig):
name = 'froide.bounce'
verbose_name = _('Bounce')
def ready(self):
from froide.account import account_canceled
from froide.account.export import registry
account_canceled.connect(cancel_user)
registry.register(export_user_data)
def cancel_user(sender, user=None, **kwargs):
from .models import Bounce
if user is None:
return
Bounce.objects.filter(user=user).delete()
def export_user_data(user):
from .models import Bounce
bounces = Bounce.objects.filter(user=user)
if not bounces:
return
yield ('bounces.json', json.dumps([
{
'last_update': (
b.last_update.isoformat() if b.last_update else None
),
'bounces': b.bounces,
'email': b.email,
}
for b in bounces]).encode('utf-8')
)
|
Add user data export for bounce handling
|
Add user data export for bounce handling
|
Python
|
mit
|
fin/froide,fin/froide,stefanw/froide,stefanw/froide,fin/froide,stefanw/froide,stefanw/froide,fin/froide,stefanw/froide
|
---
+++
@@ -1,3 +1,5 @@
+import json
+
from django.apps import AppConfig
from django.utils.translation import ugettext_lazy as _
@@ -8,8 +10,10 @@
def ready(self):
from froide.account import account_canceled
+ from froide.account.export import registry
account_canceled.connect(cancel_user)
+ registry.register(export_user_data)
def cancel_user(sender, user=None, **kwargs):
@@ -18,3 +22,21 @@
if user is None:
return
Bounce.objects.filter(user=user).delete()
+
+
+def export_user_data(user):
+ from .models import Bounce
+
+ bounces = Bounce.objects.filter(user=user)
+ if not bounces:
+ return
+ yield ('bounces.json', json.dumps([
+ {
+ 'last_update': (
+ b.last_update.isoformat() if b.last_update else None
+ ),
+ 'bounces': b.bounces,
+ 'email': b.email,
+ }
+ for b in bounces]).encode('utf-8')
+ )
|
f9d7f69d7e8ae1dceaba09ac4412438076261744
|
tests/test_completion.py
|
tests/test_completion.py
|
import os
import subprocess
import sys
from pathlib import Path
import typer
from typer.testing import CliRunner
from first_steps import tutorial001 as mod
runner = CliRunner()
app = typer.Typer()
app.command()(mod.main)
def test_show_completion():
result = subprocess.run(
[
"bash",
"-c",
f"{sys.executable} -m coverage run {mod.__file__} --show-completion",
],
stdout=subprocess.PIPE,
stderr=subprocess.PIPE,
encoding="utf-8",
env={**os.environ, "SHELL": "/bin/bash"},
)
assert "_TUTORIAL001.PY_COMPLETE=complete-bash" in result.stdout
def test_install_completion():
bash_completion_path: Path = Path.home() / ".bash_completion"
text = ""
if bash_completion_path.is_file():
text = bash_completion_path.read_text()
result = subprocess.run(
[
"bash",
"-c",
f"{sys.executable} -m coverage run {mod.__file__} --install-completion",
],
stdout=subprocess.PIPE,
stderr=subprocess.PIPE,
encoding="utf-8",
env={**os.environ, "SHELL": "/bin/bash"},
)
new_text = bash_completion_path.read_text()
assert "_TUTORIAL001.PY_COMPLETE=complete-bash" in new_text
bash_completion_path.write_text(text)
|
import os
import subprocess
import sys
from pathlib import Path
import typer
from typer.testing import CliRunner
from first_steps import tutorial001 as mod
runner = CliRunner()
app = typer.Typer()
app.command()(mod.main)
def test_show_completion():
result = subprocess.run(
[
"bash",
"-c",
f"{sys.executable} -m coverage run {mod.__file__} --show-completion",
],
stdout=subprocess.PIPE,
stderr=subprocess.PIPE,
encoding="utf-8",
env={**os.environ, "SHELL": "/bin/bash"},
)
assert "_TUTORIAL001.PY_COMPLETE=complete-bash" in result.stdout
def test_install_completion():
bash_completion_path: Path = Path.home() / ".bash_completion"
text = ""
if bash_completion_path.is_file():
text = bash_completion_path.read_text()
result = subprocess.run(
[
"bash",
"-c",
f"{sys.executable} -m coverage run {mod.__file__} --install-completion",
],
stdout=subprocess.PIPE,
stderr=subprocess.PIPE,
encoding="utf-8",
env={**os.environ, "SHELL": "/bin/bash"},
)
new_text = bash_completion_path.read_text()
bash_completion_path.write_text(text)
assert "_TUTORIAL001.PY_COMPLETE=complete-bash" in new_text
assert "completion installed in" in result.stdout
assert "Completion will take effect once you restart the terminal." in result.stdout
|
Update completion tests, checking for printed message
|
:white_check_mark: Update completion tests, checking for printed message
|
Python
|
mit
|
tiangolo/typer,tiangolo/typer
|
---
+++
@@ -45,5 +45,7 @@
env={**os.environ, "SHELL": "/bin/bash"},
)
new_text = bash_completion_path.read_text()
+ bash_completion_path.write_text(text)
assert "_TUTORIAL001.PY_COMPLETE=complete-bash" in new_text
- bash_completion_path.write_text(text)
+ assert "completion installed in" in result.stdout
+ assert "Completion will take effect once you restart the terminal." in result.stdout
|
83ea38ee5616b1637cc2d983d4518d83793c7b72
|
lint/events.py
|
lint/events.py
|
from collections import defaultdict
import traceback
LINT_START = 'LINT_START'
LINT_RESULT = 'LINT_RESULT'
LINT_END = 'LINT_END'
listeners = defaultdict(set)
def subscribe(topic, fn):
listeners[topic].add(fn)
def unsubscribe(topic, fn):
try:
listeners[topic].remove(fn)
except KeyError:
pass
def broadcast(topic, message=None):
payload = message.copy() or {}
for fn in listeners.get(topic, []):
try:
fn(**payload)
except Exception:
traceback.print_exc()
map_fn_to_topic = {}
def on(topic):
def inner(fn):
subscribe(topic, fn)
map_fn_to_topic[fn] = topic
return fn
return inner
def off(fn):
topic = map_fn_to_topic.get(fn, None)
if topic:
unsubscribe(topic, fn)
|
from collections import defaultdict
import traceback
LINT_START = 'LINT_START' # (buffer_id)
LINT_RESULT = 'LINT_RESULT' # (buffer_id, linter_name, errors)
LINT_END = 'LINT_END' # (buffer_id)
listeners = defaultdict(set)
def subscribe(topic, fn):
listeners[topic].add(fn)
def unsubscribe(topic, fn):
try:
listeners[topic].remove(fn)
except KeyError:
pass
def broadcast(topic, message=None):
payload = message.copy() or {}
for fn in listeners.get(topic, []):
try:
fn(**payload)
except Exception:
traceback.print_exc()
map_fn_to_topic = {}
def on(topic):
def inner(fn):
subscribe(topic, fn)
map_fn_to_topic[fn] = topic
return fn
return inner
def off(fn):
topic = map_fn_to_topic.get(fn, None)
if topic:
unsubscribe(topic, fn)
|
Add very brief comments about the event types
|
Add very brief comments about the event types
|
Python
|
mit
|
SublimeLinter/SublimeLinter3,SublimeLinter/SublimeLinter3
|
---
+++
@@ -2,9 +2,9 @@
import traceback
-LINT_START = 'LINT_START'
-LINT_RESULT = 'LINT_RESULT'
-LINT_END = 'LINT_END'
+LINT_START = 'LINT_START' # (buffer_id)
+LINT_RESULT = 'LINT_RESULT' # (buffer_id, linter_name, errors)
+LINT_END = 'LINT_END' # (buffer_id)
listeners = defaultdict(set)
|
d572c7228d4c4fb84530b8a33df7bd4596e01d8b
|
app/billing/billing_schemas.py
|
app/billing/billing_schemas.py
|
from datetime import datetime
create_or_update_free_sms_fragment_limit_schema = {
"$schema": "http://json-schema.org/draft-04/schema#",
"description": "POST annual billing schema",
"type": "object",
"title": "Create",
"properties": {
"free_sms_fragment_limit": {"type": "integer", "minimum": 1},
},
"required": ["free_sms_fragment_limit"]
}
def serialize_ft_billing_remove_emails(data):
results = []
billed_notifications = [x for x in data if x.notification_type != 'email']
for notification in billed_notifications:
json_result = {
"month": (datetime.strftime(notification.month, "%B")),
"notification_type": notification.notification_type,
"billing_units": notification.billable_units,
"rate": float(notification.rate),
"postage": notification.postage,
}
results.append(json_result)
return results
def serialize_ft_billing_yearly_totals(data):
yearly_totals = []
for total in data:
json_result = {
"notification_type": total.notification_type,
"billing_units": total.billable_units,
"rate": float(total.rate),
"letter_total": float(total.billable_units * total.rate) if total.notification_type == 'letter' else 0
}
yearly_totals.append(json_result)
return yearly_totals
|
from datetime import datetime
create_or_update_free_sms_fragment_limit_schema = {
"$schema": "http://json-schema.org/draft-04/schema#",
"description": "POST annual billing schema",
"type": "object",
"title": "Create",
"properties": {
"free_sms_fragment_limit": {"type": "integer", "minimum": 0},
},
"required": ["free_sms_fragment_limit"]
}
def serialize_ft_billing_remove_emails(data):
results = []
billed_notifications = [x for x in data if x.notification_type != 'email']
for notification in billed_notifications:
json_result = {
"month": (datetime.strftime(notification.month, "%B")),
"notification_type": notification.notification_type,
"billing_units": notification.billable_units,
"rate": float(notification.rate),
"postage": notification.postage,
}
results.append(json_result)
return results
def serialize_ft_billing_yearly_totals(data):
yearly_totals = []
for total in data:
json_result = {
"notification_type": total.notification_type,
"billing_units": total.billable_units,
"rate": float(total.rate),
"letter_total": float(total.billable_units * total.rate) if total.notification_type == 'letter' else 0
}
yearly_totals.append(json_result)
return yearly_totals
|
Allow the free SMS fragment limit to be 0
|
Allow the free SMS fragment limit to be 0
This updates the schema so that the free allowance has a minimum value
of 0 instead of 1.
|
Python
|
mit
|
alphagov/notifications-api,alphagov/notifications-api
|
---
+++
@@ -6,7 +6,7 @@
"type": "object",
"title": "Create",
"properties": {
- "free_sms_fragment_limit": {"type": "integer", "minimum": 1},
+ "free_sms_fragment_limit": {"type": "integer", "minimum": 0},
},
"required": ["free_sms_fragment_limit"]
}
|
9b93ffa1662e99888c430e6deab01bd215055253
|
examples/analog_watch.py
|
examples/analog_watch.py
|
from microbit import *
hands = Image.ALL_CLOCKS
#A centre dot of brightness 2.
ticker_image = Image("2\n").crop(-2,-2,5,5)
#Adjust these to taste
MINUTE_BRIGHT = 0.1111
HOUR_BRIGHT = 0.55555
#Generate hands for 5 minute intervals
def fiveticks():
fivemins = 0
hours = 0
while True:
yield hands[fivemins]*MINUTE_BRIGHT + hands[hours]*HOUR_BRIGHT
fivemins = (fivemins+1)%12
hours = (hours + (fivemins == 0))%12
#Generate hands with ticker superimposed for 1 minute intervals.
def ticks():
on = True
for face in fiveticks():
for i in range(5):
if on:
yield face + ticker_image
else:
yield face - ticker_image
on = not on
#Run a clock speeded up 60 times, so we can watch the animation.
display.animate(ticks(), 1000)
|
from microbit import *
hands = Image.ALL_CLOCKS
#A centre dot of brightness 2.
ticker_image = Image("2\n").crop(-2,-2,5,5)
#Adjust these to taste
MINUTE_BRIGHT = 0.1111
HOUR_BRIGHT = 0.55555
#Generate hands for 5 minute intervals
def fiveticks():
fivemins = 0
hours = 0
while True:
yield hands[fivemins]*MINUTE_BRIGHT + hands[hours]*HOUR_BRIGHT
fivemins = (fivemins+1)%12
hours = (hours + (fivemins == 0))%12
#Generate hands with ticker superimposed for 1 minute intervals.
def ticks():
on = True
for face in fiveticks():
for i in range(5):
if on:
yield face + ticker_image
else:
yield face - ticker_image
on = not on
#Run a clock speeded up 60 times, so we can watch the animation.
display.show(ticks(), 1000)
|
Use display.show instead of display.animate
|
Use display.show instead of display.animate
|
Python
|
mit
|
JoeGlancy/micropython,JoeGlancy/micropython,JoeGlancy/micropython
|
---
+++
@@ -30,7 +30,7 @@
on = not on
#Run a clock speeded up 60 times, so we can watch the animation.
-display.animate(ticks(), 1000)
+display.show(ticks(), 1000)
|
4286d2d6a685571c70a8f48c3cd6802d13c4acef
|
braid/postgres.py
|
braid/postgres.py
|
from fabric.api import sudo, quiet
from braid import package
from pipes import quote
def install():
package.install(['postgresql-9.1', 'postgresql-server-dev-9.1'])
def _runQuery(query):
with quiet():
return sudo('psql --no-align --no-readline --no-password --quiet '
'--tuples-only -c {}'.format(quote(query)),
user='postgres', pty=False, combine_stderr=False)
def _dbExists(name):
res = _runQuery("select count(*) from pg_database "
"where datname = '{}';".format(name))
return res == '1'
def _userExists(name):
res = _runQuery("select count(*) from pg_user "
"where usename = '{}';".format(name))
return res == '1'
def createUser(name):
if not _userExists(name):
sudo('createuser -D -R -S {}'.format(name), user='postgres', pty=False)
def createDb(name, owner):
if not _dbExists(name):
sudo('createdb -O {} {}'.format(owner, name), user='postgres',
pty=False)
def grantRead(user, database):
"""
Grant read permissions to C{user} to all tables in C{database}.
"""
def grantReadWrite(user, database):
"""
Grant read and write permissions to C{user} to all tables in C{database}.
"""
|
from fabric.api import sudo, hide
from braid import package
from pipes import quote
def install():
package.install(['postgresql-9.1', 'postgresql-server-dev-9.1'])
def _runQuery(query):
with hide('running', 'output'):
return sudo('psql --no-align --no-readline --no-password --quiet '
'--tuples-only -c {}'.format(quote(query)),
user='postgres', pty=False, combine_stderr=False)
def _dbExists(name):
res = _runQuery("select count(*) from pg_database "
"where datname = '{}';".format(name))
return res == '1'
def _userExists(name):
res = _runQuery("select count(*) from pg_user "
"where usename = '{}';".format(name))
return res == '1'
def createUser(name):
if not _userExists(name):
sudo('createuser -D -R -S {}'.format(name), user='postgres', pty=False)
def createDb(name, owner):
if not _dbExists(name):
sudo('createdb -O {} {}'.format(owner, name), user='postgres',
pty=False)
def grantRead(user, database):
"""
Grant read permissions to C{user} to all tables in C{database}.
"""
def grantReadWrite(user, database):
"""
Grant read and write permissions to C{user} to all tables in C{database}.
"""
|
Make _runQuery to fail if the query fails, but still hide the execution messages
|
Make _runQuery to fail if the query fails, but still hide the execution messages
|
Python
|
mit
|
alex/braid,alex/braid
|
---
+++
@@ -1,4 +1,4 @@
-from fabric.api import sudo, quiet
+from fabric.api import sudo, hide
from braid import package
from pipes import quote
@@ -8,7 +8,7 @@
def _runQuery(query):
- with quiet():
+ with hide('running', 'output'):
return sudo('psql --no-align --no-readline --no-password --quiet '
'--tuples-only -c {}'.format(quote(query)),
user='postgres', pty=False, combine_stderr=False)
|
92d253fdce108162ab2ce05dd38da971ca42293d
|
keystone/contrib/kds/common/service.py
|
keystone/contrib/kds/common/service.py
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from oslo.config import cfg
from keystone.openstack.common import log
CONF = cfg.CONF
API_SERVICE_OPTS = [
cfg.StrOpt('bind_ip',
default='0.0.0.0',
help='IP for the server to bind to'),
cfg.IntOpt('port',
default=9109,
help='The port for the server'),
]
CONF.register_opts(API_SERVICE_OPTS)
def parse_args(args, default_config_files=None):
CONF(args=args[1:],
project='kds',
default_config_files=default_config_files)
def prepare_service(argv=[]):
cfg.set_defaults(log.log_opts,
default_log_levels=['sqlalchemy=WARN',
'eventlet.wsgi.server=WARN'
])
parse_args(argv)
log.setup('kds')
|
# Licensed under the Apache License, Version 2.0 (the "License"); you may
# not use this file except in compliance with the License. You may obtain
# a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT
# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the
# License for the specific language governing permissions and limitations
# under the License.
from oslo.config import cfg
from keystone.openstack.common import log
CONF = cfg.CONF
FILE_OPTIONS = {
None: [
cfg.StrOpt('bind_ip',
default='0.0.0.0',
help='IP for the server to bind to'),
cfg.IntOpt('port',
default=9109,
help='The port for the server')]}
def configure(conf=None):
if conf is None:
conf = CONF
for group in FILE_OPTIONS:
conf.register_opts(FILE_OPTIONS[group], group=group)
def parse_args(args, default_config_files=None):
CONF(args=args[1:],
project='kds',
default_config_files=default_config_files)
def prepare_service(argv=[]):
cfg.set_defaults(log.log_opts,
default_log_levels=['sqlalchemy=WARN',
'eventlet.wsgi.server=WARN'
])
parse_args(argv)
log.setup('kds')
configure()
|
Restructure KDS options to be more like Keystone's options
|
Restructure KDS options to be more like Keystone's options
Restructure the KDS options to be more closely aligned with the way
Keystone options work and allowing movement towards not registering
the options on import. This will also prevent KDS options from
appearing in the Keystone auto-generated sample config.
Change-Id: I073aa58ff3132e2714478f54c88c3a8200ff47da
|
Python
|
apache-2.0
|
rushiagr/keystone,jumpstarter-io/keystone,reeshupatel/demo,dstanek/keystone,idjaw/keystone,jonnary/keystone,vivekdhayaal/keystone,MaheshIBM/keystone,klmitch/keystone,rajalokan/keystone,rajalokan/keystone,nuxeh/keystone,ging/keystone,rushiagr/keystone,takeshineshiro/keystone,ilay09/keystone,nuxeh/keystone,roopali8/keystone,JioCloud/keystone,himanshu-setia/keystone,blueboxgroup/keystone,dims/keystone,vivekdhayaal/keystone,jamielennox/keystone,maestro-hybrid-cloud/keystone,idjaw/keystone,ilay09/keystone,maestro-hybrid-cloud/keystone,ajayaa/keystone,mahak/keystone,mahak/keystone,blueboxgroup/keystone,jamielennox/keystone,cernops/keystone,openstack/keystone,ging/keystone,rushiagr/keystone,rodrigods/keystone,JioCloud/keystone,openstack/keystone,jonnary/keystone,klmitch/keystone,reeshupatel/demo,takeshineshiro/keystone,dstanek/keystone,MaheshIBM/keystone,UTSA-ICS/keystone-kerberos,ajayaa/keystone,dstanek/keystone,rodrigods/keystone,nuxeh/keystone,vivekdhayaal/keystone,rajalokan/keystone,promptworks/keystone,cernops/keystone,jumpstarter-io/keystone,openstack/keystone,mahak/keystone,himanshu-setia/keystone,roopali8/keystone,ilay09/keystone,dims/keystone,reeshupatel/demo,promptworks/keystone,jumpstarter-io/keystone,UTSA-ICS/keystone-kerberos,promptworks/keystone
|
---
+++
@@ -16,16 +16,22 @@
CONF = cfg.CONF
-API_SERVICE_OPTS = [
- cfg.StrOpt('bind_ip',
- default='0.0.0.0',
- help='IP for the server to bind to'),
- cfg.IntOpt('port',
- default=9109,
- help='The port for the server'),
-]
+FILE_OPTIONS = {
+ None: [
+ cfg.StrOpt('bind_ip',
+ default='0.0.0.0',
+ help='IP for the server to bind to'),
+ cfg.IntOpt('port',
+ default=9109,
+ help='The port for the server')]}
-CONF.register_opts(API_SERVICE_OPTS)
+
+def configure(conf=None):
+ if conf is None:
+ conf = CONF
+
+ for group in FILE_OPTIONS:
+ conf.register_opts(FILE_OPTIONS[group], group=group)
def parse_args(args, default_config_files=None):
@@ -41,3 +47,6 @@
])
parse_args(argv)
log.setup('kds')
+
+
+configure()
|
eaa13f9005a8aaf8c748a98de697b03eee9e675b
|
salt/client/netapi.py
|
salt/client/netapi.py
|
# encoding: utf-8
'''
The main entry point for salt-api
'''
from __future__ import absolute_import
# Import python libs
import logging
# Import salt-api libs
import salt.loader
import salt.utils.process
logger = logging.getLogger(__name__)
class NetapiClient(object):
'''
Start each netapi module that is configured to run
'''
def __init__(self, opts):
self.opts = opts
self.process_manager = salt.utils.process.ProcessManager()
self.netapi = salt.loader.netapi(self.opts)
def run(self):
'''
Load and start all available api modules
'''
for fun in self.netapi:
if fun.endswith('.start'):
logger.info('Starting {0} netapi module'.format(fun))
self.process_manager.add_process(self.netapi[fun])
self.process_manager.run()
|
# encoding: utf-8
'''
The main entry point for salt-api
'''
from __future__ import absolute_import
# Import python libs
import logging
# Import salt-api libs
import salt.loader
import salt.utils.process
logger = logging.getLogger(__name__)
class NetapiClient(object):
'''
Start each netapi module that is configured to run
'''
def __init__(self, opts):
self.opts = opts
self.process_manager = salt.utils.process.ProcessManager()
self.netapi = salt.loader.netapi(self.opts)
def run(self):
'''
Load and start all available api modules
'''
if not len(self.netapi):
logger.error("Did not find any netapi configurations, nothing to start")
for fun in self.netapi:
if fun.endswith('.start'):
logger.info('Starting {0} netapi module'.format(fun))
self.process_manager.add_process(self.netapi[fun])
self.process_manager.run()
|
Add log error if we run salt-api w/ no config
|
Add log error if we run salt-api w/ no config
Currently, the salt-api script will exit with no error or hint of why it
failed if there is no netapi module configured. Added a short line if
we find no api modules to start, warning the user that the config may be
missing.
Fixes #28240
|
Python
|
apache-2.0
|
saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt
|
---
+++
@@ -26,6 +26,9 @@
'''
Load and start all available api modules
'''
+ if not len(self.netapi):
+ logger.error("Did not find any netapi configurations, nothing to start")
+
for fun in self.netapi:
if fun.endswith('.start'):
logger.info('Starting {0} netapi module'.format(fun))
|
917c2701f25f15b0d39a4dd8f93254b75aa058dd
|
script/release/release/pypi.py
|
script/release/release/pypi.py
|
from __future__ import absolute_import
from __future__ import unicode_literals
from configparser import Error
from requests.exceptions import HTTPError
from twine.commands.upload import main as twine_upload
from twine.utils import get_config
from .utils import ScriptError
def pypi_upload(args):
print('Uploading to PyPi')
try:
rel = args.release.replace('-rc', 'rc')
twine_upload([
'dist/docker_compose-{}*.whl'.format(rel),
'dist/docker-compose-{}*.tar.gz'.format(rel)
])
except HTTPError as e:
if e.response.status_code == 400 and 'File already exists' in e.message:
if not args.finalize_resume:
raise ScriptError(
'Package already uploaded on PyPi.'
)
print('Skipping PyPi upload - package already uploaded')
else:
raise ScriptError('Unexpected HTTP error uploading package to PyPi: {}'.format(e))
def check_pypirc():
try:
config = get_config()
except Error as e:
raise ScriptError('Failed to parse .pypirc file: {}'.format(e))
if config is None:
raise ScriptError('Failed to parse .pypirc file')
if 'pypi' not in config:
raise ScriptError('Missing [pypi] section in .pypirc file')
if not (config['pypi'].get('username') and config['pypi'].get('password')):
raise ScriptError('Missing login/password pair for pypi repo')
|
from __future__ import absolute_import
from __future__ import unicode_literals
from configparser import Error
from requests.exceptions import HTTPError
from twine.commands.upload import main as twine_upload
from twine.utils import get_config
from .utils import ScriptError
def pypi_upload(args):
print('Uploading to PyPi')
try:
rel = args.release.replace('-rc', 'rc')
twine_upload([
'dist/docker_compose-{}*.whl'.format(rel),
'dist/docker-compose-{}*.tar.gz'.format(rel)
])
except HTTPError as e:
if e.response.status_code == 400 and 'File already exists' in str(e):
if not args.finalize_resume:
raise ScriptError(
'Package already uploaded on PyPi.'
)
print('Skipping PyPi upload - package already uploaded')
else:
raise ScriptError('Unexpected HTTP error uploading package to PyPi: {}'.format(e))
def check_pypirc():
try:
config = get_config()
except Error as e:
raise ScriptError('Failed to parse .pypirc file: {}'.format(e))
if config is None:
raise ScriptError('Failed to parse .pypirc file')
if 'pypi' not in config:
raise ScriptError('Missing [pypi] section in .pypirc file')
if not (config['pypi'].get('username') and config['pypi'].get('password')):
raise ScriptError('Missing login/password pair for pypi repo')
|
Fix script for release file already present case
|
Fix script for release file already present case
This avoids a:
"AttributeError: 'HTTPError' object has no attribute 'message'"
Signed-off-by: Ulysses Souza <a0ff1337c6a0e43e9559f5f67fc3acb852912071@docker.com>
|
Python
|
apache-2.0
|
thaJeztah/compose,vdemeester/compose,vdemeester/compose,thaJeztah/compose
|
---
+++
@@ -18,7 +18,7 @@
'dist/docker-compose-{}*.tar.gz'.format(rel)
])
except HTTPError as e:
- if e.response.status_code == 400 and 'File already exists' in e.message:
+ if e.response.status_code == 400 and 'File already exists' in str(e):
if not args.finalize_resume:
raise ScriptError(
'Package already uploaded on PyPi.'
|
0f1ed52e7525ea5f41d63642bca1eaeb9d5af8ba
|
emission/core/wrapper/labelprediction.py
|
emission/core/wrapper/labelprediction.py
|
# Based on modeprediction.py
import emission.core.wrapper.wrapperbase as ecwb
class Labelprediction(ecwb.WrapperBase):
props = {"trip_id": ecwb.WrapperBase.Access.WORM, # the trip that this is part of
"prediction": ecwb.WrapperBase.Access.WORM, # What we predict
"start_ts": ecwb.WrapperBase.Access.WORM, # start time for the prediction, so that it can be captured in time-based queries, e.g. to reset the pipeline
"end_ts": ecwb.WrapperBase.Access.WORM, # end time for the prediction, so that it can be captured in time-based queries, e.g. to reset the pipeline
}
enums = {}
geojson = {}
local_dates = {}
def _populateDependencies(self):
pass
|
# Based on modeprediction.py
import emission.core.wrapper.wrapperbase as ecwb
# The "prediction" data structure is a list of label possibilities, each one consisting of a set of labels and a probability:
# [
# {"labels": {"labeltype1": "labelvalue1", "labeltype2": "labelvalue2"}, "p": 0.61},
# {"labels": {"labeltype1": "labelvalue3", "labeltype2": "labelvalue4"}, "p": 0.27},
# ...
# ]
class Labelprediction(ecwb.WrapperBase):
props = {"trip_id": ecwb.WrapperBase.Access.WORM, # the trip that this is part of
"prediction": ecwb.WrapperBase.Access.WORM, # What we predict -- see above
"start_ts": ecwb.WrapperBase.Access.WORM, # start time for the prediction, so that it can be captured in time-based queries, e.g. to reset the pipeline
"end_ts": ecwb.WrapperBase.Access.WORM, # end time for the prediction, so that it can be captured in time-based queries, e.g. to reset the pipeline
}
enums = {}
geojson = {}
local_dates = {}
def _populateDependencies(self):
pass
|
Add comments explaining prediction data structure
|
Add comments explaining prediction data structure
|
Python
|
bsd-3-clause
|
shankari/e-mission-server,shankari/e-mission-server,shankari/e-mission-server,shankari/e-mission-server,e-mission/e-mission-server,e-mission/e-mission-server,e-mission/e-mission-server,e-mission/e-mission-server
|
---
+++
@@ -1,9 +1,16 @@
# Based on modeprediction.py
import emission.core.wrapper.wrapperbase as ecwb
+# The "prediction" data structure is a list of label possibilities, each one consisting of a set of labels and a probability:
+# [
+# {"labels": {"labeltype1": "labelvalue1", "labeltype2": "labelvalue2"}, "p": 0.61},
+# {"labels": {"labeltype1": "labelvalue3", "labeltype2": "labelvalue4"}, "p": 0.27},
+# ...
+# ]
+
class Labelprediction(ecwb.WrapperBase):
props = {"trip_id": ecwb.WrapperBase.Access.WORM, # the trip that this is part of
- "prediction": ecwb.WrapperBase.Access.WORM, # What we predict
+ "prediction": ecwb.WrapperBase.Access.WORM, # What we predict -- see above
"start_ts": ecwb.WrapperBase.Access.WORM, # start time for the prediction, so that it can be captured in time-based queries, e.g. to reset the pipeline
"end_ts": ecwb.WrapperBase.Access.WORM, # end time for the prediction, so that it can be captured in time-based queries, e.g. to reset the pipeline
}
|
3e614788b179fa3549098870c5ba9ca0ce4a35a1
|
src/euth/setup.py
|
src/euth/setup.py
|
"""Adhocracy frontend customization package."""
import os
import version
from setuptools import setup, find_packages
here = os.path.abspath(os.path.dirname(__file__))
README = open(os.path.join(here, 'README.rst')).read()
CHANGES = open(os.path.join(here, 'CHANGES.rst')).read()
requires = ['adhocracy_frontend',
'adhocracy_kit',
]
test_requires = ['adhocracy_frontend[test]',
'adhocracy_kit[test]',
]
debug_requires = ['adhocracy_frontend[debug]',
'adhocracy_kit[debug]',
]
setup(name='euth',
version=version.get_git_version(),
description='Adhocracy meta package for backend/frontend customization.',
long_description=README + '\n\n' + CHANGES,
classifiers=["Programming Language :: Python",
"Framework :: Pylons",
"Topic :: Internet :: WWW/HTTP",
"Topic :: Internet :: WWW/HTTP :: WSGI :: Application",
],
author='',
author_email='',
url='',
keywords='web pyramid pylons adhocracy',
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=requires,
tests_require=requires,
extras_require={'test': test_requires,
'debug': debug_requires},
entry_points="""\
[paste.app_factory]
main = euth:main
""",
)
|
"""Adhocracy frontend customization package."""
import os
import version
from setuptools import setup, find_packages
here = os.path.abspath(os.path.dirname(__file__))
README = open(os.path.join(here, 'README.rst')).read()
CHANGES = open(os.path.join(here, 'CHANGES.rst')).read()
requires = ['adhocracy_frontend',
'adhocracy_euth',
]
test_requires = ['adhocracy_frontend[test]',
'adhocracy_euth[test]',
]
debug_requires = ['adhocracy_frontend[debug]',
'adhocracy_euth[debug]',
]
setup(name='euth',
version=version.get_git_version(),
description='Adhocracy meta package for backend/frontend customization.',
long_description=README + '\n\n' + CHANGES,
classifiers=["Programming Language :: Python",
"Framework :: Pylons",
"Topic :: Internet :: WWW/HTTP",
"Topic :: Internet :: WWW/HTTP :: WSGI :: Application",
],
author='',
author_email='',
url='',
keywords='web pyramid pylons adhocracy',
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=requires,
tests_require=requires,
extras_require={'test': test_requires,
'debug': debug_requires},
entry_points="""\
[paste.app_factory]
main = euth:main
""",
)
|
Remove old dependency to kit from euth frontend
|
Remove old dependency to kit from euth frontend
|
Python
|
agpl-3.0
|
liqd/adhocracy3.mercator,liqd/adhocracy3.mercator,liqd/adhocracy3.mercator,liqd/adhocracy3.mercator,liqd/adhocracy3.mercator,liqd/adhocracy3.mercator,liqd/adhocracy3.mercator
|
---
+++
@@ -9,15 +9,15 @@
CHANGES = open(os.path.join(here, 'CHANGES.rst')).read()
requires = ['adhocracy_frontend',
- 'adhocracy_kit',
+ 'adhocracy_euth',
]
test_requires = ['adhocracy_frontend[test]',
- 'adhocracy_kit[test]',
+ 'adhocracy_euth[test]',
]
debug_requires = ['adhocracy_frontend[debug]',
- 'adhocracy_kit[debug]',
+ 'adhocracy_euth[debug]',
]
setup(name='euth',
|
841289e7092a5e4bf485ef525f358aa3a58f7fb9
|
meta-iotqa/lib/oeqa/runtime/sanity/comm_btcheck.py
|
meta-iotqa/lib/oeqa/runtime/sanity/comm_btcheck.py
|
import time
from oeqa.oetest import oeRuntimeTest
from oeqa.utils.decorators import tag
@tag(TestType="FVT", FeatureID="IOTOS-453")
class CommBluetoothTest(oeRuntimeTest):
"""
@class CommBluetoothTest
"""
log = ""
def setUp(self):
self.target.run('connmanctl enable bluetooth')
time.sleep(8)
def tearDown(self):
self.target.run('connmanctl disable bluetooth')
def target_collect_info(self, cmd):
"""
@fn target_collect_info
@param self
@param cmd
@return
"""
(status, output) = self.target.run(cmd)
self.log = self.log + "\n\n[Debug] Command output --- %s: \n" % cmd
self.log = self.log + output
'''Bluetooth device check'''
def test_comm_btcheck(self):
'''check bluetooth device
@fn test_comm_btcheck
@param self
@return
'''
# un-block software rfkill lock
self.target.run('rfkill unblock all')
# This is special for edison platform
# Collect system information as log
self.target_collect_info("ifconfig")
self.target_collect_info("hciconfig")
self.target_collect_info("lsmod")
# Detect BT device status
(status, output) = self.target.run('hciconfig hci0')
##
# TESTPOINT: #1, test_comm_btcheck
#
self.assertEqual(status, 0, msg="Error messages: %s" % self.log)
|
import time
from oeqa.oetest import oeRuntimeTest
from oeqa.utils.decorators import tag
@tag(TestType="FVT", FeatureID="IOTOS-453")
class CommBluetoothTest(oeRuntimeTest):
"""
@class CommBluetoothTest
"""
log = ""
def setUp(self):
self.target.run('connmanctl enable bluetooth')
time.sleep(8)
def tearDown(self):
self.target.run('connmanctl disable bluetooth')
def target_collect_info(self, cmd):
"""
@fn target_collect_info
@param self
@param cmd
@return
"""
(status, output) = self.target.run(cmd)
self.log = self.log + "\n\n[Debug] Command output --- %s: \n" % cmd
self.log = self.log + output
'''Bluetooth device check'''
def test_comm_btcheck(self):
'''check bluetooth device
@fn test_comm_btcheck
@param self
@return
'''
# Collect system information as log
self.target_collect_info("ifconfig")
self.target_collect_info("hciconfig")
self.target_collect_info("lsmod")
# Detect BT device status
(status, output) = self.target.run('hciconfig hci0')
##
# TESTPOINT: #1, test_comm_btcheck
#
self.assertEqual(status, 0, msg="Error messages: %s" % self.log)
|
Remove Edison specific command from Bluetooth test
|
meta-iotqa: Remove Edison specific command from Bluetooth test
The platform isn't supported anymore and the command isn't needed with
current devices.
Signed-off-by: Simo Kuusela <4755938158c3c622d3884e9a75ed20dc865bc695@intel.com>
|
Python
|
mit
|
YinThong/intel-iot-refkit,mythi/intel-iot-refkit,mythi/intel-iot-refkit,klihub/intel-iot-refkit,jairglez/intel-iot-refkit,klihub/intel-iot-refkit,intel/intel-iot-refkit,YinThong/intel-iot-refkit,jairglez/intel-iot-refkit,YinThong/intel-iot-refkit,YinThong/intel-iot-refkit,mythi/intel-iot-refkit,intel/intel-iot-refkit,klihub/intel-iot-refkit,klihub/intel-iot-refkit,mythi/intel-iot-refkit,YinThong/intel-iot-refkit,intel/intel-iot-refkit,ipuustin/intel-iot-refkit,ipuustin/intel-iot-refkit,intel/intel-iot-refkit,intel/intel-iot-refkit,jairglez/intel-iot-refkit,intel/intel-iot-refkit,jairglez/intel-iot-refkit,ipuustin/intel-iot-refkit,jairglez/intel-iot-refkit,jairglez/intel-iot-refkit,mythi/intel-iot-refkit,ipuustin/intel-iot-refkit,ipuustin/intel-iot-refkit,mythi/intel-iot-refkit,klihub/intel-iot-refkit,intel/intel-iot-refkit,YinThong/intel-iot-refkit,jairglez/intel-iot-refkit,klihub/intel-iot-refkit,ipuustin/intel-iot-refkit,klihub/intel-iot-refkit,ipuustin/intel-iot-refkit,YinThong/intel-iot-refkit,mythi/intel-iot-refkit
|
---
+++
@@ -34,9 +34,6 @@
@param self
@return
'''
- # un-block software rfkill lock
- self.target.run('rfkill unblock all')
- # This is special for edison platform
# Collect system information as log
self.target_collect_info("ifconfig")
self.target_collect_info("hciconfig")
|
4a48be97254f4626eaadf86492d98e0dd640d43d
|
flower/utils/__init__.py
|
flower/utils/__init__.py
|
from __future__ import absolute_import
import uuid
import base64
import os.path
from .. import __version__
def gen_cookie_secret():
return base64.b64encode(uuid.uuid4().bytes + uuid.uuid4().bytes)
def bugreport():
try:
import celery
return 'flower -> %s' % __version__ + celery.bugreport()
except (ImportError, AttributeError):
return 'Unknown Celery version'
def abs_path(path):
path = os.path.expanduser(path)
if not os.path.isabs(path):
cwd = os.environ.get('PWD') or os.getcwd()
path = os.path.join(cwd, path)
return path
|
from __future__ import absolute_import
import uuid
import base64
import os.path
from .. import __version__
def gen_cookie_secret():
return base64.b64encode(uuid.uuid4().bytes + uuid.uuid4().bytes)
def bugreport():
try:
import celery
import tornado
import babel
return 'flower -> flower:%s tornado:%s babel:%s' %\
(__version__, tornado.version, babel.__version__)\
+ celery.bugreport()
except (ImportError, AttributeError):
return 'Unknown Celery version'
def abs_path(path):
path = os.path.expanduser(path)
if not os.path.isabs(path):
cwd = os.environ.get('PWD') or os.getcwd()
path = os.path.join(cwd, path)
return path
|
Include tornado and babel version info in bug report
|
Include tornado and babel version info in bug report
|
Python
|
bsd-3-clause
|
raphaelmerx/flower,ChinaQuants/flower,pj/flower,Lingling7/flower,tellapart/flower,alexmojaki/flower,ucb-bar/bar-crawl-web,Lingling7/flower,ChinaQuants/flower,jzhou77/flower,Lingling7/flower,lucius-feng/flower,lucius-feng/flower,ucb-bar/bar-crawl-web,allengaller/flower,asmodehn/flower,allengaller/flower,lucius-feng/flower,marrybird/flower,pygeek/flower,ucb-bar/bar-crawl-web,ChinaQuants/flower,pj/flower,pygeek/flower,alexmojaki/flower,jzhou77/flower,barseghyanartur/flower,jzhou77/flower,marrybird/flower,asmodehn/flower,pj/flower,pygeek/flower,getupcloud/flower,marrybird/flower,barseghyanartur/flower,alexmojaki/flower,getupcloud/flower,getupcloud/flower,tellapart/flower,asmodehn/flower,raphaelmerx/flower,tellapart/flower,raphaelmerx/flower,allengaller/flower,barseghyanartur/flower
|
---
+++
@@ -14,7 +14,11 @@
def bugreport():
try:
import celery
- return 'flower -> %s' % __version__ + celery.bugreport()
+ import tornado
+ import babel
+ return 'flower -> flower:%s tornado:%s babel:%s' %\
+ (__version__, tornado.version, babel.__version__)\
+ + celery.bugreport()
except (ImportError, AttributeError):
return 'Unknown Celery version'
|
98e574061dc430b1cad6c9bbef5b55aa92803206
|
command_line/rebin_images.py
|
command_line/rebin_images.py
|
# LIBTBX_SET_DISPATCHER_NAME dev.dials.rebin_images
from __future__ import division
def rebin_images(in_template, out_template, start, end):
from dials.util.rebin_images import main
in_images = [in_template % j for j in range(start, end + 1)]
out_images = [out_template % j for j in range(start, end + 1)]
main(in_images, out_images)
if __name__ == '__main__':
import sys
if len(sys.argv) != 5:
raise RuntimeError, '%s in_\%04d.cbf out_\%04d.cbf start end' % \
sys.argv[0]
in_template = sys.argv[1]
out_template = sys.argv[2]
start = int(sys.argv[3])
end = int(sys.argv[4])
rebin_images(in_template, out_template, start, end)
|
# LIBTBX_SET_DISPATCHER_NAME dev.dials.rebin_images
from __future__ import division
def rebin_images(in_template, out_template, start, end):
from dials.util.rebin_images import main
in_images = [in_template % j for j in range(start, end + 1)]
out_images = [out_template % j for j in range(start, end + 1)]
main(in_images, out_images)
return
if __name__ == '__main__':
import sys
if len(sys.argv) != 5:
raise RuntimeError, '%s in_\%04d.cbf out_\%04d.cbf start end' % \
sys.argv[0]
in_template = sys.argv[1]
out_template = sys.argv[2]
start = int(sys.argv[3])
end = int(sys.argv[4])
rebin_images(in_template, out_template, start, end)
|
Check email update on branch for YT only
|
Check email update on branch for YT only
|
Python
|
bsd-3-clause
|
dials/dials,dials/dials,dials/dials,dials/dials,dials/dials
|
---
+++
@@ -7,6 +7,7 @@
in_images = [in_template % j for j in range(start, end + 1)]
out_images = [out_template % j for j in range(start, end + 1)]
main(in_images, out_images)
+ return
if __name__ == '__main__':
import sys
|
4e483d6443e809f9e7e1a59c3fe959fd5f42f938
|
simple-cipher/simple_cipher.py
|
simple-cipher/simple_cipher.py
|
import math
import secrets
from string import ascii_lowercase
class Cipher(object):
def __init__(self, key=None):
if not key:
key = Cipher._random_key()
if not key.isalpha() or not key.islower():
raise ValueError("Key must consist only of lowercase letters")
self.key = key
self._key = [ord(k)-97 for k in key]
def encode(self, s):
key = self._key * math.ceil(len(s)/len(self._key))
chars = [c for c in s.lower() if c in ascii_lowercase]
return "".join(Cipher._shift(c, k) for c, k in zip(chars, key))
def decode(self, s):
key = self._key * math.ceil(len(s)/len(self._key))
chars = [c for c in s.lower() if c in ascii_lowercase]
return "".join(Cipher._shift(c, -k) for c, k in zip(chars, key))
@staticmethod
def _shift(char, key):
return chr(97 + ((ord(char) - 97 + key) % 26))
@staticmethod
def _random_key(length=256):
return "".join(secrets.choice(ascii_lowercase) for _ in range(length))
class Caesar(Cipher):
def __init__(self):
Cipher.__init__(self, "d")
|
import math
import secrets
from string import ascii_lowercase
class Cipher(object):
def __init__(self, key=None):
if not key:
key = self._random_key()
if not key.isalpha() or not key.islower():
raise ValueError("Key must consist only of lowercase letters")
self.key = key
self._key = [ord(k)-97 for k in key]
def encode(self, s):
key = self._key * math.ceil(len(s)/len(self._key))
chars = [c for c in s.lower() if c in ascii_lowercase]
return "".join(self._shift(c, k) for c, k in zip(chars, key))
def decode(self, s):
key = self._key * math.ceil(len(s)/len(self._key))
chars = [c for c in s.lower() if c in ascii_lowercase]
return "".join(self._shift(c, -k) for c, k in zip(chars, key))
@staticmethod
def _shift(char, key):
return chr(97 + ((ord(char) - 97 + key) % 26))
@staticmethod
def _random_key(length=256):
return "".join(secrets.choice(ascii_lowercase) for _ in range(length))
class Caesar(Cipher):
def __init__(self):
super().__init__("d")
|
Use super() and self within the Cipher and Caesar classes
|
Use super() and self within the Cipher and Caesar classes
|
Python
|
agpl-3.0
|
CubicComet/exercism-python-solutions
|
---
+++
@@ -6,7 +6,7 @@
class Cipher(object):
def __init__(self, key=None):
if not key:
- key = Cipher._random_key()
+ key = self._random_key()
if not key.isalpha() or not key.islower():
raise ValueError("Key must consist only of lowercase letters")
self.key = key
@@ -15,12 +15,12 @@
def encode(self, s):
key = self._key * math.ceil(len(s)/len(self._key))
chars = [c for c in s.lower() if c in ascii_lowercase]
- return "".join(Cipher._shift(c, k) for c, k in zip(chars, key))
+ return "".join(self._shift(c, k) for c, k in zip(chars, key))
def decode(self, s):
key = self._key * math.ceil(len(s)/len(self._key))
chars = [c for c in s.lower() if c in ascii_lowercase]
- return "".join(Cipher._shift(c, -k) for c, k in zip(chars, key))
+ return "".join(self._shift(c, -k) for c, k in zip(chars, key))
@staticmethod
def _shift(char, key):
@@ -33,4 +33,4 @@
class Caesar(Cipher):
def __init__(self):
- Cipher.__init__(self, "d")
+ super().__init__("d")
|
5ca96beb26dd2ab5285a57f5cade6f01160df368
|
joequery/blog/posts/code/notes-on-dynamic-programming-part-1/meta.py
|
joequery/blog/posts/code/notes-on-dynamic-programming-part-1/meta.py
|
title="Notes on dynamic programming - part 1"
description="""
Part 1 of extensive notes discussing the fundamentals of dynamic programming.
Examples in these notes include the Fibonacci sequence and Warshall's
algorithm. Pseudocode and Python implementations of the algorithms are
provided.
"""
time="2012-12-10 Mon 02:28 AM"
# related=[("Some article", "its/url")]
|
title="Notes on dynamic programming - part 1"
description="""
Part 1 of extensive notes discussing the fundamentals of dynamic programming.
Examples in these notes include the Fibonacci sequence, the Binomial Formula,
and Warshall's algorithm. Python implementations of the algorithms are
provided.
"""
time="2012-12-10 Mon 02:48 AM"
# related=[("Some article", "its/url")]
|
Update description and timestamp for dynamic programming part 1
|
Update description and timestamp for dynamic programming part 1
|
Python
|
mit
|
joequery/joequery.me,joequery/joequery.me,joequery/joequery.me,joequery/joequery.me
|
---
+++
@@ -1,9 +1,9 @@
title="Notes on dynamic programming - part 1"
description="""
Part 1 of extensive notes discussing the fundamentals of dynamic programming.
-Examples in these notes include the Fibonacci sequence and Warshall's
-algorithm. Pseudocode and Python implementations of the algorithms are
+Examples in these notes include the Fibonacci sequence, the Binomial Formula,
+and Warshall's algorithm. Python implementations of the algorithms are
provided.
"""
-time="2012-12-10 Mon 02:28 AM"
+time="2012-12-10 Mon 02:48 AM"
# related=[("Some article", "its/url")]
|
c04b9813b5d6d3f8bc8eaa7be2d49d32f150aaf2
|
tests/test_authentication.py
|
tests/test_authentication.py
|
import unittest
from flask import json
from api import db
from api.BucketListAPI import app
from instance.config import application_config
class AuthenticationTestCase(unittest.TestCase):
def setUp(self):
app.config.from_object(application_config['TestingEnv'])
self.client = app.test_client()
# Binds the app to current context
with app.app_context():
# Create all tables
db.create_all()
def test_index_route(self):
response = self.client.get('/')
self.assertEqual(response.status_code, 201)
self.assertIn('Welcome Message', response.data.decode())
def tearDown(self):
# Drop all tables
with app.app_context():
# Drop all tables
db.session.remove()
db.drop_all()
if __name__ == '__main__':
unittest.main()
|
import unittest
from flask import json
from api import db
from api.BucketListAPI import app
from instance.config import application_config
class AuthenticationTestCase(unittest.TestCase):
def setUp(self):
app.config.from_object(application_config['TestingEnv'])
self.client = app.test_client()
# Binds the app to current context
with app.app_context():
# Create all tables
db.create_all()
def test_index_route(self):
response = self.client.get('/')
self.assertEqual(response.status_code, 201)
self.assertIn('Welcome Message', response.data.decode())
def test_registration_with_missing_dredentials(self):
"""Should throw error for missing credentials"""
user = json.dumps({
'name': '',
'email': '',
'password': ''
})
response = self.client.post('/auth/register', data=user)
self.assertEqual(response.status_code, 400)
self.assertIn('Missing', response.data.decode())
def tearDown(self):
# Drop all tables
with app.app_context():
# Drop all tables
db.session.remove()
db.drop_all()
if __name__ == '__main__':
unittest.main()
|
Add test for user with missing credentials
|
Add test for user with missing credentials
|
Python
|
mit
|
patlub/BucketListAPI,patlub/BucketListAPI
|
---
+++
@@ -20,6 +20,17 @@
self.assertEqual(response.status_code, 201)
self.assertIn('Welcome Message', response.data.decode())
+ def test_registration_with_missing_dredentials(self):
+ """Should throw error for missing credentials"""
+ user = json.dumps({
+ 'name': '',
+ 'email': '',
+ 'password': ''
+ })
+ response = self.client.post('/auth/register', data=user)
+ self.assertEqual(response.status_code, 400)
+ self.assertIn('Missing', response.data.decode())
+
def tearDown(self):
# Drop all tables
with app.app_context():
|
47f1b47f37da4f9a3444a2ac6cc7b7a0affafbf3
|
node_bridge.py
|
node_bridge.py
|
import os
import platform
import subprocess
IS_MACOS = platform.system() == 'Darwin'
IS_WINDOWS = platform.system() == 'Windows'
def node_bridge(data, bin, args=[]):
env = None
startupinfo = None
if IS_MACOS:
# GUI apps on macOS doesn't contain .bashrc/.zshrc set paths
env = os.environ.copy()
env['PATH'] += ':/usr/local/bin'
if IS_WINDOWS:
startupinfo = subprocess.STARTUPINFO()
startupinfo.dwFlags |= subprocess.STARTF_USESHOWWINDOW
try:
p = subprocess.Popen(['node', bin] + args,
stdout=subprocess.PIPE, stdin=subprocess.PIPE, stderr=subprocess.PIPE,
env=env, startupinfo=startupinfo)
except OSError:
raise Exception('Couldn\'t find Node.js. Make sure it\'s in your $PATH by running `node -v` in your command-line.')
stdout, stderr = p.communicate(input=data.encode('utf-8'))
stdout = stdout.decode('utf-8')
stderr = stderr.decode('utf-8')
if stderr:
raise Exception('Error: %s' % stderr)
else:
return stdout
|
import os
import platform
import subprocess
IS_MACOS = platform.system() == 'Darwin'
IS_WINDOWS = platform.system() == 'Windows'
def node_bridge(data, bin, args=[]):
env = None
startupinfo = None
if IS_MACOS:
# GUI apps on macOS doesn't contain .bashrc/.zshrc set paths
env = os.environ.copy()
env['PATH'] += os.path.expanduser('~/n/bin')
env['PATH'] += ':/usr/local/bin'
if IS_WINDOWS:
startupinfo = subprocess.STARTUPINFO()
startupinfo.dwFlags |= subprocess.STARTF_USESHOWWINDOW
try:
p = subprocess.Popen(['node', bin] + args,
stdout=subprocess.PIPE, stdin=subprocess.PIPE, stderr=subprocess.PIPE,
env=env, startupinfo=startupinfo)
except OSError:
raise Exception('Couldn\'t find Node.js. Make sure it\'s in your $PATH by running `node -v` in your command-line.')
stdout, stderr = p.communicate(input=data.encode('utf-8'))
stdout = stdout.decode('utf-8')
stderr = stderr.decode('utf-8')
if stderr:
raise Exception('Error: %s' % stderr)
else:
return stdout
|
Add support for `n` Node.js version manager
|
Add support for `n` Node.js version manager
|
Python
|
mit
|
sindresorhus/sublime-autoprefixer,sindresorhus/sublime-autoprefixer,sindresorhus/sublime-autoprefixer
|
---
+++
@@ -11,6 +11,7 @@
if IS_MACOS:
# GUI apps on macOS doesn't contain .bashrc/.zshrc set paths
env = os.environ.copy()
+ env['PATH'] += os.path.expanduser('~/n/bin')
env['PATH'] += ':/usr/local/bin'
if IS_WINDOWS:
startupinfo = subprocess.STARTUPINFO()
|
c7660db45e0275a685a6cc450fd4341a69c52b92
|
threaded_multihost/fields.py
|
threaded_multihost/fields.py
|
from django.db.models import ForeignKey
from django.contrib.auth.models import User
import threadlocals
class UserField(ForeignKey):
""" UserField
By defaults, foreign key to User; null=True, blank=True
"""
def __init__(self, **kwargs):
kwargs.setdefault('null', True)
kwargs.setdefault('blank', True)
ForeignKey.__init__(self, User, **kwargs)
class CreatorField(UserField):
""" CreatorField
By default, sets editable=False, default=threadlocals.get_current_user
"""
def __init__(self, **kwargs):
kwargs.setdefault('editable', False)
kwargs.setdefault('default', threadlocals.get_current_user)
UserField.__init__(self, **kwargs)
class EditorField(CreatorField):
""" EditorField
By default, sets editable=False, default=threadlocals.get_current_user
Sets value to get_current_user() on each save of the model.
"""
def __init__(self, **kwargs):
super(CreatorField, self).__init__(**kwargs)
def pre_save(self, model_instance, add):
value = threadlocals.get_current_user()
setattr(model_instance, self.name, value)
if value:
value = value.pk
setattr(model_instance, self.attname, value)
return value
|
from django.db.models import ForeignKey
from django.contrib.auth.models import User
import threadlocals
class UserField(ForeignKey):
""" UserField
By defaults, foreign key to User; null=True, blank=True
"""
def __init__(self, **kwargs):
kwargs.setdefault('to', User)
kwargs.setdefault('null', True)
kwargs.setdefault('blank', True)
ForeignKey.__init__(self, **kwargs)
class CreatorField(UserField):
""" CreatorField
By default, sets editable=False, default=threadlocals.get_current_user
"""
def __init__(self, **kwargs):
kwargs.setdefault('editable', False)
kwargs.setdefault('default', threadlocals.get_current_user)
UserField.__init__(self, **kwargs)
class EditorField(CreatorField):
""" EditorField
By default, sets editable=False, default=threadlocals.get_current_user
Sets value to get_current_user() on each save of the model.
"""
def __init__(self, **kwargs):
super(CreatorField, self).__init__(**kwargs)
def pre_save(self, model_instance, add):
value = threadlocals.get_current_user()
setattr(model_instance, self.name, value)
if value:
value = value.pk
setattr(model_instance, self.attname, value)
return value
try:
from south.modelsinspector import add_introspection_rules
except ImportError:
add_introspection_rules = False
if add_introspection_rules:
add_introspection_rules([], [r"^threaded_multihost\.fields\.(User|Creator|Editor)Field"])
|
Patch from chrischambers to enable south migrations.
|
Patch from chrischambers to enable south migrations.
|
Python
|
bsd-3-clause
|
diver-in-sky/django-threaded-multihost
|
---
+++
@@ -11,11 +11,12 @@
"""
def __init__(self, **kwargs):
+ kwargs.setdefault('to', User)
kwargs.setdefault('null', True)
kwargs.setdefault('blank', True)
- ForeignKey.__init__(self, User, **kwargs)
+ ForeignKey.__init__(self, **kwargs)
-
+
class CreatorField(UserField):
""" CreatorField
@@ -46,3 +47,11 @@
value = value.pk
setattr(model_instance, self.attname, value)
return value
+
+try:
+ from south.modelsinspector import add_introspection_rules
+except ImportError:
+ add_introspection_rules = False
+
+if add_introspection_rules:
+ add_introspection_rules([], [r"^threaded_multihost\.fields\.(User|Creator|Editor)Field"])
|
305e88780fc2d3638fb3a9f33bfec8d6c295535e
|
feincms/views/base.py
|
feincms/views/base.py
|
from django.contrib.auth.decorators import permission_required
from django.shortcuts import get_object_or_404, render_to_response
from django.template import RequestContext
from feincms.module.page.models import Page
def build_page_response(page, request):
response = page.setup_request(request)
if response is None:
extra_context = request._feincms_extra_context
response = render_to_response(page.template.path, {
'feincms_page': page,
}, context_instance=RequestContext(request, extra_context))
return response
def handler(request, path=None):
"""
This is the default handler for feincms page content.
"""
if path is None:
path = request.path
page = Page.objects.page_for_path_or_404(path)
response = build_page_response(page, request)
page.finalize_response(request, response)
return response
@permission_required('page.change_page')
def preview_handler(request, page_id):
"""
This handler is for previewing site content; it takes a page_id so
the page is uniquely identified and does not care whether the page
is active or expired. To balance that, it requires a logged in user.
"""
page = get_object_or_404(Page, pk=page_id)
return build_page_response(page, request)
|
from django.contrib.auth.decorators import permission_required
from django.shortcuts import get_object_or_404, render_to_response
from django.template import RequestContext
from feincms.module.page.models import Page
def _build_page_response(page, request):
response = page.setup_request(request)
if response is None:
extra_context = request._feincms_extra_context
response = render_to_response(page.template.path, {
'feincms_page': page,
}, context_instance=RequestContext(request, extra_context))
return response
def build_page_response(page, request):
response = _build_page_response(page, request)
page.finalize_response(request, response)
return response
def handler(request, path=None):
"""
This is the default handler for feincms page content.
"""
if path is None:
path = request.path
page = Page.objects.page_for_path_or_404(path)
response = build_page_response(page, request)
return response
@permission_required('page.change_page')
def preview_handler(request, page_id):
"""
This handler is for previewing site content; it takes a page_id so
the page is uniquely identified and does not care whether the page
is active or expired. To balance that, it requires a logged in user.
"""
page = get_object_or_404(Page, pk=page_id)
return _build_page_response(page, request)
|
Make sure we invoke the response processors even for app content.
|
Make sure we invoke the response processors even for app content.
|
Python
|
bsd-3-clause
|
matthiask/feincms2-content,joshuajonah/feincms,nickburlett/feincms,nickburlett/feincms,hgrimelid/feincms,feincms/feincms,mjl/feincms,mjl/feincms,joshuajonah/feincms,pjdelport/feincms,matthiask/django-content-editor,feincms/feincms,matthiask/feincms2-content,matthiask/django-content-editor,hgrimelid/feincms,pjdelport/feincms,michaelkuty/feincms,joshuajonah/feincms,michaelkuty/feincms,mjl/feincms,feincms/feincms,michaelkuty/feincms,joshuajonah/feincms,michaelkuty/feincms,nickburlett/feincms,matthiask/django-content-editor,matthiask/django-content-editor,pjdelport/feincms,hgrimelid/feincms,matthiask/feincms2-content,nickburlett/feincms
|
---
+++
@@ -5,7 +5,7 @@
from feincms.module.page.models import Page
-def build_page_response(page, request):
+def _build_page_response(page, request):
response = page.setup_request(request)
if response is None:
@@ -16,6 +16,10 @@
return response
+def build_page_response(page, request):
+ response = _build_page_response(page, request)
+ page.finalize_response(request, response)
+ return response
def handler(request, path=None):
"""
@@ -27,7 +31,6 @@
page = Page.objects.page_for_path_or_404(path)
response = build_page_response(page, request)
- page.finalize_response(request, response)
return response
@@ -40,4 +43,4 @@
is active or expired. To balance that, it requires a logged in user.
"""
page = get_object_or_404(Page, pk=page_id)
- return build_page_response(page, request)
+ return _build_page_response(page, request)
|
c7ef639ac9bab4a01e4c8cbafb71bf09d973c355
|
src/adhocracy/lib/auth/shibboleth.py
|
src/adhocracy/lib/auth/shibboleth.py
|
from pylons import config
def get_userbadge_mapping(config=config):
mapping = config.get('adhocracy.shibboleth.userbadge_mapping', u'')
return (line.strip().split(u' ')
for line in mapping.strip().split(u'\n')
if line is not u'')
def _attribute_equals(request, key, value):
"""
exact match
"""
return request.headers.get(key) == value
def _attribute_contains(request, key, value):
"""
contains element
"""
elements = (e.strip() for e in request.headers.get(key).split(','))
return value in elements
def _attribute_contains_substring(request, key, value):
"""
contains substring
"""
return value in request.headers.get(key)
USERBADGE_MAPPERS = {
'attribute_equals': _attribute_equals,
'attribute_contains': _attribute_contains,
'attribute_contains_substring': _attribute_contains_substring,
}
|
from pylons import config
def get_userbadge_mapping(config=config):
mapping = config.get('adhocracy.shibboleth.userbadge_mapping', u'')
return (line.strip().split(u' ')
for line in mapping.strip().split(u'\n')
if line is not u'')
def _attribute_equals(request, key, value):
"""
exact match
"""
return request.headers.get(key) == value
def _attribute_contains(request, key, value):
"""
contains element
"""
elements = (e.strip() for e in request.headers.get(key).split(';'))
return value in elements
def _attribute_contains_substring(request, key, value):
"""
contains substring
"""
return value in request.headers.get(key)
USERBADGE_MAPPERS = {
'attribute_equals': _attribute_equals,
'attribute_contains': _attribute_contains,
'attribute_contains_substring': _attribute_contains_substring,
}
|
Use ; instead of , for Shibboleth multi value fields
|
Use ; instead of , for Shibboleth multi value fields
This corresponds to what mod_shibboleth does.
|
Python
|
agpl-3.0
|
liqd/adhocracy,liqd/adhocracy,phihag/adhocracy,DanielNeugebauer/adhocracy,liqd/adhocracy,alkadis/vcv,phihag/adhocracy,phihag/adhocracy,phihag/adhocracy,SysTheron/adhocracy,alkadis/vcv,DanielNeugebauer/adhocracy,SysTheron/adhocracy,alkadis/vcv,alkadis/vcv,DanielNeugebauer/adhocracy,DanielNeugebauer/adhocracy,liqd/adhocracy,SysTheron/adhocracy,alkadis/vcv,DanielNeugebauer/adhocracy,phihag/adhocracy
|
---
+++
@@ -19,7 +19,7 @@
"""
contains element
"""
- elements = (e.strip() for e in request.headers.get(key).split(','))
+ elements = (e.strip() for e in request.headers.get(key).split(';'))
return value in elements
|
1702fbc607816544c28a8f2895a82d234226e48b
|
euler/p007.py
|
euler/p007.py
|
"""Solution to Project Euler Problem 7
https://projecteuler.net/problem=7
"""
INDEX = 10001
def compute(index=INDEX):
"""Find the `index`th prime number."""
primes = list()
test_number = 2
while len(primes) < index:
for prime in primes:
if test_number % prime == 0:
break
else:
primes.append(test_number)
test_number += 1
return primes[-1]
|
"""Solution to Project Euler Problem 7
https://projecteuler.net/problem=7
"""
INDEX = 10001
def compute(index=INDEX):
"""Find the `index`th prime number."""
primes = [2]
test_number = 3
while len(primes) < index:
for prime in primes:
if test_number % prime == 0:
break
else:
primes.append(test_number)
test_number += 2
return primes[-1]
|
Check only odd numbers after 2
|
P007: Check only odd numbers after 2
|
Python
|
mit
|
2Cubed/ProjectEuler
|
---
+++
@@ -8,8 +8,8 @@
def compute(index=INDEX):
"""Find the `index`th prime number."""
- primes = list()
- test_number = 2
+ primes = [2]
+ test_number = 3
while len(primes) < index:
for prime in primes:
@@ -17,6 +17,6 @@
break
else:
primes.append(test_number)
- test_number += 1
+ test_number += 2
return primes[-1]
|
fc472d043e81c2b5687a0f83dbbdd0dd02b73e35
|
flowtype/commands/exec_flow.py
|
flowtype/commands/exec_flow.py
|
import os
import json
import threading
import subprocess
import sublime
class ExecFlowCommand(threading.Thread):
"""Threaded class used for running flow commands in a different thread.
The subprocess must be threaded so we don't lockup the UI.
"""
def __init__(self, cmd, content):
"""Initialize with the command and the file content to send."""
self.cmd = cmd
self.content = content
self.stdout = None
self.returncode = 0
self.stderr = None
threading.Thread.__init__(self)
def run(self):
"""Execute the command in a subprocess."""
read, write = os.pipe()
os.write(write, str.encode(self.content))
os.close(write)
try:
output = subprocess.check_output(
self.cmd,
shell=sublime.platform() == 'windows',
stdin=read,
stderr=subprocess.STDOUT
)
if type(output) is bytes:
output = output.decode('utf-8')
try:
self.stdout = json.loads(output)
except ValueError:
self.stdout = output
os.close(read)
except subprocess.CalledProcessError as err:
self.stderr = str(err)
self.returncode = 1
|
import os
import json
import threading
import subprocess
import sublime
class ExecFlowCommand(threading.Thread):
"""Threaded class used for running flow commands in a different thread.
The subprocess must be threaded so we don't lockup the UI.
"""
def __init__(self, cmd, content):
"""Initialize with the command and the file content to send."""
self.cmd = cmd
self.content = content
self.stdout = None
self.returncode = 0
self.stderr = None
threading.Thread.__init__(self)
def run(self):
"""Execute the command in a subprocess."""
read, write = os.pipe()
os.write(write, str.encode(self.content))
os.close(write)
try:
output = subprocess.check_output(
self.cmd,
shell=sublime.platform() == 'windows',
stdin=read,
stderr=subprocess.STDOUT
)
if type(output) is bytes:
output = output.decode('utf-8')
try:
self.stdout = json.loads(output)
except ValueError:
self.stdout = output
os.close(read)
except subprocess.CalledProcessError as err:
if type(err.output) is bytes:
output = err.output.decode('utf-8')
else:
output = err.output
self.stderr = str(err) + ': ' + str(output)
self.returncode = 1
|
Add error output to exec error messages
|
Add error output to exec error messages
e.g. for an error like "env: ‘node’: No such file or directory"
the sublime console was only reporting "exited with code 127"
which wasn't very helpful in determining the cause.
|
Python
|
mit
|
Pegase745/sublime-flowtype
|
---
+++
@@ -46,5 +46,9 @@
os.close(read)
except subprocess.CalledProcessError as err:
- self.stderr = str(err)
+ if type(err.output) is bytes:
+ output = err.output.decode('utf-8')
+ else:
+ output = err.output
+ self.stderr = str(err) + ': ' + str(output)
self.returncode = 1
|
3aa13efa28b4ededa465541a7db8df5fc5878ce3
|
tempora/tests/test_timing.py
|
tempora/tests/test_timing.py
|
import datetime
import time
import contextlib
import os
from unittest import mock
from tempora import timing
def test_IntervalGovernor():
"""
IntervalGovernor should prevent a function from being called more than
once per interval.
"""
func_under_test = mock.MagicMock()
# to look like a function, it needs a __name__ attribute
func_under_test.__name__ = 'func_under_test'
interval = datetime.timedelta(seconds=1)
governed = timing.IntervalGovernor(interval)(func_under_test)
governed('a')
governed('b')
governed(3, 'sir')
func_under_test.assert_called_once_with('a')
@contextlib.contextmanager
def change(alt_tz, monkeypatch):
monkeypatch.setitem(os.environ, 'TZ', alt_tz)
time.tzset()
try:
yield
finally:
monkeypatch.delitem(os.environ, 'TZ')
time.tzset()
def test_Stopwatch_timezone_change(monkeypatch):
"""
The stopwatch should provide a consistent duration even
if the timezone changes.
"""
watch = timing.Stopwatch()
with change('AEST-10AEDT-11,M10.5.0,M3.5.0', monkeypatch):
assert abs(watch.split().total_seconds()) < 0.1
|
import datetime
import time
import contextlib
import os
from unittest import mock
import pytest
from tempora import timing
def test_IntervalGovernor():
"""
IntervalGovernor should prevent a function from being called more than
once per interval.
"""
func_under_test = mock.MagicMock()
# to look like a function, it needs a __name__ attribute
func_under_test.__name__ = 'func_under_test'
interval = datetime.timedelta(seconds=1)
governed = timing.IntervalGovernor(interval)(func_under_test)
governed('a')
governed('b')
governed(3, 'sir')
func_under_test.assert_called_once_with('a')
@pytest.fixture
def alt_tz(monkeypatch):
if not hasattr(time, 'tzset'):
pytest.skip("tzset not available")
@contextlib.contextmanager
def change():
val = 'AEST-10AEDT-11,M10.5.0,M3.5.0'
with monkeypatch.context() as ctx:
ctx.setitem(os.environ, 'TZ', val)
time.tzset()
yield
time.tzset()
return change()
def test_Stopwatch_timezone_change(alt_tz):
"""
The stopwatch should provide a consistent duration even
if the timezone changes.
"""
watch = timing.Stopwatch()
with alt_tz:
assert abs(watch.split().total_seconds()) < 0.1
|
Rewrite alt_tz as proper fixture. Skip when tzset isn't available.
|
Rewrite alt_tz as proper fixture. Skip when tzset isn't available.
|
Python
|
mit
|
jaraco/tempora
|
---
+++
@@ -4,6 +4,7 @@
import os
from unittest import mock
+import pytest
from tempora import timing
@@ -23,22 +24,28 @@
func_under_test.assert_called_once_with('a')
-@contextlib.contextmanager
-def change(alt_tz, monkeypatch):
- monkeypatch.setitem(os.environ, 'TZ', alt_tz)
- time.tzset()
- try:
- yield
- finally:
- monkeypatch.delitem(os.environ, 'TZ')
+@pytest.fixture
+def alt_tz(monkeypatch):
+ if not hasattr(time, 'tzset'):
+ pytest.skip("tzset not available")
+
+ @contextlib.contextmanager
+ def change():
+ val = 'AEST-10AEDT-11,M10.5.0,M3.5.0'
+ with monkeypatch.context() as ctx:
+ ctx.setitem(os.environ, 'TZ', val)
+ time.tzset()
+ yield
time.tzset()
+ return change()
-def test_Stopwatch_timezone_change(monkeypatch):
+
+def test_Stopwatch_timezone_change(alt_tz):
"""
The stopwatch should provide a consistent duration even
if the timezone changes.
"""
watch = timing.Stopwatch()
- with change('AEST-10AEDT-11,M10.5.0,M3.5.0', monkeypatch):
+ with alt_tz:
assert abs(watch.split().total_seconds()) < 0.1
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.