commit
stringlengths
40
40
old_file
stringlengths
4
150
new_file
stringlengths
4
150
old_contents
stringlengths
0
3.26k
new_contents
stringlengths
1
4.43k
subject
stringlengths
15
501
message
stringlengths
15
4.06k
lang
stringclasses
4 values
license
stringclasses
13 values
repos
stringlengths
5
91.5k
diff
stringlengths
0
4.35k
91ad56ea892d2f2fdb2af97f81ec70a7b9f9305c
analysis/sanity-check-velocity.py
analysis/sanity-check-velocity.py
#!/usr/bin/env python import climate import joblib import lmj.cubes import numpy as np def _check(t): t.load() t.add_velocities(smooth=0) vel = abs(t.df[t.marker_velocity_columns].values).flatten() vel = vel[np.isfinite(vel)] pct = np.percentile(vel, [1, 2, 5, 10, 20, 50, 80, 90, 95, 98, 99]) print(t.subject.key, t.block.key, t.key, *pct) def main(root): trials = lmj.cubes.Experiment(root).trials_matching('*') check = joblib.delayed(_check) joblib.Parallel(-1)(check(t) for t in trials) if __name__ == '__main__': climate.call(main)
#!/usr/bin/env python import climate import joblib import lmj.cubes import numpy as np def _check(t): t.load() t.add_velocities(smooth=0) t.add_accelerations(smooth=0) vel = abs(t.df[t.marker_velocity_columns].values).flatten() vel = vel[np.isfinite(vel)] pct = np.percentile(vel, [1, 2, 5, 10, 20, 50, 80, 90, 95, 98, 99]) np.set_printoptions(suppress=True, linewidth=1000, precision=2) t.log('%s', pct) def main(root): trials = lmj.cubes.Experiment(root).trials_matching('*') check = joblib.delayed(_check) joblib.Parallel(-1)(check(t) for t in trials) if __name__ == '__main__': climate.call(main)
Use trial logging. Tweak numpy logging output.
Use trial logging. Tweak numpy logging output.
Python
mit
lmjohns3/cube-experiment,lmjohns3/cube-experiment,lmjohns3/cube-experiment
--- +++ @@ -9,10 +9,12 @@ def _check(t): t.load() t.add_velocities(smooth=0) + t.add_accelerations(smooth=0) vel = abs(t.df[t.marker_velocity_columns].values).flatten() vel = vel[np.isfinite(vel)] pct = np.percentile(vel, [1, 2, 5, 10, 20, 50, 80, 90, 95, 98, 99]) - print(t.subject.key, t.block.key, t.key, *pct) + np.set_printoptions(suppress=True, linewidth=1000, precision=2) + t.log('%s', pct) def main(root):
0cccd467ac4c0bd8b8110fcfe47f81d73a238aa9
plugins/uptime.py
plugins/uptime.py
import time class Plugin: def __init__(self, vk_bot): self.vk_bot = vk_bot self.vk_bot.add_command('uptime', self.uptime) async def uptime(self, vk_api, sender, message): await self.vk_bot.send_message(sender, 'Total uptime: {} seconds'.format(round(time.time() - self.vk_bot.start_time)))
import time class Plugin: def __init__(self, vk_bot): self.vk_bot = vk_bot self.vk_bot.add_command('uptime', self.uptime) async def uptime(self, vk_api, sender, message): await self.vk_bot.send_message(sender, 'Total uptime: {} seconds'.format(round(time.time() - self.vk_bot.start_time)))
Fix code string length (PEP8)
Fix code string length (PEP8)
Python
mit
roman901/vk_bot
--- +++ @@ -7,4 +7,5 @@ self.vk_bot.add_command('uptime', self.uptime) async def uptime(self, vk_api, sender, message): - await self.vk_bot.send_message(sender, 'Total uptime: {} seconds'.format(round(time.time() - self.vk_bot.start_time))) + await self.vk_bot.send_message(sender, + 'Total uptime: {} seconds'.format(round(time.time() - self.vk_bot.start_time)))
e26434ee69545b8c16b62ebd78e5bec0c95d579a
lib/rapidsms/webui/urls.py
lib/rapidsms/webui/urls.py
#!/usr/bin/env python # vim: ai ts=4 sts=4 et sw=4 import os urlpatterns = [] # load the rapidsms configuration from rapidsms.config import Config conf = Config(os.environ["RAPIDSMS_INI"]) # iterate each of the active rapidsms apps (from the ini), # and (attempt to) import the urls.py from each. it's okay # if this fails, since not all apps have a webui for rs_app in conf["rapidsms"]["apps"]: try: package_name = "apps.%s.urls" % (rs_app["type"]) module = __import__(package_name, {}, {}, ["urlpatterns"]) urlpatterns += module.urlpatterns except Exception, e: continue
#!/usr/bin/env python # vim: ai ts=4 sts=4 et sw=4 import os, sys urlpatterns = [] loaded = [] # load the rapidsms configuration from rapidsms.config import Config conf = Config(os.environ["RAPIDSMS_INI"]) # iterate each of the active rapidsms apps (from the ini), # and (attempt to) import the urls.py from each. it's okay # if this fails, since not all apps have a webui for rs_app in conf["rapidsms"]["apps"]: try: package_name = "apps.%s.urls" % (rs_app["type"]) module = __import__(package_name, {}, {}, ["urlpatterns"]) urlpatterns += module.urlpatterns loaded += [rs_app["type"]] except Exception, e: continue print >>sys.stderr, "Loaded url patterns from %s" % ", ".join(loaded)
Print a list of which URLs got loaded. This doesn't help that much when trying to debug errors that keep URLs from getting loaded. But it's a start.
Print a list of which URLs got loaded. This doesn't help that much when trying to debug errors that keep URLs from getting loaded. But it's a start.
Python
bsd-3-clause
rapidsms/rapidsms-legacy,rapidsms/rapidsms-legacy,rapidsms/rapidsms-legacy
--- +++ @@ -1,16 +1,14 @@ #!/usr/bin/env python # vim: ai ts=4 sts=4 et sw=4 -import os - +import os, sys urlpatterns = [] - +loaded = [] # load the rapidsms configuration from rapidsms.config import Config conf = Config(os.environ["RAPIDSMS_INI"]) - # iterate each of the active rapidsms apps (from the ini), # and (attempt to) import the urls.py from each. it's okay @@ -21,6 +19,8 @@ package_name = "apps.%s.urls" % (rs_app["type"]) module = __import__(package_name, {}, {}, ["urlpatterns"]) urlpatterns += module.urlpatterns + loaded += [rs_app["type"]] except Exception, e: continue +print >>sys.stderr, "Loaded url patterns from %s" % ", ".join(loaded)
21e02ee04ef359abfcaacf35375176b296ae7ea1
tests/__init__.py
tests/__init__.py
from unittest import TestCase from redash import settings, db, app import redash.models # TODO: this isn't pretty... settings.DATABASE_CONFIG = { 'name': 'circle_test', 'engine': 'peewee.PostgresqlDatabase', 'threadlocals': True } app.config['DATABASE'] = settings.DATABASE_CONFIG db.load_database() for model in redash.models.all_models: model._meta.database = db.database class BaseTestCase(TestCase): def setUp(self): redash.models.create_db(True, True) def tearDown(self): db.close_db(None) redash.models.create_db(False, True)
import logging from unittest import TestCase from redash import settings, db, app import redash.models # TODO: this isn't pretty... settings.DATABASE_CONFIG = { 'name': 'circle_test', 'engine': 'peewee.PostgresqlDatabase', 'threadlocals': True } app.config['DATABASE'] = settings.DATABASE_CONFIG db.load_database() logging.getLogger('peewee').setLevel(logging.INFO) for model in redash.models.all_models: model._meta.database = db.database class BaseTestCase(TestCase): def setUp(self): redash.models.create_db(True, True) def tearDown(self): db.close_db(None) redash.models.create_db(False, True)
Reduce Peewee's logging level to INFO in tests.
Reduce Peewee's logging level to INFO in tests.
Python
bsd-2-clause
rockwotj/redash,crowdworks/redash,pubnative/redash,stefanseifert/redash,pubnative/redash,akariv/redash,amino-data/redash,jmvasquez/redashtest,ninneko/redash,useabode/redash,stefanseifert/redash,crowdworks/redash,pubnative/redash,stefanseifert/redash,stefanseifert/redash,getredash/redash,amino-data/redash,M32Media/redash,guaguadev/redash,akariv/redash,useabode/redash,rockwotj/redash,ninneko/redash,jmvasquez/redashtest,M32Media/redash,imsally/redash,guaguadev/redash,easytaxibr/redash,denisov-vlad/redash,guaguadev/redash,ninneko/redash,moritz9/redash,getredash/redash,chriszs/redash,easytaxibr/redash,hudl/redash,ninneko/redash,akariv/redash,jmvasquez/redashtest,jmvasquez/redashtest,imsally/redash,useabode/redash,crowdworks/redash,denisov-vlad/redash,ninneko/redash,getredash/redash,amino-data/redash,alexanderlz/redash,moritz9/redash,denisov-vlad/redash,akariv/redash,denisov-vlad/redash,EverlyWell/redash,chriszs/redash,denisov-vlad/redash,guaguadev/redash,pubnative/redash,EverlyWell/redash,44px/redash,chriszs/redash,M32Media/redash,hudl/redash,amino-data/redash,moritz9/redash,imsally/redash,vishesh92/redash,hudl/redash,44px/redash,vishesh92/redash,alexanderlz/redash,44px/redash,stefanseifert/redash,moritz9/redash,easytaxibr/redash,guaguadev/redash,jmvasquez/redashtest,pubnative/redash,useabode/redash,rockwotj/redash,44px/redash,EverlyWell/redash,easytaxibr/redash,vishesh92/redash,vishesh92/redash,rockwotj/redash,akariv/redash,EverlyWell/redash,chriszs/redash,alexanderlz/redash,crowdworks/redash,easytaxibr/redash,hudl/redash,getredash/redash,imsally/redash,alexanderlz/redash,M32Media/redash,getredash/redash
--- +++ @@ -1,3 +1,4 @@ +import logging from unittest import TestCase from redash import settings, db, app import redash.models @@ -11,6 +12,8 @@ app.config['DATABASE'] = settings.DATABASE_CONFIG db.load_database() +logging.getLogger('peewee').setLevel(logging.INFO) + for model in redash.models.all_models: model._meta.database = db.database
48af7d169bac32898763af671f3a30170b85d2cd
tests/__main__.py
tests/__main__.py
import unittest if __name__ == '__main__': all_tests = unittest.TestLoader().discover('./', pattern='*_tests.py') unittest.TextTestRunner().run(all_tests)
import sys import unittest if __name__ == '__main__': all_tests = unittest.TestLoader().discover('./', pattern='*_tests.py') ret = unittest.TextTestRunner().run(all_tests) sys.exit(not ret.wasSuccessful())
Fix an issue when unit tests always return 0 status.
Fix an issue when unit tests always return 0 status.
Python
mit
sergeymironov0001/twitch-chat-bot
--- +++ @@ -1,5 +1,7 @@ +import sys import unittest if __name__ == '__main__': all_tests = unittest.TestLoader().discover('./', pattern='*_tests.py') - unittest.TextTestRunner().run(all_tests) + ret = unittest.TextTestRunner().run(all_tests) + sys.exit(not ret.wasSuccessful())
9dc35ebafb3e33c3736c8d58a8cb2353695ddedb
tests/settings.py
tests/settings.py
DATABASES = {"default": {"ENGINE": "django.db.backends.sqlite3"}} SECRET_KEY = "secrekey" INSTALLED_APPS = ["phonenumber_field", "tests"]
DATABASES = {"default": {"ENGINE": "django.db.backends.sqlite3"}} DEFAULT_AUTO_FIELD = "django.db.models.BigAutoField" SECRET_KEY = "secrekey" INSTALLED_APPS = ["phonenumber_field", "tests"]
Set DEFAULT_AUTO_FIELD for the test project
Set DEFAULT_AUTO_FIELD for the test project https://docs.djangoproject.com/en/dev/releases/3.2/#customizing-type-of-auto-created-primary-keys Avoid warnings on Django master: ``` tests.TestModelPhoneNU: (models.W042) Auto-created primary key used when not defining a primary key type, by default 'django.db.models.AutoField'. HINT: Configure the DEFAULT_AUTO_FIELD setting or the AppConfig.default_auto_field attribute to point to a s ubclass of AutoField, e.g. 'django.db.models.BigAutoField ```
Python
mit
stefanfoulis/django-phonenumber-field
--- +++ @@ -1,4 +1,5 @@ DATABASES = {"default": {"ENGINE": "django.db.backends.sqlite3"}} +DEFAULT_AUTO_FIELD = "django.db.models.BigAutoField" SECRET_KEY = "secrekey"
25478444e1ec5b4b1c9f811fea7fe0b401f14514
lingcod/bookmarks/forms.py
lingcod/bookmarks/forms.py
from lingcod.features.forms import FeatureForm from lingcod.bookmarks.models import Bookmark from django import forms class BookmarkForm(FeatureForm): name = forms.CharField(label='Bookmark Name') latitude = forms.FloatField(widget=forms.HiddenInput()) longitude = forms.FloatField(widget=forms.HiddenInput()) altitude = forms.FloatField(widget=forms.HiddenInput()) heading = forms.FloatField(widget=forms.HiddenInput()) tilt = forms.FloatField(widget=forms.HiddenInput()) roll = forms.FloatField(widget=forms.HiddenInput()) altitudeMode = forms.FloatField(widget=forms.HiddenInput()) publicstate = forms.CharField(widget=forms.HiddenInput()) ip = forms.CharField(widget=forms.HiddenInput()) class Meta(FeatureForm.Meta): model = Bookmark
from lingcod.features.forms import FeatureForm from lingcod.bookmarks.models import Bookmark from django import forms class BookmarkForm(FeatureForm): name = forms.CharField(label='Bookmark Name') latitude = forms.FloatField(widget=forms.HiddenInput()) longitude = forms.FloatField(widget=forms.HiddenInput()) altitude = forms.FloatField(widget=forms.HiddenInput()) heading = forms.FloatField(widget=forms.HiddenInput()) tilt = forms.FloatField(widget=forms.HiddenInput()) roll = forms.FloatField(widget=forms.HiddenInput()) altitudeMode = forms.FloatField(widget=forms.HiddenInput()) publicstate = forms.CharField(widget=forms.HiddenInput()) ip = forms.CharField(widget=forms.HiddenInput(), required=False) class Meta(FeatureForm.Meta): model = Bookmark
Allow IP to be blank in form
Allow IP to be blank in form
Python
bsd-3-clause
Alwnikrotikz/marinemap,google-code-export/marinemap,google-code-export/marinemap,Alwnikrotikz/marinemap,google-code-export/marinemap,google-code-export/marinemap,Alwnikrotikz/marinemap,Alwnikrotikz/marinemap
--- +++ @@ -12,6 +12,6 @@ roll = forms.FloatField(widget=forms.HiddenInput()) altitudeMode = forms.FloatField(widget=forms.HiddenInput()) publicstate = forms.CharField(widget=forms.HiddenInput()) - ip = forms.CharField(widget=forms.HiddenInput()) + ip = forms.CharField(widget=forms.HiddenInput(), required=False) class Meta(FeatureForm.Meta): model = Bookmark
9217bfc6bab0d152e33d9fda60218c404b61d064
cmd2/__init__.py
cmd2/__init__.py
# # -*- coding: utf-8 -*- from .cmd2 import __version__, Cmd, CmdResult, Statement, categorize from .cmd2 import with_argument_list, with_argparser, with_argparser_and_unknown_args, with_category
# # -*- coding: utf-8 -*- from .cmd2 import __version__, Cmd, CmdResult, Statement, EmptyStatement, categorize from .cmd2 import with_argument_list, with_argparser, with_argparser_and_unknown_args, with_category
Add EmptyStatement exception to default imports
Add EmptyStatement exception to default imports
Python
mit
python-cmd2/cmd2,python-cmd2/cmd2
--- +++ @@ -1,4 +1,4 @@ # # -*- coding: utf-8 -*- -from .cmd2 import __version__, Cmd, CmdResult, Statement, categorize +from .cmd2 import __version__, Cmd, CmdResult, Statement, EmptyStatement, categorize from .cmd2 import with_argument_list, with_argparser, with_argparser_and_unknown_args, with_category
e2126518957d0e3e360af3f80b1657bde9053b23
capstone/game/players/alphabeta.py
capstone/game/players/alphabeta.py
import random from ..player import Player from ..utils import utility class AlphaBeta(Player): name = 'Alpha-Beta' def __init__(self, eval_func=utility, max_depth=1000): self._eval = eval_func self._max_depth = max_depth def __str__(self): return self.name def __repr__(self): return self.name def _ab(self, game, cur_depth, alpha, beta): if game.is_over() or cur_depth == self._max_depth: return None, self._eval(game, game.cur_player()) best_move = None best_score = -100000000 for move in game.legal_moves(): _, score = self._ab(game=game.copy().make_move(move), cur_depth=cur_depth + 1, alpha=-beta, beta=-max(alpha, best_score)) score = -score if score > best_score: best_score = score best_move = move if best_score >= beta: return best_move, best_score return best_move, best_score ########## # Player # ########## def choose_move(self, game): move, _ = self._ab(game, cur_depth=0, alpha=-100000000, beta=100000000) return move
import random import numpy as np from ..player import Player from ..utils import utility class AlphaBeta(Player): name = 'Alpha-Beta' def __init__(self, eval_func=utility, max_depth=np.inf): self._eval = eval_func self._max_depth = max_depth def __str__(self): return self.name def __repr__(self): return self.name def _ab(self, game, cur_depth, alpha, beta): if game.is_over() or cur_depth == self._max_depth: return None, self._eval(game, game.cur_player()) best_move = None best_score = -np.inf for move in game.legal_moves(): _, score = self._ab(game=game.copy().make_move(move), cur_depth=cur_depth + 1, alpha=-beta, beta=-max(alpha, best_score)) score = -score if score > best_score: best_score = score best_move = move if best_score >= beta: return best_move, best_score return best_move, best_score ########## # Player # ########## def choose_move(self, game): move, _ = self._ab(game, cur_depth=0, alpha=-np.inf, beta=np.inf) return move
Use np.inf for max/min limit values
Use np.inf for max/min limit values
Python
mit
davidrobles/mlnd-capstone-code
--- +++ @@ -1,4 +1,5 @@ import random +import numpy as np from ..player import Player from ..utils import utility @@ -7,7 +8,7 @@ name = 'Alpha-Beta' - def __init__(self, eval_func=utility, max_depth=1000): + def __init__(self, eval_func=utility, max_depth=np.inf): self._eval = eval_func self._max_depth = max_depth @@ -21,7 +22,7 @@ if game.is_over() or cur_depth == self._max_depth: return None, self._eval(game, game.cur_player()) best_move = None - best_score = -100000000 + best_score = -np.inf for move in game.legal_moves(): _, score = self._ab(game=game.copy().make_move(move), cur_depth=cur_depth + 1, @@ -40,5 +41,5 @@ ########## def choose_move(self, game): - move, _ = self._ab(game, cur_depth=0, alpha=-100000000, beta=100000000) + move, _ = self._ab(game, cur_depth=0, alpha=-np.inf, beta=np.inf) return move
a17c2ce30f30d0441b1475457b0bc9d04da9f143
coil/__init__.py
coil/__init__.py
"""Coil: A Configuration Library.""" __version__ = "0.2.2"
"""Coil: A Configuration Library.""" __version__ = "0.3.0" from coil.parser import Parser def parse_file(file_name): """Open and parse a coil file. Returns the root Struct. """ coil = open(file_name) return Parser(coil, file_name).root() def parse(string): """Parse a coil string. Returns the root Struct. """ return Parser(string.splitlines()).root()
Add helpers for parsing files and strings
Add helpers for parsing files and strings
Python
mit
tectronics/coil,marineam/coil,kovacsbalu/coil,kovacsbalu/coil,marineam/coil,tectronics/coil
--- +++ @@ -1,3 +1,20 @@ """Coil: A Configuration Library.""" -__version__ = "0.2.2" +__version__ = "0.3.0" + +from coil.parser import Parser + +def parse_file(file_name): + """Open and parse a coil file. + + Returns the root Struct. + """ + coil = open(file_name) + return Parser(coil, file_name).root() + +def parse(string): + """Parse a coil string. + + Returns the root Struct. + """ + return Parser(string.splitlines()).root()
f664609d579e7b709945756def90092f0814998e
libpb/__init__.py
libpb/__init__.py
"""FreeBSD port building infrastructure.""" from __future__ import absolute_import from . import event def stop(kill=False, kill_clean=False): """Stop building ports and cleanup.""" from os import killpg from signal import SIGTERM, SIGKILL from .builder import builders from .env import cpus, flags from .queue import attr_queue, clean_queue, queues if flags["no_op"]: raise SystemExit(254) flags["mode"] = "clean" # Stop all queues attr_queue.load = 0 for queue in queues: queue.load = 0 # Make cleaning go a bit faster if kill_clean: clean_queue.load = 0 return else: clean_queue.load = cpus # Wait for all active ports to finish so that they may be cleaned active = set() for queue in queues: for job in queue.active: port = job.port active.add(port) port.stage_completed.connect(lambda x: x.clean()) # Clean all other outstanding ports for builder in builders: for port in builder.ports: if port not in active: port.clean()
"""FreeBSD port building infrastructure.""" from __future__ import absolute_import from . import event def stop(kill=False, kill_clean=False): """Stop building ports and cleanup.""" from os import kill, killpg from signal import SIGTERM, SIGKILL from .builder import builders from .env import cpus, flags from .queue import attr_queue, clean_queue, queues if flags["no_op"]: raise SystemExit(254) flags["mode"] = "clean" kill_queues = (attr_queue,) + queues if kill_clean: kill_queues += (clean_queue,) # Kill all active jobs for queue in kill_queues: for pid in (job.pid for job in queue.active if job.pid): try: if kill: killpg(pid, SIGKILL) else: kill(pid, SIGTERM) except OSError: pass # Stop all queues attr_queue.load = 0 for queue in queues: queue.load = 0 # Make cleaning go a bit faster if kill_clean: clean_queue.load = 0 return else: clean_queue.load = cpus # Wait for all active ports to finish so that they may be cleaned active = set() for queue in queues: for job in queue.active: port = job.port active.add(port) port.stage_completed.connect(lambda x: x.clean()) # Clean all other outstanding ports for builder in builders: for port in builder.ports: if port not in active: port.clean()
Send SIGTERM and SIGKILL to child processes.
Send SIGTERM and SIGKILL to child processes. With the removal of subprocess there was no way to known what were the subprocesses, however after the introduction of Jobs tracking the PIDs it is now possible. Use those PIDs.
Python
bsd-2-clause
DragonSA/portbuilder,DragonSA/portbuilder
--- +++ @@ -6,7 +6,7 @@ def stop(kill=False, kill_clean=False): """Stop building ports and cleanup.""" - from os import killpg + from os import kill, killpg from signal import SIGTERM, SIGKILL from .builder import builders from .env import cpus, flags @@ -16,6 +16,21 @@ raise SystemExit(254) flags["mode"] = "clean" + + kill_queues = (attr_queue,) + queues + if kill_clean: + kill_queues += (clean_queue,) + + # Kill all active jobs + for queue in kill_queues: + for pid in (job.pid for job in queue.active if job.pid): + try: + if kill: + killpg(pid, SIGKILL) + else: + kill(pid, SIGTERM) + except OSError: + pass # Stop all queues attr_queue.load = 0
605339144c61c4860f1dc7dec5fc5a0ff959600f
company/forms.py
company/forms.py
from django import forms from . import models from pola.forms import (CommitDescriptionMixin, FormHorizontalMixin, SaveButtonMixin, ReadOnlyFieldsMixin) class CompanyForm(ReadOnlyFieldsMixin, SaveButtonMixin, FormHorizontalMixin, CommitDescriptionMixin, forms.ModelForm): readonly_fields = [ 'name' ] class Meta: model = models.Company fields = [ 'nip', 'name', 'official_name', 'address', 'plCapital', 'plCapital_notes', 'plTaxes', 'plTaxes_notes', 'plRnD', 'plRnD_notes', 'plWorkers', 'plWorkers_notes', 'plBrand', 'plBrand_notes', 'verified', ]
from django import forms from . import models from pola.forms import (CommitDescriptionMixin, FormHorizontalMixin, SaveButtonMixin, ReadOnlyFieldsMixin) class CompanyForm(ReadOnlyFieldsMixin, SaveButtonMixin, FormHorizontalMixin, CommitDescriptionMixin, forms.ModelForm): readonly_fields = [ 'name' ] class Meta: model = models.Company fields = [ 'nip', 'name', 'official_name', 'common_name', 'address', 'plCapital', 'plCapital_notes', 'plTaxes', 'plTaxes_notes', 'plRnD', 'plRnD_notes', 'plWorkers', 'plWorkers_notes', 'plBrand', 'plBrand_notes', 'verified', ]
Add 'common_name' to company's form
Add 'common_name' to company's form
Python
bsd-3-clause
KlubJagiellonski/pola-backend,KlubJagiellonski/pola-backend,KlubJagiellonski/pola-backend,KlubJagiellonski/pola-backend
--- +++ @@ -18,6 +18,7 @@ 'nip', 'name', 'official_name', + 'common_name', 'address', 'plCapital', 'plCapital_notes',
559fa4bf1982de6dd4a8943939b535972731bd08
comrade/core/context_processors.py
comrade/core/context_processors.py
from django.conf import settings from django.contrib.sites.models import Site from settings import DeploymentType def default(request): context = {} context['DEPLOYMENT'] = settings.DEPLOYMENT context['site'] = Site.objects.get_current() if settings.DEPLOYMENT != DeploymentType.PRODUCTION: context['GIT_COMMIT'] = settings.GIT_COMMIT context['site_email'] = settings.CONTACT_EMAIL return context def ssl_media(request): if request.is_secure(): ssl_media_url = settings.MEDIA_URL.replace('http://','https://') else: ssl_media_url = settings.MEDIA_URL return {'MEDIA_URL': ssl_media_url}
from django.conf import settings from django.contrib.sites.models import Site from settings import DeploymentType def default(request): context = {} context['DEPLOYMENT'] = settings.DEPLOYMENT context['current_site'] = Site.objects.get_current() if settings.DEPLOYMENT != DeploymentType.PRODUCTION: context['GIT_COMMIT'] = settings.GIT_COMMIT context['site_email'] = settings.CONTACT_EMAIL if request.is_secure(): context['protocol'] = 'https://' else: context['protocol'] = 'http://' context['current_site_url'] = (context['protocol'] + context['current_site'].domain) return context def ssl_media(request): if request.is_secure(): ssl_media_url = settings.MEDIA_URL.replace('http://','https://') else: ssl_media_url = settings.MEDIA_URL return {'MEDIA_URL': ssl_media_url}
Add full base URL for site to default context.
Add full base URL for site to default context.
Python
mit
bueda/django-comrade
--- +++ @@ -5,10 +5,16 @@ def default(request): context = {} context['DEPLOYMENT'] = settings.DEPLOYMENT - context['site'] = Site.objects.get_current() + context['current_site'] = Site.objects.get_current() if settings.DEPLOYMENT != DeploymentType.PRODUCTION: context['GIT_COMMIT'] = settings.GIT_COMMIT context['site_email'] = settings.CONTACT_EMAIL + if request.is_secure(): + context['protocol'] = 'https://' + else: + context['protocol'] = 'http://' + context['current_site_url'] = (context['protocol'] + + context['current_site'].domain) return context def ssl_media(request):
1b97aa2dae43a8988802ca532a3200f444f85db3
markups/common.py
markups/common.py
# This file is part of python-markups module # License: BSD # Copyright: (C) Dmitry Shachnev, 2012 import os.path # Some common constants and functions (LANGUAGE_HOME_PAGE, MODULE_HOME_PAGE, SYNTAX_DOCUMENTATION) = range(3) CONFIGURATION_DIR = (os.environ.get('XDG_CONFIG_HOME') or os.path.expanduser('~/.config')) MATHJAX_LOCAL_URL = 'file:///usr/share/javascript/mathjax/MathJax.js' MATHJAX_WEB_URL = 'http://cdn.mathjax.org/mathjax/latest/MathJax.js' def get_pygments_stylesheet(selector): try: from pygments.formatters import HtmlFormatter except ImportError: return '' else: return HtmlFormatter().get_style_defs(selector) + '\n' def get_mathjax_url(webenv): if os.path.exists(MATHJAX_LOCAL_URL[7:]) and not webenv: return MATHJAX_LOCAL_URL else: return MATHJAX_WEB_URL
# This file is part of python-markups module # License: BSD # Copyright: (C) Dmitry Shachnev, 2012 import os.path # Some common constants and functions (LANGUAGE_HOME_PAGE, MODULE_HOME_PAGE, SYNTAX_DOCUMENTATION) = range(3) CONFIGURATION_DIR = (os.environ.get('XDG_CONFIG_HOME') or os.path.expanduser('~/.config')) MATHJAX_LOCAL_URL = 'file:///usr/share/javascript/mathjax/MathJax.js' MATHJAX_WEB_URL = 'http://cdn.mathjax.org/mathjax/latest/MathJax.js' PYGMENTS_STYLE = 'default' def get_pygments_stylesheet(selector, style=None): try: from pygments.formatters import HtmlFormatter except ImportError: return '' else: return HtmlFormatter(style=(style or PYGMENTS_STYLE)).get_style_defs(selector) + '\n' def get_mathjax_url(webenv): if os.path.exists(MATHJAX_LOCAL_URL[7:]) and not webenv: return MATHJAX_LOCAL_URL else: return MATHJAX_WEB_URL
Add initial support for pygments styles
Add initial support for pygments styles
Python
bsd-3-clause
retext-project/pymarkups,mitya57/pymarkups
--- +++ @@ -11,13 +11,15 @@ MATHJAX_LOCAL_URL = 'file:///usr/share/javascript/mathjax/MathJax.js' MATHJAX_WEB_URL = 'http://cdn.mathjax.org/mathjax/latest/MathJax.js' -def get_pygments_stylesheet(selector): +PYGMENTS_STYLE = 'default' + +def get_pygments_stylesheet(selector, style=None): try: from pygments.formatters import HtmlFormatter except ImportError: return '' else: - return HtmlFormatter().get_style_defs(selector) + '\n' + return HtmlFormatter(style=(style or PYGMENTS_STYLE)).get_style_defs(selector) + '\n' def get_mathjax_url(webenv): if os.path.exists(MATHJAX_LOCAL_URL[7:]) and not webenv:
05cb698d45ce4e33e2f4bfdc38f9633083a284a7
test_project/project_specific/generic_channel_example.py
test_project/project_specific/generic_channel_example.py
import autocomplete_light from models import Contact, Address class MyGenericChannel(autocomplete_light.GenericChannelBase): def get_querysets(self): return { Contact: Contact.objects.all(), Address: Address.objects.all(), } def order_results(self, results): if results.model == Address: return results.order_by('street') elif results.model == Contact: return results.order_by('name') autocomplete_light.register(MyGenericChannel)
import autocomplete_light from models import Contact, Address class MyGenericChannel(autocomplete_light.GenericChannelBase): def get_querysets(self): return { Contact: Contact.objects.all(), Address: Address.objects.all(), } def order_results(self, results): if results.model == Address: return results.order_by('street') elif results.model == Contact: return results.order_by('name') def query_filter(self, results): q = self.request.GET.get('q', None) if q: if results.model == Address: results = results.filter(street__icontains=q) elif results.model == Contact: results = results.filter(name__icontains=q) return results autocomplete_light.register(MyGenericChannel)
Implement query_filter for MyGenericChannel, because it should search by something other than search_name in the case of Address
Implement query_filter for MyGenericChannel, because it should search by something other than search_name in the case of Address
Python
mit
Eraldo/django-autocomplete-light,spookylukey/django-autocomplete-light,Perkville/django-autocomplete-light,jonashaag/django-autocomplete-light,yourlabs/django-autocomplete-light,shubhamdipt/django-autocomplete-light,Perkville/django-autocomplete-light,Visgean/django-autocomplete-light,dsanders11/django-autocomplete-light,Visgean/django-autocomplete-light,jonashaag/django-autocomplete-light,Eraldo/django-autocomplete-light,spookylukey/django-autocomplete-light,dsanders11/django-autocomplete-light,yourlabs/django-autocomplete-light,blueyed/django-autocomplete-light,shubhamdipt/django-autocomplete-light,luzfcb/django-autocomplete-light,dsanders11/django-autocomplete-light,Visgean/django-autocomplete-light,shubhamdipt/django-autocomplete-light,yourlabs/django-autocomplete-light,jonashaag/django-autocomplete-light,Perkville/django-autocomplete-light,Perkville/django-autocomplete-light,yourlabs/django-autocomplete-light,shubhamdipt/django-autocomplete-light,luzfcb/django-autocomplete-light,Visgean/django-autocomplete-light,luzfcb/django-autocomplete-light,luzfcb/django-autocomplete-light,Eraldo/django-autocomplete-light,blueyed/django-autocomplete-light,blueyed/django-autocomplete-light,dsanders11/django-autocomplete-light,spookylukey/django-autocomplete-light,Eraldo/django-autocomplete-light
--- +++ @@ -15,4 +15,15 @@ elif results.model == Contact: return results.order_by('name') + def query_filter(self, results): + q = self.request.GET.get('q', None) + + if q: + if results.model == Address: + results = results.filter(street__icontains=q) + elif results.model == Contact: + results = results.filter(name__icontains=q) + + return results + autocomplete_light.register(MyGenericChannel)
d8c75104acb68ca648c5a3b30d6791775272e5c1
authentic2/idp/idp_openid/admin.py
authentic2/idp/idp_openid/admin.py
# -*- coding: utf-8 -*- # vim: set ts=4 sw=4 : */ from django.contrib import admin from models import TrustedRoot, Association, Nonce admin.site.register(TrustedRoot) admin.site.register(Association) admin.site.register(Nonce)
# -*- coding: utf-8 -*- from django.contrib import admin from models import TrustedRoot, Association, Nonce admin.site.register(TrustedRoot) admin.site.register(Association) admin.site.register(Nonce)
Remove vim instruction in prologue.
[idp/idp_openid] Remove vim instruction in prologue.
Python
agpl-3.0
incuna/authentic,incuna/authentic,adieu/authentic2,incuna/authentic,adieu/authentic2,BryceLohr/authentic,BryceLohr/authentic,incuna/authentic,BryceLohr/authentic,pu239ppy/authentic2,BryceLohr/authentic,pu239ppy/authentic2,pu239ppy/authentic2,adieu/authentic2,adieu/authentic2,incuna/authentic,pu239ppy/authentic2
--- +++ @@ -1,5 +1,4 @@ # -*- coding: utf-8 -*- -# vim: set ts=4 sw=4 : */ from django.contrib import admin from models import TrustedRoot, Association, Nonce
15c596ea224fb0a624404d425a2580ab3503807d
bluebottle/utils/middleware.py
bluebottle/utils/middleware.py
from importlib import import_module from django.conf import settings from django.contrib.sessions import middleware from django.db import connection from django.utils import translation from tenant_extras.middleware import tenant_translation from bluebottle.utils.models import get_languages, get_default_language class SubDomainSessionMiddleware(middleware.SessionMiddleware): def process_request(self, request): engine = import_module(settings.SESSION_ENGINE) session_key = request.COOKIES.get(settings.SESSION_COOKIE_NAME, None) if session_key is None: # Look for old cookie in request for auth purposes. session_key = request.COOKIES.get('sessionid', None) request.session = engine.SessionStore(session_key) class APILanguageMiddleware(middleware.SessionMiddleware): def process_request(self, request): if request.path.startswith('/api'): try: language = request.META['HTTP_X_APPLICATION_LANGUAGE'] if language not in [lang.code for lang in get_languages()]: language = get_default_language() except KeyError: language = get_default_language() translation.activate(language) translation._trans._active.value = tenant_translation( language, connection.tenant.client_name ) request.LANGUAGE_CODE = translation.get_language()
from importlib import import_module from django.conf import settings from django.contrib.sessions import middleware from django.db import connection from django.utils import translation from tenant_extras.middleware import tenant_translation from bluebottle.utils.models import get_languages, get_default_language class SubDomainSessionMiddleware(middleware.SessionMiddleware): def process_request(self, request): engine = import_module(settings.SESSION_ENGINE) session_key = request.COOKIES.get(settings.SESSION_COOKIE_NAME, None) if session_key is None: # Look for old cookie in request for auth purposes. session_key = request.COOKIES.get('sessionid', None) request.session = engine.SessionStore(session_key) class APILanguageMiddleware(middleware.SessionMiddleware): def process_request(self, request): if request.path.startswith('/api'): try: language = request.META['HTTP_X_APPLICATION_LANGUAGE'] if language not in [lang.full_code for lang in get_languages()]: language = get_default_language() except KeyError: language = get_default_language() translation.activate(language) translation._trans._active.value = tenant_translation( language, connection.tenant.client_name ) request.LANGUAGE_CODE = translation.get_language()
Use full code to check if api language actually exists
Use full code to check if api language actually exists
Python
bsd-3-clause
onepercentclub/bluebottle,onepercentclub/bluebottle,onepercentclub/bluebottle,onepercentclub/bluebottle,onepercentclub/bluebottle
--- +++ @@ -26,7 +26,7 @@ if request.path.startswith('/api'): try: language = request.META['HTTP_X_APPLICATION_LANGUAGE'] - if language not in [lang.code for lang in get_languages()]: + if language not in [lang.full_code for lang in get_languages()]: language = get_default_language() except KeyError: language = get_default_language()
0ad7be235135303cb9d902df2a89b17da8aac918
syntacticframes_project/syntacticframes/migrations/0012_auto_20150220_1836.py
syntacticframes_project/syntacticframes/migrations/0012_auto_20150220_1836.py
# -*- coding: utf-8 -*- from __future__ import unicode_literals from distutils.version import LooseVersion from django.db import models, migrations def set_position_value_for_levin_classes(apps, schema_editor): i = 0 LevinClass = apps.get_model('syntacticframes', 'LevinClass') levin_class_list = sorted(LevinClass.objects.all(), key=lambda l: int(l.number)) for levin_class in levin_class_list: verbnet_classes = sorted( levin_class.verbnetclass_set.all(), key=lambda v: LooseVersion(v.name.split('-')[1])) for v in verbnet_classes: v.position = i v.save() i += 10 class Migration(migrations.Migration): dependencies = [ ('syntacticframes', '0011_auto_20150121_1600'), ] operations = [ migrations.AlterModelOptions( name='verbnetclass', options={'ordering': ['position']}, ), migrations.AddField( model_name='verbnetclass', name='position', field=models.PositiveSmallIntegerField(default=0), preserve_default=False, ), migrations.RunPython(set_position_value_for_levin_classes), ]
# -*- coding: utf-8 -*- from __future__ import unicode_literals from distutils.version import LooseVersion from django.db import models, migrations def set_position_value_for_levin_classes(apps, schema_editor): i = 0 LevinClass = apps.get_model('syntacticframes', 'LevinClass') levin_class_list = sorted(LevinClass.objects.all(), key=lambda l: int(l.number)) for levin_class in levin_class_list: verbnet_classes = sorted( levin_class.verbnetclass_set.all(), key=lambda v: LooseVersion(v.name.split('-')[1])) for v in verbnet_classes: v.position = i v.save() i += 10 class Migration(migrations.Migration): dependencies = [ ('syntacticframes', '0011_auto_20150121_1600'), ] operations = [ migrations.AlterModelOptions( name='verbnetclass', options={'ordering': ['position']}, ), migrations.AddField( model_name='verbnetclass', name='position', field=models.PositiveSmallIntegerField(default=0), preserve_default=False, ), migrations.RunPython( code=set_position_value_for_levin_classes, # Nothing to reverse since this is about a new field reverse_code=lambda apps, schema_editor: None), ]
Make the 0012 migration reversible
Make the 0012 migration reversible
Python
mit
aymara/verbenet-editor,aymara/verbenet-editor,aymara/verbenet-editor
--- +++ @@ -36,5 +36,8 @@ field=models.PositiveSmallIntegerField(default=0), preserve_default=False, ), - migrations.RunPython(set_position_value_for_levin_classes), + migrations.RunPython( + code=set_position_value_for_levin_classes, + # Nothing to reverse since this is about a new field + reverse_code=lambda apps, schema_editor: None), ]
31d973be4f50e2e26d8ff65dd86d96d55e6b6250
contrib/zmq/chromecast-snoop.py
contrib/zmq/chromecast-snoop.py
#!/usr/bin/env python3 class Listener(object): def __init__(self): import zmqclient self.artist = None self.title = None pub = zmqclient.pub() def new_media_status(self, status): if status.artist != self.artist or status.title != self.title: self.artist = status.artist self.title = status.title print("{} -- {}".format(status.artist, status.title)) if status.artist != None or status.title != None: pub.send(b"CHROMECAST", zmq.SNDMORE) pub.send_json(status.media_metadata) if __name__ == '__main__': import pychromecast, time cast = pychromecast.get_chromecast() cast.wait() print("Connected to {}".format(cast.device.friendly_name)) zmq = Listener() cast.media_controller.register_status_listener(zmq) while True: time.sleep(30)
#!/usr/bin/env python3 class Listener(object): def __init__(self): import zmqclient self.artist = None self.title = None self.pub = zmqclient.pub() def new_media_status(self, status): if status.artist != self.artist or status.title != self.title: self.artist = status.artist self.title = status.title print("{} -- {}".format(status.artist, status.title)) if status.artist != None or status.title != None: try: self.pub.send(b"CHROMECAST", 2) # 2 == zmq.SNDMORE FIXME self.pub.send_json(status.media_metadata) except Exception as e: print(e) if __name__ == '__main__': import pychromecast, time cast = pychromecast.get_chromecast() cast.wait() print("Connected to {}".format(cast.device.friendly_name)) zmq = Listener() cast.media_controller.register_status_listener(zmq) while True: time.sleep(30)
Print exceptions since pychromecast just ignores them
Print exceptions since pychromecast just ignores them
Python
apache-2.0
hackeriet/nfcd,hackeriet/nfcd,hackeriet/pyhackeriet,hackeriet/pyhackeriet,hackeriet/nfcd,hackeriet/pyhackeriet
--- +++ @@ -6,7 +6,7 @@ self.artist = None self.title = None - pub = zmqclient.pub() + self.pub = zmqclient.pub() def new_media_status(self, status): if status.artist != self.artist or status.title != self.title: @@ -14,8 +14,11 @@ self.title = status.title print("{} -- {}".format(status.artist, status.title)) if status.artist != None or status.title != None: - pub.send(b"CHROMECAST", zmq.SNDMORE) - pub.send_json(status.media_metadata) + try: + self.pub.send(b"CHROMECAST", 2) # 2 == zmq.SNDMORE FIXME + self.pub.send_json(status.media_metadata) + except Exception as e: + print(e) if __name__ == '__main__': import pychromecast, time
fdbf12d560a18724b5293e4cee8aa3a03a48c90b
tcconfig/tcdel.py
tcconfig/tcdel.py
#!/usr/bin/env python # encoding: utf-8 """ .. codeauthor:: Tsuyoshi Hombashi <gogogo.vm@gmail.com> """ from __future__ import absolute_import import sys import logbook import subprocrunner import tcconfig from .traffic_control import TrafficControl from ._argparse_wrapper import ArgparseWrapper from ._common import verify_network_interface from ._error import NetworkInterfaceNotFoundError handler = logbook.StderrHandler() handler.push_application() def parse_option(): parser = ArgparseWrapper(tcconfig.VERSION) group = parser.parser.add_argument_group("Traffic Control") group.add_argument( "--device", required=True, help="network device name (e.g. eth0)") return parser.parser.parse_args() def main(): options = parse_option() logger = logbook.Logger("tcdel") logger.level = options.log_level subprocrunner.logger.level = options.log_level if options.quiet: subprocrunner.logger.disable() else: subprocrunner.logger.enable() subprocrunner.Which("tc").verify() try: verify_network_interface(options.device) except NetworkInterfaceNotFoundError as e: logger.error(e) return 1 tc = TrafficControl(options.device) return tc.delete_tc() if __name__ == '__main__': sys.exit(main())
#!/usr/bin/env python # encoding: utf-8 """ .. codeauthor:: Tsuyoshi Hombashi <gogogo.vm@gmail.com> """ from __future__ import absolute_import import sys import logbook import subprocrunner import tcconfig from .traffic_control import TrafficControl from ._argparse_wrapper import ArgparseWrapper from ._common import verify_network_interface from ._error import NetworkInterfaceNotFoundError handler = logbook.StderrHandler() handler.push_application() def parse_option(): parser = ArgparseWrapper(tcconfig.VERSION) group = parser.parser.add_argument_group("Traffic Control") group.add_argument( "--device", required=True, help="network device name (e.g. eth0)") return parser.parser.parse_args() def main(): options = parse_option() logger = logbook.Logger("tcdel") logger.level = options.log_level subprocrunner.logger.level = options.log_level if options.quiet: subprocrunner.logger.disable() else: subprocrunner.logger.enable() subprocrunner.Which("tc").verify() try: verify_network_interface(options.device) except NetworkInterfaceNotFoundError as e: logger.error(e) return 1 tc = TrafficControl(options.device) try: return tc.delete_tc() except NetworkInterfaceNotFoundError as e: logger.debug(e) return 0 return 1 if __name__ == '__main__': sys.exit(main())
Modify to handle exception properly
Modify to handle exception properly
Python
mit
thombashi/tcconfig,thombashi/tcconfig
--- +++ @@ -54,7 +54,13 @@ tc = TrafficControl(options.device) - return tc.delete_tc() + try: + return tc.delete_tc() + except NetworkInterfaceNotFoundError as e: + logger.debug(e) + return 0 + + return 1 if __name__ == '__main__':
f6dd7d0ca966856325adc50f4c5ca2cc48dda0a5
cogbot/cog_bot_server_state.py
cogbot/cog_bot_server_state.py
import json import logging import typing from datetime import datetime import discord from cogbot.types import ServerId, ChannelId log = logging.getLogger(__name__) class CogBotServerState: def __init__(self, bot, server: discord.Server, log_channel: ChannelId = None): self.bot = bot self.server: discord.Server = server # resolve log channel self.log_channel: discord.Channel = None if log_channel: self.log_channel = self.bot.get_channel(log_channel) if not self.log_channel: log.warning( f"[{self.server}] Failed to resolve log channel <{log_channel}>" ) async def mod_log( self, member: discord.Member, content: str, channel: discord.Channel = None ): if self.log_channel: now = datetime.utcnow() quote_name = f"{member.display_name} ({member.name}#{member.discriminator})" em = discord.Embed(description=content, timestamp=now) em.set_author(name=quote_name, icon_url=member.avatar_url) em.set_footer(text=f"#{channel}" if channel else None) await self.bot.send_message(self.log_channel, embed=em)
import json import logging import typing from datetime import datetime import discord from cogbot.types import ServerId, ChannelId log = logging.getLogger(__name__) class CogBotServerState: def __init__(self, bot, server: discord.Server, log_channel: ChannelId = None): self.bot = bot self.server: discord.Server = server # resolve log channel self.log_channel: discord.Channel = None if log_channel: self.log_channel = self.bot.get_channel(log_channel) if not self.log_channel: log.warning( f"[{self.server}] Failed to resolve log channel <{log_channel}>" ) async def mod_log( self, member: discord.Member, content: str, channel: discord.Channel = None ): if self.log_channel: now = datetime.utcnow() quote_name = f"{member.display_name} ({member.name}#{member.discriminator})" em = discord.Embed(description=content, timestamp=now) em.set_author(name=quote_name, icon_url=member.avatar_url) if channel: em.set_footer(text=f"#{channel}") await self.bot.send_message(self.log_channel, embed=em)
Fix mod log optional channel
Fix mod log optional channel
Python
mit
Arcensoth/cogbot
--- +++ @@ -33,5 +33,6 @@ quote_name = f"{member.display_name} ({member.name}#{member.discriminator})" em = discord.Embed(description=content, timestamp=now) em.set_author(name=quote_name, icon_url=member.avatar_url) - em.set_footer(text=f"#{channel}" if channel else None) + if channel: + em.set_footer(text=f"#{channel}") await self.bot.send_message(self.log_channel, embed=em)
67c671260858cc2c3d3041188cebda63cac1c4eb
prequ/__init__.py
prequ/__init__.py
import pkg_resources try: __version__ = pkg_resources.get_distribution(__name__).version except pkg_resources.DistributionNotFound: __version__ = None
import pkg_resources try: __version__ = pkg_resources.get_distribution(__name__).version except pkg_resources.DistributionNotFound: # pragma: no cover __version__ = None
Add "no cover" pragma to version setting code
Add "no cover" pragma to version setting code
Python
bsd-2-clause
suutari-ai/prequ,suutari/prequ,suutari/prequ
--- +++ @@ -2,5 +2,5 @@ try: __version__ = pkg_resources.get_distribution(__name__).version -except pkg_resources.DistributionNotFound: +except pkg_resources.DistributionNotFound: # pragma: no cover __version__ = None
6171b8111359cc54a4af2c3444ce0e0e2db5ba80
froide/helper/context_processors.py
froide/helper/context_processors.py
from django.conf import settings def froide(request): return {"froide": settings.FROIDE_CONFIG} def site_settings(request): return {"SITE_NAME": settings.SITE_NAME, "SITE_URL": settings.SITE_URL, "FROIDE_DRYRUN": settings.FROIDE_DRYRUN, "FROIDE_DRYRUN_DOMAIN": settings.FROIDE_DRYRUN_DOMAIN}
from django.conf import settings def froide(request): return {"froide": settings.FROIDE_CONFIG} def site_settings(request): return {"SITE_NAME": settings.SITE_NAME, "SITE_URL": settings.SITE_URL, "FROIDE_DRYRUN": settings.FROIDE_DRYRUN, "FROIDE_DRYRUN_DOMAIN": settings.FROIDE_DRYRUN_DOMAIN, "LANGUAGE_CODE": settings.LANGUAGE_CODE}
Add Froide Dry Run Domain and Language Code to context_processor
Add Froide Dry Run Domain and Language Code to context_processor
Python
mit
okfse/froide,ryankanno/froide,fin/froide,LilithWittmann/froide,okfse/froide,fin/froide,ryankanno/froide,LilithWittmann/froide,catcosmo/froide,CodeforHawaii/froide,CodeforHawaii/froide,stefanw/froide,catcosmo/froide,ryankanno/froide,catcosmo/froide,catcosmo/froide,okfse/froide,fin/froide,ryankanno/froide,stefanw/froide,ryankanno/froide,CodeforHawaii/froide,LilithWittmann/froide,okfse/froide,stefanw/froide,fin/froide,stefanw/froide,catcosmo/froide,stefanw/froide,LilithWittmann/froide,okfse/froide,CodeforHawaii/froide,LilithWittmann/froide,CodeforHawaii/froide
--- +++ @@ -7,4 +7,5 @@ return {"SITE_NAME": settings.SITE_NAME, "SITE_URL": settings.SITE_URL, "FROIDE_DRYRUN": settings.FROIDE_DRYRUN, - "FROIDE_DRYRUN_DOMAIN": settings.FROIDE_DRYRUN_DOMAIN} + "FROIDE_DRYRUN_DOMAIN": settings.FROIDE_DRYRUN_DOMAIN, + "LANGUAGE_CODE": settings.LANGUAGE_CODE}
50451c69d337228c2016851258ff7249bf906440
profiling/plot.py
profiling/plot.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- import glob import re import numpy as np import matplotlib.pyplot as plt csv_files = glob.glob('*.csv') fig = plt.figure() ax = fig.add_subplot(111) colors = iter(plt.cm.rainbow(np.linspace(0,1,len(csv_files)))) p = re.compile(r'profiling_(.*?)_(.*?)\.csv') ms_to_s = 1.0 / 1000.0 for csv_file in csv_files: data = np.genfromtxt(csv_file, delimiter=',', skip_header=1).transpose() j = data[0] N = data[1] avg = data[2] std = data[3] m = p.search(csv_file) name = m.group(2) name = name.replace('_', ' ') ax.errorbar(N, avg*ms_to_s, yerr=std*ms_to_s, label=name, color=next(colors), marker='o') ax.grid(True) ax.set_xlabel('N') ax.set_ylabel('Timing [s]') ax.set_xscale('log') ax.set_yscale('log') ax.legend(loc='best') plt.show()
#!/usr/bin/env python3 # -*- coding: utf-8 -*- import glob import re import os import sys import numpy as np import matplotlib.pyplot as plt if len(sys.argv) == 1: print('Usage: plot.py path/to/build/profiling') sys.exit(0) csv_files = glob.glob(os.path.join(sys.argv[1], '*.csv')) fig = plt.figure() ax = fig.add_subplot(111) colors = iter(plt.cm.rainbow(np.linspace(0,1,len(csv_files)))) p = re.compile(r'profiling_(.*?)_(.*?)\.csv') ms_to_s = 1.0 / 1000.0 for csv_file in csv_files: data = np.genfromtxt(csv_file, delimiter=',', skip_header=1).transpose() j = data[0] N = data[1] avg = data[2] std = data[3] m = p.search(csv_file) name = m.group(2) name = name.replace('_', ' ') ax.errorbar(N, avg*ms_to_s, yerr=std*ms_to_s, label=name, color=next(colors), marker='o') ax.grid(True) ax.set_xlabel('N') ax.set_ylabel('Timing [s]') ax.set_xscale('log') ax.set_yscale('log') ax.legend(loc='best') plt.show()
Use path given as argument
Use path given as argument
Python
bsd-3-clause
nbigaouette/sorting,nbigaouette/sorting,nbigaouette/sorting,nbigaouette/sorting
--- +++ @@ -3,12 +3,18 @@ import glob import re +import os +import sys import numpy as np import matplotlib.pyplot as plt -csv_files = glob.glob('*.csv') +if len(sys.argv) == 1: + print('Usage: plot.py path/to/build/profiling') + sys.exit(0) + +csv_files = glob.glob(os.path.join(sys.argv[1], '*.csv')) fig = plt.figure() ax = fig.add_subplot(111)
d8ce56feada64d287306d7f439ec12a42acda0d6
bot.py
bot.py
#!/usr/bin/env python3 # -*- coding: utf8 -*- import tweepy consumer_key = "" consumer_secret = "" access_token = "" access_token_secret = "" auth = tweepy.OAuthHandler(consumer_key, consumer_secret) auth.set_access_token(access_token, access_token_secret) api = tweepy.API(auth) userid = str(input("Please input id who you want fav attack\n")) count = input("input number you want to fav!\n") fav = api.user_timeline(id = userid, count = count) try: for status in fav: api.create_favorite(status.id_str) except tweepy.error.TweepError as e: if e.args[0][0]['code'] == 139: print("You have already favorited this status! \n") else: print(e.reason) finally: print("Done!")
#!/usr/bin/env python3 # -*- coding: utf8 -*- import tweepy consumer_key = "" consumer_secret = "" access_token = "" access_token_secret = "" auth = tweepy.OAuthHandler(consumer_key, consumer_secret) auth.set_access_token(access_token, access_token_secret) api = tweepy.API(auth) def getdata(): userid = str(input("Please input id who you want fav attack\n")) count = input("input number you want to fav!\n") fav = api.user_timeline(id = userid, count = count) def main(): getdata() try: for status in fav: api.create_favorite(status.id_str) except tweepy.error.TweepError as e: if e.args[0][0]['code'] == 139: print("You have already favorited this status! \n") else: print(e.reason) finally: print("Done!") if __name__ == "__main__": main()
Make it more complex (((
Make it more complex (((
Python
mit
zhangyubaka/tweepy_favbot
--- +++ @@ -15,18 +15,23 @@ api = tweepy.API(auth) -userid = str(input("Please input id who you want fav attack\n")) -count = input("input number you want to fav!\n") +def getdata(): + userid = str(input("Please input id who you want fav attack\n")) + count = input("input number you want to fav!\n") + fav = api.user_timeline(id = userid, count = count) -fav = api.user_timeline(id = userid, count = count) +def main(): + getdata() + try: + for status in fav: + api.create_favorite(status.id_str) + except tweepy.error.TweepError as e: + if e.args[0][0]['code'] == 139: + print("You have already favorited this status! \n") + else: + print(e.reason) + finally: + print("Done!") -try: - for status in fav: - api.create_favorite(status.id_str) -except tweepy.error.TweepError as e: - if e.args[0][0]['code'] == 139: - print("You have already favorited this status! \n") - else: - print(e.reason) -finally: - print("Done!") +if __name__ == "__main__": + main()
f511af4fc89a170914a86de1704e8e842ffd6b6d
test/test_configuration.py
test/test_configuration.py
#!/usr/bin/env python """Test coordinate classes.""" import sys try: import unittest2 as unittest # Python 2.6 except ImportError: import unittest import heatmap as hm class Tests(unittest.TestCase): # To remove Python 3's # "DeprecationWarning: Please use assertRaisesRegex instead" if sys.version_info[0] == 2: assertRaisesRegex = unittest.TestCase.assertRaisesRegexp def test_basic(self): '''Test Configuration class.''' # Act config = hm.Configuration(use_defaults=True) # Assert self.assertEqual(config.margin, 0) self.assertEqual(config.frequency, 1) def test_fill_missing_no_input(self): '''Test Configuration class.''' # Arrange config = hm.Configuration(use_defaults=True) # Act / Assert with self.assertRaisesRegex(ValueError, "no input specified"): config.fill_missing() if __name__ == '__main__': unittest.main()
#!/usr/bin/env python """Test coordinate classes.""" import sys try: import unittest2 as unittest # Python 2.6 except ImportError: import unittest ROOT_DIR = os.path.split(os.path.abspath(os.path.dirname(__file__)))[0] sys.path.append(ROOT_DIR) import heatmap as hm class Tests(unittest.TestCase): # To remove Python 3's # "DeprecationWarning: Please use assertRaisesRegex instead" if sys.version_info[0] == 2: assertRaisesRegex = unittest.TestCase.assertRaisesRegexp def test_basic(self): '''Test Configuration class.''' # Act config = hm.Configuration(use_defaults=True) # Assert self.assertEqual(config.margin, 0) self.assertEqual(config.frequency, 1) def test_fill_missing_no_input(self): '''Test Configuration class.''' # Arrange config = hm.Configuration(use_defaults=True) # Act / Assert with self.assertRaisesRegex(ValueError, "no input specified"): config.fill_missing() if __name__ == '__main__': unittest.main()
Update sys.path to import heatmap
Update sys.path to import heatmap
Python
agpl-3.0
hugovk/heatmap,hugovk/heatmap,sethoscope/heatmap,sethoscope/heatmap
--- +++ @@ -8,6 +8,8 @@ except ImportError: import unittest +ROOT_DIR = os.path.split(os.path.abspath(os.path.dirname(__file__)))[0] +sys.path.append(ROOT_DIR) import heatmap as hm @@ -20,7 +22,6 @@ def test_basic(self): '''Test Configuration class.''' - # Act config = hm.Configuration(use_defaults=True)
c3883ad36139db4592c259a94e7b254702ec2f00
run.py
run.py
import glpi.server;glpi.server.app.run(debug = True)
import glpi.server;glpi.server.app.run(debug = True, host='0.0.0.0')
Add option to rlisten on al interfaces
Add option to rlisten on al interfaces
Python
agpl-3.0
ddurieux/poc_glpi
--- +++ @@ -1 +1 @@ -import glpi.server;glpi.server.app.run(debug = True) +import glpi.server;glpi.server.app.run(debug = True, host='0.0.0.0')
f62ec2304bb42b0bf0eba121c1b88adb35cec6d3
run.py
run.py
#!/usr/bin/env python import os import argparse def run(): os.system("gunicorn server:app --reload --config gunicorn_config.py") def deploy(): os.system("git push heroku master") def main(): parser = argparse.ArgumentParser() parser.add_argument('--deploy', action="store_true", required=False) args = parser.parse_args() if args.deploy: deploy() else: run() if __name__ == '__main__': main()
#!/usr/bin/env python import os import argparse def run(): os.system("gunicorn server:app --reload --config gunicorn_config.py") def deploy(): os.system("git push dokku master") def main(): parser = argparse.ArgumentParser() parser.add_argument('--deploy', action="store_true", required=False) args = parser.parse_args() if args.deploy: deploy() else: run() if __name__ == '__main__': main()
Switch to dokku for deployment.
Switch to dokku for deployment.
Python
mit
EmilStenstrom/json-tagger,EmilStenstrom/json-tagger,EmilStenstrom/json-tagger,EmilStenstrom/json-tagger,EmilStenstrom/json-tagger
--- +++ @@ -6,7 +6,7 @@ os.system("gunicorn server:app --reload --config gunicorn_config.py") def deploy(): - os.system("git push heroku master") + os.system("git push dokku master") def main(): parser = argparse.ArgumentParser()
a24faf712d8dfba0f6ac9fc295807552dca37ae9
custom/inddex/reports/utils.py
custom/inddex/reports/utils.py
from corehq.apps.reports.datatables import DataTablesColumn, DataTablesHeader from corehq.apps.reports.generic import GenericTabularReport from corehq.apps.reports.standard import CustomProjectReport, DatespanMixin class MultiTabularReport(DatespanMixin, CustomProjectReport, GenericTabularReport): report_template_path = 'inddex/multi_report.html' exportable = True export_only = False @property def data_providers(self): # data providers should supply a title, slug, headers, and rows return [] @property def report_context(self): context = { 'name': self.name, 'export_only': self.export_only } if not self.needs_filters: context['data_providers'] = [{ 'title': data_provider.title, 'slug': data_provider.slug, 'headers': DataTablesHeader( *(DataTablesColumn(header) for header in data_provider.headers), ), 'rows': data_provider.rows, } for data_provider in self.data_providers] return context @property def export_table(self): return [ [dp.slug, [dp.headers] + dp.rows] for dp in self.data_providers ]
from itertools import chain from corehq.apps.reports.datatables import DataTablesColumn, DataTablesHeader from corehq.apps.reports.generic import GenericTabularReport from corehq.apps.reports.standard import CustomProjectReport, DatespanMixin class MultiTabularReport(DatespanMixin, CustomProjectReport, GenericTabularReport): report_template_path = 'inddex/multi_report.html' exportable = True exportable_all = True export_only = False @property def data_providers(self): # data providers should supply a title, slug, headers, and rows return [] @property def report_context(self): context = { 'name': self.name, 'export_only': self.export_only } if not self.needs_filters: context['data_providers'] = [{ 'title': data_provider.title, 'slug': data_provider.slug, 'headers': DataTablesHeader( *(DataTablesColumn(header) for header in data_provider.headers), ), 'rows': data_provider.rows, } for data_provider in self.data_providers] return context @property def export_table(self): return [ [dp.slug, chain([dp.headers], dp.rows)] for dp in self.data_providers ]
Move export to a background process
Move export to a background process
Python
bsd-3-clause
dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq
--- +++ @@ -1,3 +1,5 @@ +from itertools import chain + from corehq.apps.reports.datatables import DataTablesColumn, DataTablesHeader from corehq.apps.reports.generic import GenericTabularReport from corehq.apps.reports.standard import CustomProjectReport, DatespanMixin @@ -6,6 +8,7 @@ class MultiTabularReport(DatespanMixin, CustomProjectReport, GenericTabularReport): report_template_path = 'inddex/multi_report.html' exportable = True + exportable_all = True export_only = False @property @@ -33,6 +36,6 @@ @property def export_table(self): return [ - [dp.slug, [dp.headers] + dp.rows] + [dp.slug, chain([dp.headers], dp.rows)] for dp in self.data_providers ]
46c33ca68c1124fb06c4ba62306cb00ba61d7e5c
tests/__init__.py
tests/__init__.py
from flexmock import flexmock from flask.ext.storage import MockStorage from flask_uploads import init class TestCase(object): added_objects = [] committed_objects = [] created_objects = [] deleted_objects = [] def setup_method(self, method, resizer=None): init(db_mock, MockStorage, resizer) self.db = db_mock self.Storage = MockStorage self.storage = MockStorage() self.resizer = resizer def teardown_method(self, method): # Empty the stacks. TestCase.added_objects[:] = [] TestCase.committed_objects[:] = [] TestCase.created_objects[:] = [] TestCase.deleted_objects[:] = [] class MockModel(object): def __init__(self, **kw): TestCase.created_objects.append(self) for key, val in kw.iteritems(): setattr(self, key, val) db_mock = flexmock( Column=lambda *a, **kw: ('column', a, kw), Integer=('integer', [], {}), Unicode=lambda *a, **kw: ('unicode', a, kw), Model=MockModel, session=flexmock( add=TestCase.added_objects.append, commit=lambda: TestCase.committed_objects.extend( TestCase.added_objects + TestCase.deleted_objects ), delete=TestCase.deleted_objects.append, ), )
from flexmock import flexmock from flask.ext.storage import MockStorage from flask_uploads import init class TestCase(object): added_objects = [] committed_objects = [] created_objects = [] deleted_objects = [] def setup_method(self, method, resizer=None): init(db_mock, MockStorage, resizer) self.db = db_mock self.Storage = MockStorage self.storage = MockStorage() self.resizer = resizer def teardown_method(self, method): # Empty the stacks. TestCase.added_objects[:] = [] TestCase.committed_objects[:] = [] TestCase.created_objects[:] = [] TestCase.deleted_objects[:] = [] class MockModel(object): def __init__(self, **kw): TestCase.created_objects.append(self) for key, val in kw.iteritems(): setattr(self, key, val) db_mock = flexmock( Column=lambda *a, **kw: ('column', a, kw), Integer=('integer', [], {}), Unicode=lambda *a, **kw: ('unicode', a, kw), Model=MockModel, metadata=flexmock(tables={}), session=flexmock( add=TestCase.added_objects.append, commit=lambda: TestCase.committed_objects.extend( TestCase.added_objects + TestCase.deleted_objects ), delete=TestCase.deleted_objects.append, ), )
Add metadata.tables to mock db.
Add metadata.tables to mock db.
Python
mit
FelixLoether/flask-uploads,FelixLoether/flask-image-upload-thing
--- +++ @@ -36,6 +36,7 @@ Integer=('integer', [], {}), Unicode=lambda *a, **kw: ('unicode', a, kw), Model=MockModel, + metadata=flexmock(tables={}), session=flexmock( add=TestCase.added_objects.append, commit=lambda: TestCase.committed_objects.extend(
3037c15a658ee05a4483c5b1c8c137a06e9d7707
tests/conftest.py
tests/conftest.py
''' General-purpose fixtures for vdirsyncer's testsuite. ''' import logging import os import click_log from hypothesis import HealthCheck, Verbosity, settings import pytest @pytest.fixture(autouse=True) def setup_logging(): click_log.basic_config('vdirsyncer').setLevel(logging.DEBUG) try: import pytest_benchmark except ImportError: @pytest.fixture def benchmark(): return lambda x: x() else: del pytest_benchmark settings.register_profile("ci", settings( max_examples=1000, verbosity=Verbosity.verbose, suppress_health_check=[HealthCheck.too_slow], )) settings.register_profile("deterministic", settings( derandomize=True, suppress_health_check=HealthCheck.all(), )) if os.environ.get('DETERMINISTIC_TESTS', 'false').lower() == 'true': settings.load_profile("deterministic") elif os.environ.get('CI', 'false').lower() == 'true': settings.load_profile("ci")
''' General-purpose fixtures for vdirsyncer's testsuite. ''' import logging import os import click_log from hypothesis import HealthCheck, Verbosity, settings import pytest @pytest.fixture(autouse=True) def setup_logging(): click_log.basic_config('vdirsyncer').setLevel(logging.DEBUG) try: import pytest_benchmark except ImportError: @pytest.fixture def benchmark(): return lambda x: x() else: del pytest_benchmark settings.register_profile("ci", settings( max_examples=1000, verbosity=Verbosity.verbose, suppress_health_check=[HealthCheck.too_slow], )) settings.register_profile("deterministic", settings( derandomize=True, suppress_health_check=HealthCheck.all(), )) settings.register_profile("dev", settings( suppress_health_check=[HealthCheck.too_slow] )) if os.environ.get('DETERMINISTIC_TESTS', 'false').lower() == 'true': settings.load_profile("deterministic") elif os.environ.get('CI', 'false').lower() == 'true': settings.load_profile("ci") else: settings.load_profile("dev")
Use hypothesis setting load_profile to setup health check
Use hypothesis setting load_profile to setup health check Fixes #779
Python
mit
untitaker/vdirsyncer,untitaker/vdirsyncer,untitaker/vdirsyncer
--- +++ @@ -35,8 +35,13 @@ derandomize=True, suppress_health_check=HealthCheck.all(), )) +settings.register_profile("dev", settings( + suppress_health_check=[HealthCheck.too_slow] +)) if os.environ.get('DETERMINISTIC_TESTS', 'false').lower() == 'true': settings.load_profile("deterministic") elif os.environ.get('CI', 'false').lower() == 'true': settings.load_profile("ci") +else: + settings.load_profile("dev")
24b78a4d510606563106da24d568d5fb79ddca2b
IPython/__main__.py
IPython/__main__.py
# encoding: utf-8 """Terminal-based IPython entry point. """ #----------------------------------------------------------------------------- # Copyright (c) 2012, IPython Development Team. # # Distributed under the terms of the Modified BSD License. # # The full license is in the file COPYING.txt, distributed with this software. #----------------------------------------------------------------------------- from IPython.terminal.ipapp import launch_new_instance launch_new_instance()
# encoding: utf-8 """Terminal-based IPython entry point. """ #----------------------------------------------------------------------------- # Copyright (c) 2012, IPython Development Team. # # Distributed under the terms of the Modified BSD License. # # The full license is in the file COPYING.txt, distributed with this software. #----------------------------------------------------------------------------- from IPython import start_ipython start_ipython()
Use new entry point for python -m IPython
Use new entry point for python -m IPython
Python
bsd-3-clause
ipython/ipython,ipython/ipython
--- +++ @@ -9,6 +9,6 @@ # The full license is in the file COPYING.txt, distributed with this software. #----------------------------------------------------------------------------- -from IPython.terminal.ipapp import launch_new_instance +from IPython import start_ipython -launch_new_instance() +start_ipython()
8c8b668ba3684c3e756bf9fccafbd1bd8e1a7cfe
mediapipe/__init__.py
mediapipe/__init__.py
"""Copyright 2019 - 2020 The MediaPipe Authors. Licensed under the Apache License, Version 2.0 (the "License"); you may not use this file except in compliance with the License. You may obtain a copy of the License at http://www.apache.org/licenses/LICENSE-2.0 Unless required by applicable law or agreed to in writing, software distributed under the License is distributed on an "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the License for the specific language governing permissions and limitations under the License. """
# Copyright 2019 - 2022 The MediaPipe Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License.
Fix comment for `mediapipe` license.
Fix comment for `mediapipe` license. The `"""` comment indicates a public docstring for the module, and will end up in the generated docs. By using a "private" comment (`#`) we will not document the license as part of the API. The Apache license is noted in the footer of generated docs, and this is sufficient. PiperOrigin-RevId: 487688719
Python
apache-2.0
google/mediapipe,google/mediapipe,google/mediapipe,google/mediapipe,google/mediapipe,google/mediapipe,google/mediapipe,google/mediapipe
--- +++ @@ -1,14 +1,13 @@ -"""Copyright 2019 - 2020 The MediaPipe Authors. - -Licensed under the Apache License, Version 2.0 (the "License"); -you may not use this file except in compliance with the License. -You may obtain a copy of the License at - - http://www.apache.org/licenses/LICENSE-2.0 - -Unless required by applicable law or agreed to in writing, software -distributed under the License is distributed on an "AS IS" BASIS, -WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. -See the License for the specific language governing permissions and -limitations under the License. -""" +# Copyright 2019 - 2022 The MediaPipe Authors. +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License.
189847ffdca0264ddd6248faa9974ba35eaea373
tests/test_aur.py
tests/test_aur.py
# MIT licensed # Copyright (c) 2013-2017 lilydjwg <lilydjwg@gmail.com>, et al. from flaky import flaky import pytest pytestmark = pytest.mark.asyncio @flaky(max_runs=5) async def test_aur(get_version): assert await get_version("asciidoc-fake", {"aur": None}) == "1.0-1" @flaky(max_runs=5) async def test_aur_strip_release(get_version): assert await get_version("asciidoc-fake", {"aur": None, "strip-release": 1}) == "1.0"
# MIT licensed # Copyright (c) 2013-2017 lilydjwg <lilydjwg@gmail.com>, et al. from flaky import flaky import pytest pytestmark = pytest.mark.asyncio @flaky async def test_aur(get_version): assert await get_version("asciidoc-fake", {"aur": None}) == "1.0-1" @flaky async def test_aur_strip_release(get_version): assert await get_version("asciidoc-fake", {"aur": None, "strip-release": 1}) == "1.0"
Revert "make AUR tests more flaky"
Revert "make AUR tests more flaky" This reverts commit 61df628bd8bc97acbed40a4af67b124c47584f5f. It doesn't help :-(
Python
mit
lilydjwg/nvchecker
--- +++ @@ -5,10 +5,10 @@ import pytest pytestmark = pytest.mark.asyncio -@flaky(max_runs=5) +@flaky async def test_aur(get_version): assert await get_version("asciidoc-fake", {"aur": None}) == "1.0-1" -@flaky(max_runs=5) +@flaky async def test_aur_strip_release(get_version): assert await get_version("asciidoc-fake", {"aur": None, "strip-release": 1}) == "1.0"
6d22cc47174139b56fad7d94696b08d9830a7ea4
lettuce_webdriver/tests/__init__.py
lettuce_webdriver/tests/__init__.py
from __future__ import print_function import os from contextlib import contextmanager from selenium import webdriver from aloe import around, world here = os.path.dirname(__file__) html_pages = os.path.join(here, 'html_pages') @around.each_feature @contextmanager def with_browser(feature): world.browser = webdriver.Firefox() world.browser.get('') yield world.browser.quit() delattr(world, 'browser') @around.each_step @contextmanager def print_source(step): try: yield except: print(world.browser.page_source) print(world.browser.get_screenshot_as_base64()) raise
from __future__ import print_function import os from contextlib import contextmanager from selenium import webdriver from aloe import around, world here = os.path.dirname(__file__) html_pages = os.path.join(here, 'html_pages') @around.each_feature @contextmanager def with_browser(feature): world.browser = webdriver.Firefox() world.browser.get('') yield world.browser.quit() delattr(world, 'browser') @around.each_step @contextmanager def print_source(step): try: yield except: try: step_container = step.scenario except AttributeError: step_container = step.background print(step_container.feature.name) print(step_container.name) print(step.sentence) print(world.browser.page_source) print(world.browser.get_screenshot_as_base64()) raise
Print scenario/step names on failure
Print scenario/step names on failure
Python
mit
koterpillar/aloe_webdriver,infoxchange/aloe_webdriver,infoxchange/aloe_webdriver,aloetesting/aloe_webdriver,aloetesting/aloe_webdriver,aloetesting/aloe_webdriver,koterpillar/aloe_webdriver
--- +++ @@ -27,6 +27,14 @@ try: yield except: + try: + step_container = step.scenario + except AttributeError: + step_container = step.background + + print(step_container.feature.name) + print(step_container.name) + print(step.sentence) print(world.browser.page_source) print(world.browser.get_screenshot_as_base64()) raise
73c1900a05fa3e4f68224f4e0d5dce2c08687254
opwen_email_server/backend/email_sender.py
opwen_email_server/backend/email_sender.py
from typing import Tuple from opwen_email_server import azure_constants as constants from opwen_email_server import config from opwen_email_server.services.queue import AzureQueue from opwen_email_server.services.sendgrid import SendgridEmailSender QUEUE = AzureQueue(account=config.QUEUES_ACCOUNT, key=config.QUEUES_KEY, name=constants.QUEUE_EMAIL_SEND) EMAIL = SendgridEmailSender(key=config.EMAIL_SENDER_KEY) def send(email: dict) -> Tuple[str, int]: success = EMAIL.send_email(email) if not success: return 'error', 500 return 'sent', 200 if __name__ == '__main__': from argparse import ArgumentParser from json import loads from uuid import uuid4 parser = ArgumentParser() parser.add_argument('email') args = parser.parse_args() email = loads(args.email) email.setdefault('_uid', str(uuid4())) send(email)
from typing import Tuple from opwen_email_server import azure_constants as constants from opwen_email_server import config from opwen_email_server.services.queue import AzureQueue from opwen_email_server.services.sendgrid import SendgridEmailSender QUEUE = AzureQueue(account=config.QUEUES_ACCOUNT, key=config.QUEUES_KEY, name=constants.QUEUE_EMAIL_SEND) EMAIL = SendgridEmailSender(key=config.EMAIL_SENDER_KEY) def send(email: dict) -> Tuple[str, int]: success = EMAIL.send_email(email) if not success: return 'error', 500 return 'sent', 200 if __name__ == '__main__': from argparse import ArgumentParser from argparse import FileType from base64 import b64encode from json import loads from os.path import basename from uuid import uuid4 parser = ArgumentParser() parser.add_argument('email') parser.add_argument('--attachment', type=FileType('rb')) args = parser.parse_args() email = loads(args.email) email.setdefault('_uid', str(uuid4())) if args.attachment: email.setdefault('attachments', []).append({ 'filename': basename(args.attachment.name), 'content': b64encode(args.attachment.read()).decode('ascii') }) args.attachment.close() send(email)
Add attachment support to email sender CLI
Add attachment support to email sender CLI
Python
apache-2.0
ascoderu/opwen-cloudserver,ascoderu/opwen-cloudserver
--- +++ @@ -22,14 +22,25 @@ if __name__ == '__main__': from argparse import ArgumentParser + from argparse import FileType + from base64 import b64encode from json import loads + from os.path import basename from uuid import uuid4 parser = ArgumentParser() parser.add_argument('email') + parser.add_argument('--attachment', type=FileType('rb')) args = parser.parse_args() email = loads(args.email) email.setdefault('_uid', str(uuid4())) + if args.attachment: + email.setdefault('attachments', []).append({ + 'filename': basename(args.attachment.name), + 'content': b64encode(args.attachment.read()).decode('ascii') + }) + args.attachment.close() + send(email)
dd5d81263842066bf1faa3699c12da47f9453401
packages/flake8-config-4catalyzer/setup.py
packages/flake8-config-4catalyzer/setup.py
from setuptools import setup setup( name='flake8-config-4catalyzer', version='0.2.1', url='https://github.com/4Catalyzer/python/tree/packages/flake8-config-4catalyzer', author="Alex Rothberg", author_email='arothberg@4catalyzer.com', license='MIT', classifiers=( 'Development Status :: 3 - Alpha', 'Environment :: Console', 'Framework :: Flake8', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 3', 'Topic :: Software Development :: Libraries :: Python Modules', 'Topic :: Software Development :: Quality Assurance', ), keywords='flake8', install_requires=( 'flake8', 'flake8-commas', 'flake8-import-order', ), extras_require={ ':python_version>="3.5"': ('flake8-bugbear',), }, )
from setuptools import setup setup( name='flake8-config-4catalyzer', version='0.2.1', url='https://github.com/4Catalyzer/python/tree/packages/flake8-config-4catalyzer', author="Alex Rothberg", author_email='arothberg@4catalyzer.com', license='MIT', classifiers=( 'Development Status :: 3 - Alpha', 'Environment :: Console', 'Framework :: Flake8', 'Intended Audience :: Developers', 'License :: OSI Approved :: MIT License', 'Operating System :: OS Independent', 'Programming Language :: Python', 'Programming Language :: Python :: 2', 'Programming Language :: Python :: 3', 'Topic :: Software Development :: Libraries :: Python Modules', 'Topic :: Software Development :: Quality Assurance', ), keywords='flake8', install_requires=( 'flake8', 'flake8-commas', 'flake8-debugger', 'flake8-import-order', ), extras_require={ ':python_version>="3.5"': ('flake8-bugbear',), }, )
Add flake8-debugger to list of flake8 checks
Add flake8-debugger to list of flake8 checks
Python
mit
4Catalyzer/python
--- +++ @@ -25,6 +25,7 @@ install_requires=( 'flake8', 'flake8-commas', + 'flake8-debugger', 'flake8-import-order', ), extras_require={
dcb1fc943ec4fe39bd752b1015ba11f6d8145c27
modules/status.py
modules/status.py
import discord from modules.botModule import BotModule from modules.help import * import time import datetime class Status(BotModule): name = 'status' description = 'Allow for the assignment and removal of roles.' help_text = 'Usage: `!status` shows information about this instance of scubot.' trigger_string = '!status' init_time = 0 def __init__(self): init_time = time.time() def uptime_convert(self,seconds): minutes, seconds = divmod(seconds, 60) hours, minutes = divmod(minutes, 60) days, hours = divmod(hours, 24) return days, hours, minutes, seconds async def parse_command(self, message, client): uptime = self.init_time - time.time() uptime_string = self.uptime_convert(uptime) uptime_string = [str(round(x,0))[:-2] for x in uptime_string] uptime_string = uptime_string[0] + 'd ' + uptime_string[1] + 'h ' + uptime_string[2] + 'm ' + uptime_string[3] + 's' print(self.loaded_modules) module_string = '' for botModule in self.loaded_modules: module_string += botModule.name + ', ' module_string = module_string[:-2] msg = '```\n Uptime: ' + uptime_string + '\n Loaded modules: ' + module_string + '\n```' await client.send_message(message.channel, msg)
import discord from modules.botModule import BotModule from modules.help import * import time import datetime class Status(BotModule): name = 'status' description = 'Allow for the assignment and removal of roles.' help_text = 'Usage: `!status` shows information about this instance of scubot.' trigger_string = '!status' init_time = time.time() def __init__(self): init_time = time.time() def uptime_convert(self,seconds): minutes, seconds = divmod(seconds, 60) hours, minutes = divmod(minutes, 60) days, hours = divmod(hours, 24) return days, hours, minutes, seconds async def parse_command(self, message, client): uptime = time.time() - self.init_time uptime_string = self.uptime_convert(uptime) uptime_string = [str(round(x,0))[:-2] for x in uptime_string] uptime_string = uptime_string[0] + 'd ' + uptime_string[1] + 'h ' + uptime_string[2] + 'm ' + uptime_string[3] + 's' module_string = '' for botModule in self.loaded_modules: module_string += botModule.name + ', ' module_string = module_string[:-2] msg = '```\n Uptime: ' + uptime_string + '\n Loaded modules: ' + module_string + '\n```' await client.send_message(message.channel, msg)
Fix uptime and uptime timing
Fix uptime and uptime timing
Python
mit
suclearnub/scubot
--- +++ @@ -13,7 +13,7 @@ trigger_string = '!status' - init_time = 0 + init_time = time.time() def __init__(self): init_time = time.time() @@ -25,11 +25,10 @@ return days, hours, minutes, seconds async def parse_command(self, message, client): - uptime = self.init_time - time.time() + uptime = time.time() - self.init_time uptime_string = self.uptime_convert(uptime) uptime_string = [str(round(x,0))[:-2] for x in uptime_string] uptime_string = uptime_string[0] + 'd ' + uptime_string[1] + 'h ' + uptime_string[2] + 'm ' + uptime_string[3] + 's' - print(self.loaded_modules) module_string = '' for botModule in self.loaded_modules: module_string += botModule.name + ', '
a9e80e81fe2e6ad1325047cb3045ab12640f984f
cla_backend/apps/cla_eventlog/management/commands/find_and_delete_old_cases.py
cla_backend/apps/cla_eventlog/management/commands/find_and_delete_old_cases.py
from django.core.management.base import BaseCommand from dateutil.relativedelta import relativedelta from legalaid.models import Case from cla_butler.tasks import DeleteOldData class FindAndDeleteCasesUsingCreationTime(DeleteOldData): def get_eligible_cases(self): self._setup() two_years = self.now - relativedelta(years=2) return Case.objects.filter(created__lte=two_years).exclude(log__created__gte=two_years) class Command(BaseCommand): help = ( "Find or delete cases that are 2 years old or over that were not deleted prior to the task command being fixed" ) def handle(self, *args, **kwargs): instance = FindAndDeleteCasesUsingCreationTime() if len(args) == 0: cases = instance.get_eligible_cases() print(cases.count()) elif args[0] == "delete": instance.run()
from django.core.management.base import BaseCommand from dateutil.relativedelta import relativedelta from legalaid.models import Case from cla_butler.tasks import DeleteOldData class FindAndDeleteCasesUsingCreationTime(DeleteOldData): def get_eligible_cases(self): self._setup() two_years = self.now - relativedelta(years=2) return Case.objects.filter(created__lte=two_years).exclude(log__created__gte=two_years) class Command(BaseCommand): help = ( "Find or delete cases that are 2 years old or over that were not deleted prior to the task command being fixed" ) def handle(self, *args, **kwargs): instance = FindAndDeleteCasesUsingCreationTime() cases = instance.get_eligible_cases() if len(args) == 0: print(cases.count()) elif args[0] == "test_find": return cases elif args[0] == "delete": instance.run()
Refactor code so we can use command in tests
Refactor code so we can use command in tests
Python
mit
ministryofjustice/cla_backend,ministryofjustice/cla_backend,ministryofjustice/cla_backend,ministryofjustice/cla_backend
--- +++ @@ -20,8 +20,10 @@ def handle(self, *args, **kwargs): instance = FindAndDeleteCasesUsingCreationTime() + cases = instance.get_eligible_cases() if len(args) == 0: - cases = instance.get_eligible_cases() print(cases.count()) + elif args[0] == "test_find": + return cases elif args[0] == "delete": instance.run()
cb2a6c44fcff3453bbf20dbecbd807ee8e640e8e
database/queries/select_queries.py
database/queries/select_queries.py
SELECT_MOVIES_ORDERED_BY_RATING = ''' SELECT * FROM MOVIE ORDER BY RATING; ''' SELECT_PROJECTION_FOR_MOVIE = ''' SELECT p.*, COUNT(r.ROW * r.COL) FROM PROJECTION as p LEFT JOIN RESERVATION as r ON r.PROJECTION_ID = p.ID WHERE p.MOVIE_ID = ? GROUP BY p.ID ORDER BY p.DATE; ''' SELECT_PROJECTION_ORDERED_BY_DATE = ''' SELECT PROJECTION.*, COUNT(RESERVATION.ROW * RESERVATION.COL) FROM PROJECTION LEFT JOIN RESERVATION ON RESERVATION.PROJECTION_ID = PROJECTION.ID WHERE PROJECTION.MOVIE_ID = ? AND PROJECTION.DATE BETWEEN ? AND ? GROUP BY PROJECTION.ID ORDER BY PROJECTION.DATE; ''' IS_USER_IN_USERS = ''' SELECT * FROM USERS WHERE USERNAME = ? and PASSWORD LIKE ?; ''' SELECT_AVAILABLE_SEATS = ''' SELECT ROW, COL FROM RESERVATIONS WHERE PROJECTION_ID = ? AND ROW IS NOT NULL AND COL IS NOT NULL; ''' SELECT_PROJECTION_BY_ID = ''' SELECT * FROM PROJECTION as p WHERE p.ID = ? ''' SELECT_MOVIE_BY_ID = ''' SELECT * FROM MOVIE as m WHERE m.ID = ? '''
SELECT_MOVIES_ORDERED_BY_RATING = ''' SELECT * FROM MOVIE ORDER BY RATING; ''' SELECT_PROJECTION_FOR_MOVIE = ''' SELECT p.*, COUNT(r.ROW * r.COL) FROM PROJECTION as p LEFT JOIN RESERVATION as r ON r.PROJECTION_ID = p.ID WHERE p.MOVIE_ID = ? GROUP BY p.ID ORDER BY p.DATE; ''' SELECT_PROJECTION_ORDERED_BY_DATE = ''' SELECT PROJECTION.*, COUNT(RESERVATION.ROW * RESERVATION.COL) FROM PROJECTION LEFT JOIN RESERVATION ON RESERVATION.PROJECTION_ID = PROJECTION.ID WHERE PROJECTION.MOVIE_ID = ? AND PROJECTION.DATE BETWEEN ? AND ? GROUP BY PROJECTION.ID ORDER BY PROJECTION.DATE; ''' IS_USER_IN_USERS = ''' SELECT * FROM USER WHERE USERNAME = ? and PASSWORD LIKE ?; ''' SELECT_AVAILABLE_SEATS = ''' SELECT ROW, COL FROM RESERVATION WHERE PROJECTION_ID = ? AND ROW IS NOT NULL AND COL IS NOT NULL; ''' SELECT_PROJECTION_BY_ID = ''' SELECT * FROM PROJECTION as p WHERE p.ID = ? ''' SELECT_MOVIE_BY_ID = ''' SELECT * FROM MOVIE as m WHERE m.ID = ? '''
Select Queries Table names changed
Select Queries Table names changed
Python
mit
BrickText/JHROM
--- +++ @@ -27,13 +27,13 @@ IS_USER_IN_USERS = ''' SELECT * - FROM USERS + FROM USER WHERE USERNAME = ? and PASSWORD LIKE ?; ''' SELECT_AVAILABLE_SEATS = ''' SELECT ROW, COL - FROM RESERVATIONS + FROM RESERVATION WHERE PROJECTION_ID = ? AND ROW IS NOT NULL AND COL IS NOT NULL; '''
01cd080395533b9e8d53f4c203ef6be185d97ebc
dbaas/integrations/iaas/manager.py
dbaas/integrations/iaas/manager.py
from dbaas_cloudstack.provider import CloudStackProvider from pre_provisioned.pre_provisioned_provider import PreProvisionedProvider import logging LOG = logging.getLogger(__name__) class IaaSManager(): @classmethod def destroy_instance(cls, database, *args, **kwargs): plan = database.plan provider = plan.provider if provider == plan.PREPROVISIONED: LOG.info("Destroying pre provisioned database...") PreProvisionedProvider().destroy_instance(database, *args, **kwargs) elif provider == plan.CLOUDSTACK: LOG.info("Destroying cloud stack instance...") CloudStackProvider().destroy_instance(database, *args, **kwargs) @classmethod def create_instance(cls, plan, environment, name): if plan.provider == plan.PREPROVISIONED: LOG.info("Creating pre provisioned instance...") return PreProvisionedProvider().create_instance(plan, environment) elif plan.provider == plan.CLOUDSTACK: LOG.info("Creating cloud stack instance...") return CloudStackProvider().create_instance(plan, environment, name)
from dbaas_cloudstack.provider import CloudStackProvider from pre_provisioned.pre_provisioned_provider import PreProvisionedProvider from integrations.monitoring.manager import MonitoringManager import logging LOG = logging.getLogger(__name__) class IaaSManager(): @classmethod def destroy_instance(cls, database, *args, **kwargs): plan = database.plan provider = plan.provider if provider == plan.PREPROVISIONED: LOG.info("Destroying pre provisioned database...") PreProvisionedProvider().destroy_instance(database, *args, **kwargs) elif provider == plan.CLOUDSTACK: LOG.info("Destroying cloud stack instance...") MonitoringManager.remove_monitoring(database.databaseinfra) CloudStackProvider().destroy_instance(database, *args, **kwargs) @classmethod def create_instance(cls, plan, environment, name): if plan.provider == plan.PREPROVISIONED: LOG.info("Creating pre provisioned instance...") return PreProvisionedProvider().create_instance(plan, environment) elif plan.provider == plan.CLOUDSTACK: LOG.info("Creating cloud stack instance...") databaseinfra = CloudStackProvider().create_instance(plan, environment, name) if databaseinfra is not None: MonitoringManager.create_monitoring(databaseinfra) return databaseinfra
Add call to monitoring app
Add call to monitoring app
Python
bsd-3-clause
globocom/database-as-a-service,globocom/database-as-a-service,globocom/database-as-a-service,globocom/database-as-a-service
--- +++ @@ -1,5 +1,6 @@ from dbaas_cloudstack.provider import CloudStackProvider from pre_provisioned.pre_provisioned_provider import PreProvisionedProvider +from integrations.monitoring.manager import MonitoringManager import logging LOG = logging.getLogger(__name__) @@ -15,6 +16,7 @@ PreProvisionedProvider().destroy_instance(database, *args, **kwargs) elif provider == plan.CLOUDSTACK: LOG.info("Destroying cloud stack instance...") + MonitoringManager.remove_monitoring(database.databaseinfra) CloudStackProvider().destroy_instance(database, *args, **kwargs) @classmethod @@ -24,4 +26,7 @@ return PreProvisionedProvider().create_instance(plan, environment) elif plan.provider == plan.CLOUDSTACK: LOG.info("Creating cloud stack instance...") - return CloudStackProvider().create_instance(plan, environment, name) + databaseinfra = CloudStackProvider().create_instance(plan, environment, name) + if databaseinfra is not None: + MonitoringManager.create_monitoring(databaseinfra) + return databaseinfra
2313424f811c59563090e77966d906dd3eb7f127
tools/buildbot.py
tools/buildbot.py
import os import sys def usage(): print '%s all -- build all bsp' % os.path.basename(sys.argv[0]) print '%s clean -- clean all bsp' % os.path.basename(sys.argv[0]) print '%s project -- update all prject files' % os.path.basename(sys.argv[0]) BSP_ROOT = '../bsp' if len(sys.argv) != 2: usage() sys.exit(0) # get command options command = '' if sys.argv[1] == 'all': command = ' ' elif sys.argv[1] == 'clean': command = ' -c' elif sys.argv[1] == 'project': command = ' --target=mdk -s' else: usage() sys.exit(0) projects = os.listdir(BSP_ROOT) for item in projects: project_dir = os.path.join(BSP_ROOT, item) if os.path.isfile(os.path.join(project_dir, 'SConstruct')): if os.system('scons --directory=' + project_dir + command) != 0: print 'build failed!!' break
import os import sys def usage(): print '%s all -- build all bsp' % os.path.basename(sys.argv[0]) print '%s clean -- clean all bsp' % os.path.basename(sys.argv[0]) print '%s project -- update all prject files' % os.path.basename(sys.argv[0]) BSP_ROOT = '../bsp' if len(sys.argv) != 2: usage() sys.exit(0) # get command options command = '' if sys.argv[1] == 'all': command = ' ' elif sys.argv[1] == 'clean': command = ' -c' elif sys.argv[1] == 'project': command = ' --target=mdk -s' projects = os.listdir(BSP_ROOT) for item in projects: project_dir = os.path.join(BSP_ROOT, item) if os.path.isfile(os.path.join(project_dir, 'template.uvproj')): print ('prepare MDK project file on ' + project_dir) os.system('scons --directory=' + project_dir + command) sys.exit(0) else: usage() sys.exit(0) projects = os.listdir(BSP_ROOT) for item in projects: project_dir = os.path.join(BSP_ROOT, item) if os.path.isfile(os.path.join(project_dir, 'SConstruct')): if os.system('scons --directory=' + project_dir + command) != 0: print 'build failed!!' break
Add better way to generate MDK project file.
[tools] Add better way to generate MDK project file.
Python
apache-2.0
RT-Thread/rt-thread,FlyLu/rt-thread,weiyuliang/rt-thread,ArdaFu/rt-thread,weiyuliang/rt-thread,geniusgogo/rt-thread,weety/rt-thread,igou/rt-thread,weiyuliang/rt-thread,hezlog/rt-thread,armink/rt-thread,wolfgangz2013/rt-thread,armink/rt-thread,ArdaFu/rt-thread,hezlog/rt-thread,AubrCool/rt-thread,geniusgogo/rt-thread,yongli3/rt-thread,hezlog/rt-thread,nongxiaoming/rt-thread,weety/rt-thread,igou/rt-thread,zhaojuntao/rt-thread,zhaojuntao/rt-thread,yongli3/rt-thread,gbcwbz/rt-thread,zhaojuntao/rt-thread,RT-Thread/rt-thread,igou/rt-thread,nongxiaoming/rt-thread,weety/rt-thread,zhaojuntao/rt-thread,FlyLu/rt-thread,armink/rt-thread,wolfgangz2013/rt-thread,gbcwbz/rt-thread,igou/rt-thread,nongxiaoming/rt-thread,geniusgogo/rt-thread,armink/rt-thread,weiyuliang/rt-thread,zhaojuntao/rt-thread,AubrCool/rt-thread,zhaojuntao/rt-thread,wolfgangz2013/rt-thread,hezlog/rt-thread,weety/rt-thread,hezlog/rt-thread,geniusgogo/rt-thread,geniusgogo/rt-thread,FlyLu/rt-thread,wolfgangz2013/rt-thread,ArdaFu/rt-thread,hezlog/rt-thread,igou/rt-thread,RT-Thread/rt-thread,FlyLu/rt-thread,yongli3/rt-thread,weiyuliang/rt-thread,nongxiaoming/rt-thread,gbcwbz/rt-thread,yongli3/rt-thread,armink/rt-thread,zhaojuntao/rt-thread,RT-Thread/rt-thread,nongxiaoming/rt-thread,geniusgogo/rt-thread,gbcwbz/rt-thread,yongli3/rt-thread,igou/rt-thread,RT-Thread/rt-thread,wolfgangz2013/rt-thread,yongli3/rt-thread,wolfgangz2013/rt-thread,weety/rt-thread,igou/rt-thread,ArdaFu/rt-thread,gbcwbz/rt-thread,RT-Thread/rt-thread,weiyuliang/rt-thread,FlyLu/rt-thread,FlyLu/rt-thread,yongli3/rt-thread,AubrCool/rt-thread,RT-Thread/rt-thread,geniusgogo/rt-thread,FlyLu/rt-thread,weety/rt-thread,armink/rt-thread,gbcwbz/rt-thread,weety/rt-thread,ArdaFu/rt-thread,nongxiaoming/rt-thread,gbcwbz/rt-thread,AubrCool/rt-thread,armink/rt-thread,nongxiaoming/rt-thread,ArdaFu/rt-thread,wolfgangz2013/rt-thread,AubrCool/rt-thread,AubrCool/rt-thread,ArdaFu/rt-thread,weiyuliang/rt-thread,hezlog/rt-thread,AubrCool/rt-thread
--- +++ @@ -19,6 +19,17 @@ command = ' -c' elif sys.argv[1] == 'project': command = ' --target=mdk -s' + + projects = os.listdir(BSP_ROOT) + for item in projects: + project_dir = os.path.join(BSP_ROOT, item) + + if os.path.isfile(os.path.join(project_dir, 'template.uvproj')): + print ('prepare MDK project file on ' + project_dir) + + os.system('scons --directory=' + project_dir + command) + + sys.exit(0) else: usage() sys.exit(0)
155822548be11161aefdb0d93d5ec86095ab3624
rt.py
rt.py
import queue import threading def loop(queue, actor): while True: message = queue.get() actor.behavior(message) class Actor(object): def __init__(self): pass def _start_loop(self): self.queue = queue.Queue() self.dispatcher = threading.Thread( target=loop, args=(self.queue, self)) self.dispatcher.start() def __call__(self, message): self.queue.put(message) @classmethod def create(cls, *args): actor = cls(*args) actor._start_loop() return actor
import queue import threading def indiviual_loop(queue, actor): while True: message = queue.get() actor.behavior(message) def global_loop(queue): while True: actor, message = queue.get() actor.behavior(message) class EventLoop(object): loop = None def __init__(self): self.queue = queue.Queue() self.thread = threading.Thread( target=global_loop, args=(self.queue,), name='global-loop') self.thread.start() def schedule(self, message, target): self.queue.put((target, message)) @classmethod def get_loop(cls): if cls.loop is None: cls.loop = cls() return cls.loop class AbstractActor(object): def __call__(self, message): self._put(message) def _put(self, message): raise NotImplementedError() def _ensure_loop(self): pass @classmethod def create(cls, *args): actor = cls(*args) actor._ensure_loop() return actor class ActorOwnLoop(AbstractActor): def _put(self, message): self.queue.put(message) def _ensure_loop(self): self.queue = queue.Queue() self.dispatcher = threading.Thread( target=indiviual_loop, args=(self.queue, self), name=self._thread_name()) self.dispatcher.start() def _thread_name(self): return '{}-{}'.format( self.__class__.__name__, hex(id(self))) class ActorGlobalLoop(AbstractActor): def _put(self, message): self.loop.schedule(message, self) def _ensure_loop(self): self.loop = EventLoop.get_loop() Actor = ActorGlobalLoop
Refactor to allow different event loops
Refactor to allow different event loops
Python
mit
waltermoreira/tartpy
--- +++ @@ -2,29 +2,84 @@ import threading -def loop(queue, actor): +def indiviual_loop(queue, actor): while True: message = queue.get() actor.behavior(message) + +def global_loop(queue): + while True: + actor, message = queue.get() + actor.behavior(message) + + +class EventLoop(object): + + loop = None + + def __init__(self): + self.queue = queue.Queue() + self.thread = threading.Thread( + target=global_loop, + args=(self.queue,), + name='global-loop') + self.thread.start() + + def schedule(self, message, target): + self.queue.put((target, message)) -class Actor(object): + @classmethod + def get_loop(cls): + if cls.loop is None: + cls.loop = cls() + return cls.loop + - def __init__(self): +class AbstractActor(object): + + def __call__(self, message): + self._put(message) + + def _put(self, message): + raise NotImplementedError() + + def _ensure_loop(self): pass - def _start_loop(self): - self.queue = queue.Queue() - self.dispatcher = threading.Thread( - target=loop, - args=(self.queue, self)) - self.dispatcher.start() - - def __call__(self, message): - self.queue.put(message) - @classmethod def create(cls, *args): actor = cls(*args) - actor._start_loop() + actor._ensure_loop() return actor + + +class ActorOwnLoop(AbstractActor): + + def _put(self, message): + self.queue.put(message) + + def _ensure_loop(self): + self.queue = queue.Queue() + self.dispatcher = threading.Thread( + target=indiviual_loop, + args=(self.queue, self), + name=self._thread_name()) + self.dispatcher.start() + + def _thread_name(self): + return '{}-{}'.format( + self.__class__.__name__, + hex(id(self))) + + +class ActorGlobalLoop(AbstractActor): + + def _put(self, message): + self.loop.schedule(message, self) + + def _ensure_loop(self): + self.loop = EventLoop.get_loop() + + +Actor = ActorGlobalLoop
ef73cc38e08db709a33321cde47ceb6dfce3504d
chrome/test/functional/chromeos_private_view.py
chrome/test/functional/chromeos_private_view.py
#!/usr/bin/env python # Copyright (c) 2012 The Chromium Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. import os import pyauto_functional # must be imported before pyauto import pyauto class ChromeosPrivateViewTest(pyauto.PyUITest): """Basic tests for ChromeOS Private View. Requires ChromeOS to be logged in. """ def _GetExtensionInfoById(self, extensions, id): for x in extensions: if x['id'] == id: return x return None def testInstallPrivateViewExtension(self): """Basic installation test for Private View on ChromeOS.""" crx_file_path = os.path.abspath( os.path.join(self.DataDir(), 'pyauto_private', 'apps', 'privateview-chrome-1.0.800_RELEASE.crx')) ext_id = self.InstallExtension(crx_file_path) self.assertTrue(ext_id, 'Failed to install extension.') extension = self._GetExtensionInfoById(self.GetExtensionsInfo(), ext_id) self.assertTrue(extension['is_enabled'], msg='Extension was not enabled on installation') self.assertFalse(extension['allowed_in_incognito'], msg='Extension was allowed in incognito on installation.') if __name__ == '__main__': pyauto_functional.Main()
#!/usr/bin/env python # Copyright (c) 2012 The Chromium Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. import os import pyauto_functional # must be imported before pyauto import pyauto class ChromeosPrivateViewTest(pyauto.PyUITest): """Basic tests for ChromeOS Private View. Requires ChromeOS to be logged in. """ def _GetExtensionInfoById(self, extensions, id): for x in extensions: if x['id'] == id: return x return None def testInstallPrivateViewExtension(self): """Basic installation test for Private View on ChromeOS.""" crx_file_path = os.path.abspath( os.path.join(self.DataDir(), 'pyauto_private', 'apps', 'privateview-chrome-2.2.28_RELEASE.crx')) ext_id = self.InstallExtension(crx_file_path) self.assertTrue(ext_id, 'Failed to install extension.') extension = self._GetExtensionInfoById(self.GetExtensionsInfo(), ext_id) self.assertTrue(extension['is_enabled'], msg='Extension was not enabled on installation') self.assertFalse(extension['allowed_in_incognito'], msg='Extension was allowed in incognito on installation.') if __name__ == '__main__': pyauto_functional.Main()
Modify pyauto test ChromeosPrivateViewTest to use 2.2.28 data file.
Modify pyauto test ChromeosPrivateViewTest to use 2.2.28 data file. BUG=none TEST=This is a test. Review URL: https://chromiumcodereview.appspot.com/10389084 git-svn-id: de016e52bd170d2d4f2344f9bf92d50478b649e0@136454 0039d316-1c4b-4281-b951-d872f2087c98
Python
bsd-3-clause
TheTypoMaster/chromium-crosswalk,dednal/chromium.src,zcbenz/cefode-chromium,chuan9/chromium-crosswalk,Fireblend/chromium-crosswalk,ChromiumWebApps/chromium,timopulkkinen/BubbleFish,Jonekee/chromium.src,zcbenz/cefode-chromium,krieger-od/nwjs_chromium.src,Chilledheart/chromium,TheTypoMaster/chromium-crosswalk,hujiajie/pa-chromium,jaruba/chromium.src,Just-D/chromium-1,hgl888/chromium-crosswalk-efl,Just-D/chromium-1,markYoungH/chromium.src,hujiajie/pa-chromium,axinging/chromium-crosswalk,PeterWangIntel/chromium-crosswalk,axinging/chromium-crosswalk,dednal/chromium.src,Fireblend/chromium-crosswalk,chuan9/chromium-crosswalk,axinging/chromium-crosswalk,M4sse/chromium.src,PeterWangIntel/chromium-crosswalk,junmin-zhu/chromium-rivertrail,junmin-zhu/chromium-rivertrail,krieger-od/nwjs_chromium.src,keishi/chromium,timopulkkinen/BubbleFish,crosswalk-project/chromium-crosswalk-efl,M4sse/chromium.src,mogoweb/chromium-crosswalk,ChromiumWebApps/chromium,Jonekee/chromium.src,mohamed--abdel-maksoud/chromium.src,Just-D/chromium-1,ChromiumWebApps/chromium,hgl888/chromium-crosswalk-efl,crosswalk-project/chromium-crosswalk-efl,pozdnyakov/chromium-crosswalk,ltilve/chromium,patrickm/chromium.src,mohamed--abdel-maksoud/chromium.src,krieger-od/nwjs_chromium.src,PeterWangIntel/chromium-crosswalk,PeterWangIntel/chromium-crosswalk,dushu1203/chromium.src,jaruba/chromium.src,pozdnyakov/chromium-crosswalk,ChromiumWebApps/chromium,hgl888/chromium-crosswalk,robclark/chromium,Pluto-tv/chromium-crosswalk,dednal/chromium.src,bright-sparks/chromium-spacewalk,patrickm/chromium.src,krieger-od/nwjs_chromium.src,ltilve/chromium,ChromiumWebApps/chromium,krieger-od/nwjs_chromium.src,M4sse/chromium.src,nacl-webkit/chrome_deps,crosswalk-project/chromium-crosswalk-efl,chuan9/chromium-crosswalk,ltilve/chromium,Chilledheart/chromium,markYoungH/chromium.src,TheTypoMaster/chromium-crosswalk,hgl888/chromium-crosswalk,bright-sparks/chromium-spacewalk,robclark/chromium,ondra-novak/chromium.src,ltilve/chromium,Jonekee/chromium.src,hgl888/chromium-crosswalk,jaruba/chromium.src,ChromiumWebApps/chromium,pozdnyakov/chromium-crosswalk,PeterWangIntel/chromium-crosswalk,Jonekee/chromium.src,Jonekee/chromium.src,M4sse/chromium.src,hgl888/chromium-crosswalk-efl,crosswalk-project/chromium-crosswalk-efl,fujunwei/chromium-crosswalk,jaruba/chromium.src,dushu1203/chromium.src,keishi/chromium,hgl888/chromium-crosswalk-efl,axinging/chromium-crosswalk,dushu1203/chromium.src,hgl888/chromium-crosswalk-efl,Just-D/chromium-1,dushu1203/chromium.src,dushu1203/chromium.src,robclark/chromium,hujiajie/pa-chromium,bright-sparks/chromium-spacewalk,ltilve/chromium,Just-D/chromium-1,mohamed--abdel-maksoud/chromium.src,Chilledheart/chromium,Pluto-tv/chromium-crosswalk,hujiajie/pa-chromium,markYoungH/chromium.src,chuan9/chromium-crosswalk,fujunwei/chromium-crosswalk,mogoweb/chromium-crosswalk,Fireblend/chromium-crosswalk,keishi/chromium,ChromiumWebApps/chromium,junmin-zhu/chromium-rivertrail,nacl-webkit/chrome_deps,Jonekee/chromium.src,littlstar/chromium.src,junmin-zhu/chromium-rivertrail,dushu1203/chromium.src,junmin-zhu/chromium-rivertrail,mogoweb/chromium-crosswalk,Chilledheart/chromium,keishi/chromium,krieger-od/nwjs_chromium.src,TheTypoMaster/chromium-crosswalk,ondra-novak/chromium.src,markYoungH/chromium.src,pozdnyakov/chromium-crosswalk,timopulkkinen/BubbleFish,zcbenz/cefode-chromium,timopulkkinen/BubbleFish,junmin-zhu/chromium-rivertrail,nacl-webkit/chrome_deps,krieger-od/nwjs_chromium.src,hgl888/chromium-crosswalk,robclark/chromium,mohamed--abdel-maksoud/chromium.src,TheTypoMaster/chromium-crosswalk,bright-sparks/chromium-spacewalk,fujunwei/chromium-crosswalk,anirudhSK/chromium,keishi/chromium,axinging/chromium-crosswalk,M4sse/chromium.src,fujunwei/chromium-crosswalk,TheTypoMaster/chromium-crosswalk,Pluto-tv/chromium-crosswalk,bright-sparks/chromium-spacewalk,dednal/chromium.src,keishi/chromium,zcbenz/cefode-chromium,markYoungH/chromium.src,timopulkkinen/BubbleFish,anirudhSK/chromium,mogoweb/chromium-crosswalk,hgl888/chromium-crosswalk-efl,dushu1203/chromium.src,PeterWangIntel/chromium-crosswalk,Jonekee/chromium.src,hgl888/chromium-crosswalk,littlstar/chromium.src,jaruba/chromium.src,ChromiumWebApps/chromium,keishi/chromium,TheTypoMaster/chromium-crosswalk,mogoweb/chromium-crosswalk,timopulkkinen/BubbleFish,axinging/chromium-crosswalk,markYoungH/chromium.src,Jonekee/chromium.src,Fireblend/chromium-crosswalk,mohamed--abdel-maksoud/chromium.src,patrickm/chromium.src,M4sse/chromium.src,M4sse/chromium.src,Chilledheart/chromium,markYoungH/chromium.src,Chilledheart/chromium,mogoweb/chromium-crosswalk,hujiajie/pa-chromium,dednal/chromium.src,mohamed--abdel-maksoud/chromium.src,junmin-zhu/chromium-rivertrail,Chilledheart/chromium,zcbenz/cefode-chromium,axinging/chromium-crosswalk,patrickm/chromium.src,mohamed--abdel-maksoud/chromium.src,ltilve/chromium,anirudhSK/chromium,mohamed--abdel-maksoud/chromium.src,keishi/chromium,hujiajie/pa-chromium,littlstar/chromium.src,nacl-webkit/chrome_deps,chuan9/chromium-crosswalk,bright-sparks/chromium-spacewalk,markYoungH/chromium.src,zcbenz/cefode-chromium,dednal/chromium.src,zcbenz/cefode-chromium,pozdnyakov/chromium-crosswalk,anirudhSK/chromium,hujiajie/pa-chromium,fujunwei/chromium-crosswalk,Just-D/chromium-1,timopulkkinen/BubbleFish,littlstar/chromium.src,Jonekee/chromium.src,chuan9/chromium-crosswalk,ondra-novak/chromium.src,anirudhSK/chromium,Just-D/chromium-1,krieger-od/nwjs_chromium.src,robclark/chromium,dushu1203/chromium.src,axinging/chromium-crosswalk,robclark/chromium,timopulkkinen/BubbleFish,Fireblend/chromium-crosswalk,nacl-webkit/chrome_deps,jaruba/chromium.src,fujunwei/chromium-crosswalk,mohamed--abdel-maksoud/chromium.src,ltilve/chromium,patrickm/chromium.src,ondra-novak/chromium.src,nacl-webkit/chrome_deps,hgl888/chromium-crosswalk,TheTypoMaster/chromium-crosswalk,jaruba/chromium.src,PeterWangIntel/chromium-crosswalk,timopulkkinen/BubbleFish,robclark/chromium,chuan9/chromium-crosswalk,Jonekee/chromium.src,junmin-zhu/chromium-rivertrail,ChromiumWebApps/chromium,patrickm/chromium.src,pozdnyakov/chromium-crosswalk,hgl888/chromium-crosswalk-efl,axinging/chromium-crosswalk,Pluto-tv/chromium-crosswalk,patrickm/chromium.src,robclark/chromium,hgl888/chromium-crosswalk,junmin-zhu/chromium-rivertrail,dednal/chromium.src,Just-D/chromium-1,mogoweb/chromium-crosswalk,Pluto-tv/chromium-crosswalk,anirudhSK/chromium,zcbenz/cefode-chromium,markYoungH/chromium.src,Chilledheart/chromium,PeterWangIntel/chromium-crosswalk,anirudhSK/chromium,keishi/chromium,dednal/chromium.src,Jonekee/chromium.src,Fireblend/chromium-crosswalk,nacl-webkit/chrome_deps,jaruba/chromium.src,chuan9/chromium-crosswalk,pozdnyakov/chromium-crosswalk,hgl888/chromium-crosswalk-efl,mohamed--abdel-maksoud/chromium.src,nacl-webkit/chrome_deps,M4sse/chromium.src,jaruba/chromium.src,ltilve/chromium,littlstar/chromium.src,ondra-novak/chromium.src,ondra-novak/chromium.src,littlstar/chromium.src,hgl888/chromium-crosswalk-efl,Pluto-tv/chromium-crosswalk,ChromiumWebApps/chromium,zcbenz/cefode-chromium,keishi/chromium,junmin-zhu/chromium-rivertrail,zcbenz/cefode-chromium,littlstar/chromium.src,ltilve/chromium,hujiajie/pa-chromium,dednal/chromium.src,axinging/chromium-crosswalk,krieger-od/nwjs_chromium.src,jaruba/chromium.src,ChromiumWebApps/chromium,hujiajie/pa-chromium,pozdnyakov/chromium-crosswalk,Pluto-tv/chromium-crosswalk,nacl-webkit/chrome_deps,keishi/chromium,anirudhSK/chromium,anirudhSK/chromium,M4sse/chromium.src,fujunwei/chromium-crosswalk,timopulkkinen/BubbleFish,PeterWangIntel/chromium-crosswalk,anirudhSK/chromium,littlstar/chromium.src,pozdnyakov/chromium-crosswalk,M4sse/chromium.src,chuan9/chromium-crosswalk,nacl-webkit/chrome_deps,dushu1203/chromium.src,crosswalk-project/chromium-crosswalk-efl,hgl888/chromium-crosswalk,mogoweb/chromium-crosswalk,robclark/chromium,krieger-od/nwjs_chromium.src,crosswalk-project/chromium-crosswalk-efl,markYoungH/chromium.src,crosswalk-project/chromium-crosswalk-efl,jaruba/chromium.src,dednal/chromium.src,mohamed--abdel-maksoud/chromium.src,hujiajie/pa-chromium,anirudhSK/chromium,bright-sparks/chromium-spacewalk,ondra-novak/chromium.src,bright-sparks/chromium-spacewalk,crosswalk-project/chromium-crosswalk-efl,junmin-zhu/chromium-rivertrail,ondra-novak/chromium.src,axinging/chromium-crosswalk,Fireblend/chromium-crosswalk,patrickm/chromium.src,markYoungH/chromium.src,dushu1203/chromium.src,krieger-od/nwjs_chromium.src,mogoweb/chromium-crosswalk,zcbenz/cefode-chromium,Chilledheart/chromium,ChromiumWebApps/chromium,ondra-novak/chromium.src,M4sse/chromium.src,TheTypoMaster/chromium-crosswalk,Pluto-tv/chromium-crosswalk,Pluto-tv/chromium-crosswalk,crosswalk-project/chromium-crosswalk-efl,Fireblend/chromium-crosswalk,pozdnyakov/chromium-crosswalk,hujiajie/pa-chromium,robclark/chromium,hgl888/chromium-crosswalk-efl,nacl-webkit/chrome_deps,timopulkkinen/BubbleFish,pozdnyakov/chromium-crosswalk,Fireblend/chromium-crosswalk,anirudhSK/chromium,hgl888/chromium-crosswalk,fujunwei/chromium-crosswalk,dushu1203/chromium.src,mogoweb/chromium-crosswalk,patrickm/chromium.src,fujunwei/chromium-crosswalk,dednal/chromium.src,Just-D/chromium-1,bright-sparks/chromium-spacewalk
--- +++ @@ -25,7 +25,7 @@ """Basic installation test for Private View on ChromeOS.""" crx_file_path = os.path.abspath( os.path.join(self.DataDir(), 'pyauto_private', 'apps', - 'privateview-chrome-1.0.800_RELEASE.crx')) + 'privateview-chrome-2.2.28_RELEASE.crx')) ext_id = self.InstallExtension(crx_file_path) self.assertTrue(ext_id, 'Failed to install extension.') extension = self._GetExtensionInfoById(self.GetExtensionsInfo(), ext_id)
db43b3b3079842fb2baf6d181ef39374acf0053c
.gitlab/linters/check-makefiles.py
.gitlab/linters/check-makefiles.py
#!/usr/bin/env python3 """ Warn for use of `--interactive` inside Makefiles (#11468). Encourage the use of `$(TEST_HC_OPTS_INTERACTIVE)` instead of `$(TEST_HC_OPTS) --interactive -ignore-dot-ghci -v0`. It's too easy to forget one of those flags when adding a new test. """ from linter import run_linters, RegexpLinter linters = [ RegexpLinter(r'--interactive', message = "Warning: Use `$(TEST_HC_OPTS_INTERACTIVE)` instead of `--interactive -ignore-dot-ghci -v0`." ).add_path_filter(lambda path: path.name == 'Makefile') ] if __name__ == '__main__': run_linters(linters, subdir='testsuite')
#!/usr/bin/env python3 """ Linters for testsuite makefiles """ from linter import run_linters, RegexpLinter """ Warn for use of `--interactive` inside Makefiles (#11468). Encourage the use of `$(TEST_HC_OPTS_INTERACTIVE)` instead of `$(TEST_HC_OPTS) --interactive -ignore-dot-ghci -v0`. It's too easy to forget one of those flags when adding a new test. """ interactive_linter = \ RegexpLinter(r'--interactive', message = "Warning: Use `$(TEST_HC_OPTS_INTERACTIVE)` instead of `--interactive -ignore-dot-ghci -v0`." ).add_path_filter(lambda path: path.name == 'Makefile') test_hc_quotes_linter = \ RegexpLinter('\t\\$\\(TEST_HC\\)', message = "Warning: $(TEST_HC) should be quoted in Makefiles.", ).add_path_filter(lambda path: path.name == 'Makefile') linters = [ interactive_linter, test_hc_quotes_linter, ] if __name__ == '__main__': run_linters(linters, subdir='testsuite')
Add linter to catch unquoted use of $(TEST_HC)
linters: Add linter to catch unquoted use of $(TEST_HC) This is a common bug that creeps into Makefiles (e.g. see T12674).
Python
bsd-3-clause
sdiehl/ghc,sdiehl/ghc,sdiehl/ghc,sdiehl/ghc,sdiehl/ghc,sdiehl/ghc,sdiehl/ghc
--- +++ @@ -1,4 +1,10 @@ #!/usr/bin/env python3 + +""" +Linters for testsuite makefiles +""" + +from linter import run_linters, RegexpLinter """ Warn for use of `--interactive` inside Makefiles (#11468). @@ -7,13 +13,19 @@ `$(TEST_HC_OPTS) --interactive -ignore-dot-ghci -v0`. It's too easy to forget one of those flags when adding a new test. """ - -from linter import run_linters, RegexpLinter - -linters = [ +interactive_linter = \ RegexpLinter(r'--interactive', message = "Warning: Use `$(TEST_HC_OPTS_INTERACTIVE)` instead of `--interactive -ignore-dot-ghci -v0`." ).add_path_filter(lambda path: path.name == 'Makefile') + +test_hc_quotes_linter = \ + RegexpLinter('\t\\$\\(TEST_HC\\)', + message = "Warning: $(TEST_HC) should be quoted in Makefiles.", + ).add_path_filter(lambda path: path.name == 'Makefile') + +linters = [ + interactive_linter, + test_hc_quotes_linter, ] if __name__ == '__main__':
dc1a7bc4d674fd6e7235222612f1d147112d77db
src/nodeconductor_assembly_waldur/packages/migrations/0002_openstack_packages.py
src/nodeconductor_assembly_waldur/packages/migrations/0002_openstack_packages.py
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import migrations, models import nodeconductor.core.fields class Migration(migrations.Migration): dependencies = [ ('openstack', '0022_volume_device'), ('structure', '0037_remove_customer_billing_backend_id'), ('packages', '0001_initial'), ] operations = [ migrations.CreateModel( name='OpenStackPackage', fields=[ ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)), ('uuid', nodeconductor.core.fields.UUIDField()), ('service_settings', models.ForeignKey(related_name='+', to='structure.ServiceSettings')), ], options={ 'abstract': False, }, ), migrations.RemoveField( model_name='packagetemplate', name='type', ), migrations.AddField( model_name='packagetemplate', name='service_settings', field=models.ForeignKey(related_name='+', default=1, to='structure.ServiceSettings'), preserve_default=False, ), migrations.AddField( model_name='openstackpackage', name='template', field=models.ForeignKey(related_name='openstack_packages', to='packages.PackageTemplate', help_text='Tenant will be created based on this template.'), ), migrations.AddField( model_name='openstackpackage', name='tenant', field=models.ForeignKey(related_name='+', to='openstack.Tenant'), ), ]
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import migrations, models import nodeconductor.core.fields class Migration(migrations.Migration): dependencies = [ ('openstack', '0022_volume_device'), ('structure', '0037_remove_customer_billing_backend_id'), ('packages', '0001_initial'), ] operations = [ migrations.CreateModel( name='OpenStackPackage', fields=[ ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)), ('uuid', nodeconductor.core.fields.UUIDField()), ('service_settings', models.ForeignKey(related_name='+', to='structure.ServiceSettings')), ], options={ 'abstract': False, }, ), migrations.RemoveField( model_name='packagetemplate', name='type', ), migrations.AddField( model_name='packagetemplate', name='service_settings', field=models.ForeignKey(related_name='+', to='structure.ServiceSettings'), preserve_default=False, ), migrations.AddField( model_name='openstackpackage', name='template', field=models.ForeignKey(related_name='openstack_packages', to='packages.PackageTemplate', help_text='Tenant will be created based on this template.'), ), migrations.AddField( model_name='openstackpackage', name='tenant', field=models.ForeignKey(related_name='+', to='openstack.Tenant'), ), ]
Remove useless default from migration
Remove useless default from migration - wal-26
Python
mit
opennode/waldur-mastermind,opennode/nodeconductor-assembly-waldur,opennode/nodeconductor-assembly-waldur,opennode/nodeconductor-assembly-waldur,opennode/waldur-mastermind,opennode/waldur-mastermind,opennode/waldur-mastermind
--- +++ @@ -32,7 +32,7 @@ migrations.AddField( model_name='packagetemplate', name='service_settings', - field=models.ForeignKey(related_name='+', default=1, to='structure.ServiceSettings'), + field=models.ForeignKey(related_name='+', to='structure.ServiceSettings'), preserve_default=False, ), migrations.AddField(
5e504ae82c37b47ad2c1781cc74d13cdccc72495
dthm4kaiako/config/__init__.py
dthm4kaiako/config/__init__.py
"""Configuration for Django system.""" __version__ = "0.16.2" __version_info__ = tuple( [ int(num) if num.isdigit() else num for num in __version__.replace("-", ".", 1).split(".") ] )
"""Configuration for Django system.""" __version__ = "0.16.3" __version_info__ = tuple( [ int(num) if num.isdigit() else num for num in __version__.replace("-", ".", 1).split(".") ] )
Increment version number to 0.16.3
Increment version number to 0.16.3
Python
mit
uccser/cs4teachers,uccser/cs4teachers,uccser/cs4teachers,uccser/cs4teachers
--- +++ @@ -1,6 +1,6 @@ """Configuration for Django system.""" -__version__ = "0.16.2" +__version__ = "0.16.3" __version_info__ = tuple( [ int(num) if num.isdigit() else num
3dbc981e62c2d153913557b62083f60888fa7e83
ynr/apps/ynr_refactoring/management/commands/ynr_refactoring_remove_legacy_IDs.py
ynr/apps/ynr_refactoring/management/commands/ynr_refactoring_remove_legacy_IDs.py
import json from django.core.management.base import BaseCommand from django.db import transaction from people.models import Person from candidates.views.version_data import get_change_metadata from popolo.models import Identifier class Command(BaseCommand): def handle(self, *args, **options): schemes = ("yournextmp-candidate", "popit-person") # We can't use the GFK any more because we just deleted it, but the # content is still there identifiers = Identifier.objects.filter(scheme__in=schemes).values_list( "object_id", flat=True ) for person in Person.objects.filter(pk__in=identifiers).filter(pk=502): with transaction.atomic(): meta_data = get_change_metadata( None, "Removing legacy identifiers" ) meta_data["username"] = "CandidateBot" person.record_version(meta_data) person.save()
import json from django.core.management.base import BaseCommand from django.db import transaction from people.models import Person from candidates.views.version_data import get_change_metadata from popolo.models import Identifier class Command(BaseCommand): def handle(self, *args, **options): schemes = ("yournextmp-candidate", "popit-person") # We can't use the GFK any more because we just deleted it, but the # content is still there identifiers = Identifier.objects.filter(scheme__in=schemes).values_list( "object_id", flat=True ) for person in Person.objects.filter(pk__in=identifiers): with transaction.atomic(): meta_data = get_change_metadata( None, "Removing legacy identifiers" ) meta_data["username"] = "CandidateBot" person.record_version(meta_data) person.save()
Remove IDs for all candidates, not just Zac
Remove IDs for all candidates, not just Zac
Python
agpl-3.0
DemocracyClub/yournextrepresentative,DemocracyClub/yournextrepresentative,DemocracyClub/yournextrepresentative
--- +++ @@ -17,7 +17,7 @@ identifiers = Identifier.objects.filter(scheme__in=schemes).values_list( "object_id", flat=True ) - for person in Person.objects.filter(pk__in=identifiers).filter(pk=502): + for person in Person.objects.filter(pk__in=identifiers): with transaction.atomic(): meta_data = get_change_metadata( None, "Removing legacy identifiers"
e5f4dc01e94694bf9bfcae3ecd6eca34a33a24eb
openquake/__init__.py
openquake/__init__.py
# -*- coding: utf-8 -*- # vim: tabstop=4 shiftwidth=4 softtabstop=4 # # Copyright (C) 2010-2016 GEM Foundation # # OpenQuake is free software: you can redistribute it and/or modify it # under the terms of the GNU Affero General Public License as published # by the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # OpenQuake is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with OpenQuake. If not, see <http://www.gnu.org/licenses/>. __import__('pkg_resources').declare_namespace(__name__)
# -*- coding: utf-8 -*- # vim: tabstop=4 shiftwidth=4 softtabstop=4 # # Copyright (C) 2010-2016 GEM Foundation # # OpenQuake is free software: you can redistribute it and/or modify it # under the terms of the GNU Affero General Public License as published # by the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # OpenQuake is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with OpenQuake. If not, see <http://www.gnu.org/licenses/>. # Make the namespace compatible with old setuptools, like the one # provided by QGIS 2.1x on Windows try: __import__('pkg_resources').declare_namespace(__name__) except ImportError: __path__ = __import__('pkgutil').extend_path(__path__, __name__)
Make the openquake namespace compatible with old setuptools
Make the openquake namespace compatible with old setuptools
Python
agpl-3.0
gem/oq-engine,gem/oq-engine,gem/oq-engine,gem/oq-engine,gem/oq-engine
--- +++ @@ -16,4 +16,9 @@ # You should have received a copy of the GNU Affero General Public License # along with OpenQuake. If not, see <http://www.gnu.org/licenses/>. -__import__('pkg_resources').declare_namespace(__name__) +# Make the namespace compatible with old setuptools, like the one +# provided by QGIS 2.1x on Windows +try: + __import__('pkg_resources').declare_namespace(__name__) +except ImportError: + __path__ = __import__('pkgutil').extend_path(__path__, __name__)
5a69162e82c2c6031587448b975f5867c94873ed
pyramid_es/dotdict.py
pyramid_es/dotdict.py
class DotDict(dict): __getattr__ = dict.__getitem__ __setattr__ = dict.__setitem__ __delattr__ = dict.__delitem__ def __init__(self, d={}): for key, value in d.items(): if hasattr(value, 'keys'): value = DotDict(value) if isinstance(value, list): value = [DotDict(el) if hasattr(el, 'keys') else el for el in value] self[key] = value def __repr__(self): return '<DotDict(%s)>' % dict.__repr__(self)
class DotDict(dict): __getattr__ = dict.__getitem__ __setattr__ = dict.__setitem__ __delattr__ = dict.__delitem__ def __init__(self, d={}): for key, value in d.items(): if hasattr(value, 'keys'): value = DotDict(value) if isinstance(value, list): value = [DotDict(el) if hasattr(el, 'keys') else el for el in value] self[key] = value def __repr__(self): return '<%s(%s)>' % (self.__class__.__name__, dict.__repr__(self))
Make DotDict repr() use class name so that it doesn't print misleading results if subclassed
Make DotDict repr() use class name so that it doesn't print misleading results if subclassed
Python
mit
storborg/pyramid_es
--- +++ @@ -14,4 +14,4 @@ self[key] = value def __repr__(self): - return '<DotDict(%s)>' % dict.__repr__(self) + return '<%s(%s)>' % (self.__class__.__name__, dict.__repr__(self))
9a240f0efab9be036fe39f9b2b63cc399e5f8134
registration/admin.py
registration/admin.py
from django.contrib import admin from registration.models import RegistrationProfile class RegistrationAdmin(admin.ModelAdmin): list_display = ('__unicode__', 'activation_key_expired') search_fields = ('user__username', 'user__first_name') admin.site.register(RegistrationProfile, RegistrationAdmin)
from django.contrib import admin from registration.models import RegistrationProfile class RegistrationAdmin(admin.ModelAdmin): list_display = ('__unicode__', 'activation_key_expired') raw_id_fields = ['user'] search_fields = ('user__username', 'user__first_name') admin.site.register(RegistrationProfile, RegistrationAdmin)
Use raw_id_fields for the relation from RegistrationProfile to User, for sites which have huge numbers of users.
Use raw_id_fields for the relation from RegistrationProfile to User, for sites which have huge numbers of users.
Python
bsd-3-clause
dinie/django-registration,Avenza/django-registration,FundedByMe/django-registration,dinie/django-registration,FundedByMe/django-registration
--- +++ @@ -5,6 +5,7 @@ class RegistrationAdmin(admin.ModelAdmin): list_display = ('__unicode__', 'activation_key_expired') + raw_id_fields = ['user'] search_fields = ('user__username', 'user__first_name')
552d216f2cbcb780c08cc351a7456fb97ac3c3dd
app.py
app.py
from flask import Flask, request, render_template import os app = Flask(__name__) app.debug = True # Secret Key setting based on debug setting if app.debug: app.secret_key = "T3st_s3cret_k3y!~$@" else: app.secret_key = os.urandom(30) @app.route("/domain", methods=["GET", "POST"]) def domain(): if request.method == "GET": # search domain pass elif request.method == "POST": # register domain pass @app.route("/") def index(): return render_template("index.html") if __name__ == "__main__": app.run(host="0.0.0.0", threaded=True)
from flask import Flask, request, render_template from route import Route from route.db import init_db, db_session import os app = Flask(__name__) app.debug = True # Secret Key setting based on debug setting if app.debug: app.secret_key = "T3st_s3cret_k3y!~$@" else: init_db() app.secret_key = os.urandom(30) @app.teardown_request def remove_session(exception=None): db_session.remove() @app.route("/domain", methods=["GET", "POST"]) def domain(): ip = request.args.get("ip") domain = request.args.get("domain") if (ip and domain) is None: return "parameter error", 400 r = Route(ip, domain) if request.method == "GET": # search domain r.search() elif request.method == "POST": # register domain user = register.args.get("user") if user is not None: r.register(user) else: return "user not found", 400 del r return "{} - {}".format(ip, domain) @app.route("/") def index(): return render_template("index.html") if __name__ == "__main__": app.run(host="0.0.0.0", threaded=True)
Add domain route and case handling
Add domain route and case handling
Python
apache-2.0
bunseokbot/proxy_register,bunseokbot/proxy_register
--- +++ @@ -1,4 +1,8 @@ from flask import Flask, request, render_template + +from route import Route + +from route.db import init_db, db_session import os @@ -10,18 +14,41 @@ if app.debug: app.secret_key = "T3st_s3cret_k3y!~$@" else: + init_db() app.secret_key = os.urandom(30) + + +@app.teardown_request +def remove_session(exception=None): + db_session.remove() @app.route("/domain", methods=["GET", "POST"]) def domain(): + ip = request.args.get("ip") + domain = request.args.get("domain") + + if (ip and domain) is None: + return "parameter error", 400 + + r = Route(ip, domain) + if request.method == "GET": # search domain - pass + r.search() elif request.method == "POST": # register domain - pass + user = register.args.get("user") + + if user is not None: + r.register(user) + else: + return "user not found", 400 + + del r + + return "{} - {}".format(ip, domain) @app.route("/")
f4500e6422f1c6af8e9ce7d2d79d81e7479f0b7f
Instanssi/admin_programme/forms.py
Instanssi/admin_programme/forms.py
# -*- coding: utf-8 -*- from django import forms from uni_form.helper import FormHelper from uni_form.layout import Submit, Layout, Fieldset, ButtonHolder from Instanssi.ext_programme.models import ProgrammeEvent class ProgrammeEventForm(forms.ModelForm): def __init__(self, *args, **kwargs): super(ProgrammeEventForm, self).__init__(*args, **kwargs) self.helper = FormHelper() self.helper.layout = Layout( Fieldset( u'', 'title', 'description', 'presenters', 'presenters_titles', 'icon_original', 'email', 'home_url', 'twitter_url', 'github_url', 'facebook_url', 'linkedin_url', 'wiki_url', ButtonHolder ( Submit('submit', u'Tallenna') ) ) ) class Meta: model = ProgrammeEvent exclude = ('event',)
# -*- coding: utf-8 -*- from django import forms from uni_form.helper import FormHelper from uni_form.layout import Submit, Layout, Fieldset, ButtonHolder from Instanssi.ext_programme.models import ProgrammeEvent class ProgrammeEventForm(forms.ModelForm): def __init__(self, *args, **kwargs): super(ProgrammeEventForm, self).__init__(*args, **kwargs) self.helper = FormHelper() self.helper.layout = Layout( Fieldset( u'', 'title', 'description', 'start', 'end', 'presenters', 'presenters_titles', 'icon_original', 'email', 'home_url', 'twitter_url', 'github_url', 'facebook_url', 'linkedin_url', 'wiki_url', ButtonHolder ( Submit('submit', u'Tallenna') ) ) ) class Meta: model = ProgrammeEvent exclude = ('event','icon_small',)
Fix form to reflect model change
admin_programme: Fix form to reflect model change
Python
mit
Instanssi/Instanssi.org,Instanssi/Instanssi.org,Instanssi/Instanssi.org,Instanssi/Instanssi.org
--- +++ @@ -14,6 +14,8 @@ u'', 'title', 'description', + 'start', + 'end', 'presenters', 'presenters_titles', 'icon_original', @@ -32,4 +34,4 @@ class Meta: model = ProgrammeEvent - exclude = ('event',) + exclude = ('event','icon_small',)
71cffcb8a8ec7e36dc389a5aa6dc2cc9769a9e97
distutils/tests/test_ccompiler.py
distutils/tests/test_ccompiler.py
import os import sys import platform import textwrap import sysconfig import pytest from distutils import ccompiler def _make_strs(paths): """ Convert paths to strings for legacy compatibility. """ if sys.version_info > (3, 8) and platform.system() != "Windows": return paths return list(map(os.fspath, paths)) @pytest.fixture def c_file(tmp_path): c_file = tmp_path / 'foo.c' gen_headers = ('Python.h',) is_windows = platform.system() == "Windows" plat_headers = ('windows.h',) * is_windows all_headers = gen_headers + plat_headers headers = '\n'.join(f'#include <{header}>\n' for header in all_headers) payload = ( textwrap.dedent( """ #headers void PyInit_foo(void) {} """ ) .lstrip() .replace('#headers', headers) ) c_file.write_text(payload) return c_file def test_set_include_dirs(c_file): """ Extensions should build even if set_include_dirs is invoked. In particular, compiler-specific paths should not be overridden. """ compiler = ccompiler.new_compiler() python = sysconfig.get_paths()['include'] compiler.set_include_dirs([python]) compiler.compile(_make_strs([c_file]))
import os import sys import platform import textwrap import sysconfig import pytest from distutils import ccompiler def _make_strs(paths): """ Convert paths to strings for legacy compatibility. """ if sys.version_info > (3, 8) and platform.system() != "Windows": return paths return list(map(os.fspath, paths)) @pytest.fixture def c_file(tmp_path): c_file = tmp_path / 'foo.c' gen_headers = ('Python.h',) is_windows = platform.system() == "Windows" plat_headers = ('windows.h',) * is_windows all_headers = gen_headers + plat_headers headers = '\n'.join(f'#include <{header}>\n' for header in all_headers) payload = ( textwrap.dedent( """ #headers void PyInit_foo(void) {} """ ) .lstrip() .replace('#headers', headers) ) c_file.write_text(payload) return c_file def test_set_include_dirs(c_file): """ Extensions should build even if set_include_dirs is invoked. In particular, compiler-specific paths should not be overridden. """ compiler = ccompiler.new_compiler() python = sysconfig.get_paths()['include'] compiler.set_include_dirs([python]) compiler.compile(_make_strs([c_file])) # do it again, setting include dirs after any initialization compiler.set_include_dirs([python]) compiler.compile(_make_strs([c_file]))
Extend the test to compile a second time after setting include dirs again.
Extend the test to compile a second time after setting include dirs again.
Python
mit
pypa/setuptools,pypa/setuptools,pypa/setuptools
--- +++ @@ -49,3 +49,7 @@ python = sysconfig.get_paths()['include'] compiler.set_include_dirs([python]) compiler.compile(_make_strs([c_file])) + + # do it again, setting include dirs after any initialization + compiler.set_include_dirs([python]) + compiler.compile(_make_strs([c_file]))
792b9d42167208b39a5e37f2ddf608547238af4f
django_graph_api/graphql/utils.py
django_graph_api/graphql/utils.py
from traceback import format_exc from django.conf import settings class GraphQLError(Exception): def __init__(self, message): super(GraphQLError, self).__init__(message) self.message = message self.location = {'line': 0, 'column': 0} if settings.DEBUG: print(format_exc().split('/n')) def format(self): return {'message': self.message} def __eq__(self, other): return self.__class__ == other.__class__ and self.message == other.message def __hash__(self): return super(GraphQLError, self).__hash__() + self.message.__hash__()
from traceback import format_exc from django.conf import settings class GraphQLError(Exception): def __init__(self, message): super(GraphQLError, self).__init__(message) self.message = message if settings.DEBUG: self.traceback = format_exc().split('\n') def format(self): return {'message': self.message} def __eq__(self, other): return self.__class__ == other.__class__ and self.message == other.message def __hash__(self): return super(GraphQLError, self).__hash__() + self.message.__hash__()
Undo error changes for now
Undo error changes for now
Python
mit
melinath/django-graph-api,melinath/django-graph-api
--- +++ @@ -7,9 +7,8 @@ def __init__(self, message): super(GraphQLError, self).__init__(message) self.message = message - self.location = {'line': 0, 'column': 0} if settings.DEBUG: - print(format_exc().split('/n')) + self.traceback = format_exc().split('\n') def format(self): return {'message': self.message}
08489ea2c1596a067b482878ff4450db43c08612
conf.py
conf.py
# -*- coding: utf-8 -*- # # on_rtd is whether we are on readthedocs.org import os on_rtd = os.environ.get('READTHEDOCS', None) == 'True' if not on_rtd: # only import and set the theme if we're building docs locally import sphinx_rtd_theme html_theme = 'sphinx_rtd_theme' html_theme_path = [sphinx_rtd_theme.get_html_theme_path()] # otherwise, readthedocs.org uses their theme by default, so no need to specify it project = 'FIWARE-Stream-Oriented-GE'
# -*- coding: utf-8 -*- # # on_rtd is whether we are on readthedocs.org import os on_rtd = os.environ.get('READTHEDOCS', None) == 'True' if not on_rtd: # only import and set the theme if we're building docs locally import sphinx_rtd_theme html_theme = 'sphinx_rtd_theme' html_theme_path = [sphinx_rtd_theme.get_html_theme_path()] # otherwise, readthedocs.org uses their theme by default, so no need to specify it project = 'FIWARE-Stream-Oriented-GE' html_theme_options = { 'cssfiles': ['https://fiware.org/style/fiware_readthedocs.css'] }
Add custom CSS style to FIWARE doc
Add custom CSS style to FIWARE doc Change-Id: I74293d488e0cd762ad023b94879ee618a4016110
Python
apache-2.0
Kurento/doc-kurento,SanMi86/doc-kurento,SanMi86/doc-kurento,SanMi86/doc-kurento,Kurento/doc-kurento,Kurento/doc-kurento,SanMi86/doc-kurento
--- +++ @@ -12,3 +12,6 @@ # otherwise, readthedocs.org uses their theme by default, so no need to specify it project = 'FIWARE-Stream-Oriented-GE' +html_theme_options = { + 'cssfiles': ['https://fiware.org/style/fiware_readthedocs.css'] +}
0fbc02b40f4414d96686d879aa9f7611e8fbb85d
singlet/config.py
singlet/config.py
# vim: fdm=indent # author: Fabio Zanini # date: 02/08/17 # content: Support module for filenames related to the Google Sheet APIs. # Modules import os import yaml # Globals config_filename = os.getenv( 'SINGLET_CONFIG_FILENAME', os.getenv('HOME') + '/.singlet/config.yml') with open(config_filename) as stream: config = yaml.load(stream) # Warnings that should be seen only once config['_once_warnings'] = [] config.reset_once_warings = lambda: config['_once_warnings'] = []
# vim: fdm=indent # author: Fabio Zanini # date: 02/08/17 # content: Support module for filenames related to the Google Sheet APIs. # Modules import os import yaml # Globals config_filename = os.getenv( 'SINGLET_CONFIG_FILENAME', os.getenv('HOME') + '/.singlet/config.yml') with open(config_filename) as stream: config = yaml.load(stream) # Warnings that should be seen only once config['_once_warnings'] = []
Remove function to reset _once_warnings (messy)
Remove function to reset _once_warnings (messy)
Python
mit
iosonofabio/singlet,iosonofabio/singlet
--- +++ @@ -16,4 +16,3 @@ # Warnings that should be seen only once config['_once_warnings'] = [] -config.reset_once_warings = lambda: config['_once_warnings'] = []
ed7d0c5f8b64185f9fc612b44e4182b12a0fa62e
yunity/users/factories.py
yunity/users/factories.py
from django.contrib.auth import get_user_model from factory import DjangoModelFactory, CREATE_STRATEGY, LazyAttribute, PostGeneration, SubFactory from yunity.walls.factories import Wall from yunity.utils.tests.fake import faker class User(DjangoModelFactory): class Meta: model = get_user_model() strategy = CREATE_STRATEGY is_active = True is_staff = False display_name = LazyAttribute(lambda _: faker.name()) first_name = LazyAttribute(lambda _: faker.name()) last_name = LazyAttribute(lambda _: faker.name()) email = LazyAttribute(lambda _: faker.email()) password = PostGeneration(lambda obj, *args, **kwargs: obj.set_password(obj.display_name)) wall = SubFactory(Wall)
from django.contrib.auth import get_user_model from factory import DjangoModelFactory, CREATE_STRATEGY, LazyAttribute, PostGeneration, SubFactory from yunity.walls.factories import Wall from yunity.utils.tests.fake import faker class User(DjangoModelFactory): class Meta: model = get_user_model() strategy = CREATE_STRATEGY is_active = True is_staff = False display_name = LazyAttribute(lambda _: faker.name()) first_name = LazyAttribute(lambda _: faker.name()) last_name = LazyAttribute(lambda _: faker.name()) email = LazyAttribute(lambda _: faker.email()) # Use display_name as password, as it is readable password = PostGeneration(lambda obj, *args, **kwargs: obj.set_password(obj.display_name)) wall = SubFactory(Wall)
Comment about display_name == password
Comment about display_name == password
Python
agpl-3.0
yunity/foodsaving-backend,yunity/yunity-core,yunity/foodsaving-backend,yunity/yunity-core,yunity/foodsaving-backend
--- +++ @@ -16,6 +16,8 @@ first_name = LazyAttribute(lambda _: faker.name()) last_name = LazyAttribute(lambda _: faker.name()) email = LazyAttribute(lambda _: faker.email()) + + # Use display_name as password, as it is readable password = PostGeneration(lambda obj, *args, **kwargs: obj.set_password(obj.display_name)) wall = SubFactory(Wall)
aed9b3066f9d796e5c89e38d833c87e130a421c3
auth0/v2/blacklists.py
auth0/v2/blacklists.py
from .rest import RestClient class Blacklists(object): def __init__(self, domain, jwt_token): url = 'https://%s/api/v2/blacklists/tokens' % domain self.client = RestClient(endpoint=url, jwt=jwt_token) def get(self, aud=None): params = { 'aud': aud } return self.client.get(params=params) def create(self, jti, aud=''): return self.client.post(data={'jti': jti, 'aud': aud})
from .rest import RestClient class Blacklists(object): def __init__(self, domain, jwt_token): self.url = 'https://%s/api/v2/blacklists/tokens' % domain self.client = RestClient(jwt=jwt_token) def get(self, aud=None): params = { 'aud': aud } return self.client.get(self.url, params=params) def create(self, jti, aud=''): return self.client.post(self.url, data={'jti': jti, 'aud': aud})
Fix Blacklists usage of RestClient
Fix Blacklists usage of RestClient
Python
mit
auth0/auth0-python,auth0/auth0-python
--- +++ @@ -3,16 +3,15 @@ class Blacklists(object): def __init__(self, domain, jwt_token): - url = 'https://%s/api/v2/blacklists/tokens' % domain - - self.client = RestClient(endpoint=url, jwt=jwt_token) + self.url = 'https://%s/api/v2/blacklists/tokens' % domain + self.client = RestClient(jwt=jwt_token) def get(self, aud=None): params = { 'aud': aud } - return self.client.get(params=params) + return self.client.get(self.url, params=params) def create(self, jti, aud=''): - return self.client.post(data={'jti': jti, 'aud': aud}) + return self.client.post(self.url, data={'jti': jti, 'aud': aud})
e49163ceecc5da949fe01281a87b56be513784d5
abbr/languages/pt_br/dictionary.py
abbr/languages/pt_br/dictionary.py
# Copyright 2016 Adler Brediks Medrado # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. word_list = { 'Professor': 'Prof.', 'Senhor': 'Sr.', 'Doutor': 'Dr.', }
# Copyright 2016 Adler Brediks Medrado # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. word_list = { 'Professor': 'Prof.', 'Senhor': 'Sr.', 'Doutor': 'Dr.', 'Avenida': 'Av.', 'Travessa': 'Tv.', 'Número': 'Nº', 'Primeiro': '1º', 'Primeira': '1ª', 'Segundo': '2º', 'Segunda': '2ª', }
Add new words to wordlist
Add new words to wordlist
Python
apache-2.0
adlermedrado/abbr
--- +++ @@ -16,4 +16,11 @@ 'Professor': 'Prof.', 'Senhor': 'Sr.', 'Doutor': 'Dr.', + 'Avenida': 'Av.', + 'Travessa': 'Tv.', + 'Número': 'Nº', + 'Primeiro': '1º', + 'Primeira': '1ª', + 'Segundo': '2º', + 'Segunda': '2ª', }
92e5ff34737feef0d196e25b97dbc817b502a59d
demo.py
demo.py
from FbFeed import NewsFeed username = raw_input('Enter your email id registered with facebook : ') password = raw_input('Enter your Password : ') print('Creating new session on Firefox..') fb = NewsFeed(username,password) print('Logging into your facebook account') fb.login() #Add people to group print('Add people to Feed Group') count = int(raw_input('How many people would you like to add ?: ')) for i in range(count): name = raw_input() fb.add(name) print('Leave me running on your system , I will notify you whenever these people perform any public activity') fb.read()
from FbFeed import NewsFeed import getpass username = raw_input('Enter your email id registered with facebook : ') password = getpass.getpass(prompt='Enter your Password : ',stream=None) print('Creating new session on Firefox..') fb = NewsFeed(username,password) print('Logging into your facebook account') fb.login() #Add people to group print('Add people to Feed Group') count = int(raw_input('How many people would you like to add ?: ')) for i in range(count): name = raw_input() fb.add(name) print('Leave me running on your system , I will notify you whenever these people perform any public activity') fb.read()
Hide password in terminal input
Hide password in terminal input
Python
mit
ashishpahwa7/Fb-Feedirator
--- +++ @@ -1,8 +1,8 @@ from FbFeed import NewsFeed - +import getpass username = raw_input('Enter your email id registered with facebook : ') -password = raw_input('Enter your Password : ') +password = getpass.getpass(prompt='Enter your Password : ',stream=None) print('Creating new session on Firefox..') fb = NewsFeed(username,password) print('Logging into your facebook account')
deeaed14e40b9deca39c46ec7879f775606898c0
Instanssi/dblog/handlers.py
Instanssi/dblog/handlers.py
# -*- coding: utf-8 -*- from logging import Handler from datetime import datetime class DBLogHandler(Handler, object): def __init__(self): super(DBLogHandler, self).__init__() def emit(self, record): from models import DBLogEntry as _LogEntry entry = _LogEntry() entry.level = record.levelname entry.message = self.format(record) if record.event: entry.event = record.event if record.user: entry.user = record.user entry.save()
# -*- coding: utf-8 -*- from logging import Handler from datetime import datetime class DBLogHandler(Handler, object): def __init__(self): super(DBLogHandler, self).__init__() def emit(self, record): from models import DBLogEntry as _LogEntry entry = _LogEntry() entry.level = record.levelname entry.message = self.format(record) entry.module = record.name try: entry.event = record.event except: pass try: entry.user = record.user except: pass entry.save()
Handle optional field saving with exceptions, save module name.
dblog: Handle optional field saving with exceptions, save module name.
Python
mit
Instanssi/Instanssi.org,Instanssi/Instanssi.org,Instanssi/Instanssi.org,Instanssi/Instanssi.org
--- +++ @@ -13,9 +13,14 @@ entry = _LogEntry() entry.level = record.levelname entry.message = self.format(record) - if record.event: + entry.module = record.name + try: entry.event = record.event - if record.user: + except: + pass + try: entry.user = record.user + except: + pass entry.save()
407a032acb307e5f936437aec4975ef69133d0c5
DisplayAdapter/testing/test_display_adapter/test_display_driver/test_display_drivers.py
DisplayAdapter/testing/test_display_adapter/test_display_driver/test_display_drivers.py
""" This module contains the testing framework for the display driver functionality, and is responsible for testing whether the pi can correctly and sufficiently connect to the display. """ from mock import patch from display_adapter.display_driver.display_drivers import DisplayDriver class TestDisplayDriver(object): """ This class tests the functionality of the DisplayDriver class, ensuring that it can correctly connect the raspberry pi to the display. """ @patch("serial.Serial") def test_init(self, dc_mock): """ This method tests initialisation of the display driver, ensuring it has correctly set up the database helper and the display controller. """ dd = DisplayDriver() assert dd assert hasattr(dd, "_db_helper") assert hasattr(dd, "_display_controller")
""" This module contains the testing framework for the display driver functionality, and is responsible for testing whether the pi can correctly and sufficiently connect to the display. """ from mock import patch from datetime import datetime from display_adapter.display_driver.display_drivers import minutify, DisplayDriver class TestDisplayDriver(object): """ This class tests the functionality of the DisplayDriver class, ensuring that it can correctly connect the raspberry pi to the display. """ @patch("serial.Serial") def test_init(self, dc_mock): """ This method tests initialisation of the display driver, ensuring it has correctly set up the database helper and the display controller. """ dd = DisplayDriver() assert dd assert hasattr(dd, "_db_helper") assert hasattr(dd, "_display_controller") def test_minutify(): """ This function tests the functionality of the minutify function linked to the Display Driver. The expected result of this test is for a datetime object to be made accurate to the minute. """ dt = datetime.now().replace(second=30, microsecond=40000) accurate_dt = minutify(dt) # Assert the datetime object has been minutified correctly (seconds and microseconds == 0) assert accurate_dt.second == 0 and accurate_dt.microsecond == 0
Test functionality has been added. Paired by Richard and Michael.
Test functionality has been added. Paired by Richard and Michael. The functionality for the minutify function has now been tested; and the tests work (Support 231)
Python
mit
CO600GOL/Game_of_life,CO600GOL/Game_of_life,CO600GOL/Game_of_life
--- +++ @@ -4,7 +4,8 @@ """ from mock import patch -from display_adapter.display_driver.display_drivers import DisplayDriver +from datetime import datetime +from display_adapter.display_driver.display_drivers import minutify, DisplayDriver class TestDisplayDriver(object): """ @@ -23,3 +24,15 @@ assert dd assert hasattr(dd, "_db_helper") assert hasattr(dd, "_display_controller") + + +def test_minutify(): + """ + This function tests the functionality of the minutify function linked to the Display Driver. The expected result + of this test is for a datetime object to be made accurate to the minute. + """ + + dt = datetime.now().replace(second=30, microsecond=40000) + accurate_dt = minutify(dt) + # Assert the datetime object has been minutified correctly (seconds and microseconds == 0) + assert accurate_dt.second == 0 and accurate_dt.microsecond == 0
3e4094d838df26afadfa7074a16c6a21f06ba6b8
bamp/logs.py
bamp/logs.py
import logging class ExceptionFilter(logging.Filter): def __init__(self, debug=None): self.debug = debug def filter(self, record): if self.debug: return True # clear exceptions when not in debug if record.levelname == 'ERROR' and record.exc_info: record.exc_info = None return True LOGGING = { 'version': 1, 'filters': { 'exc_filter': { '()': ExceptionFilter, 'debug': False } }, 'handlers': { 'console': { 'class': 'logging.StreamHandler', 'filters': ['exc_filter'] } }, 'loggers': { 'bamp': { 'level': 'DEBUG', 'handlers': ['console'] }, } }
import logging class DebugFilter(logging.Filter): def __init__(self, debug=None): self.debug = debug def filter(self, record): if self.debug: return True return False LOGGING = { 'version': 1, 'filters': { 'exc_filter': { '()': DebugFilter, 'debug': False } }, 'handlers': { 'console': { 'class': 'logging.StreamHandler', 'filters': ['exc_filter'] } }, 'loggers': { 'bamp': { 'level': 'DEBUG', 'handlers': ['console'] }, } }
Disable printing error messages not in debug
Disable printing error messages not in debug
Python
mit
inirudebwoy/bamp
--- +++ @@ -1,24 +1,20 @@ import logging -class ExceptionFilter(logging.Filter): +class DebugFilter(logging.Filter): def __init__(self, debug=None): self.debug = debug def filter(self, record): if self.debug: return True - - # clear exceptions when not in debug - if record.levelname == 'ERROR' and record.exc_info: - record.exc_info = None - return True + return False LOGGING = { 'version': 1, 'filters': { 'exc_filter': { - '()': ExceptionFilter, + '()': DebugFilter, 'debug': False } },
61d7c9e99398874745d11720cd8d985bdc3d7514
demoapp/views.py
demoapp/views.py
from demoapp.forms import DemoLoginForm from django.shortcuts import render_to_response from django.shortcuts import redirect from demoapp import app_settings from demoapp.utils import get_salt def login_view(request): if request.method == 'POST': form = DemoLoginForm(request.POST) if form.is_valid(): response = redirect('/') response.set_signed_cookie(app_settings.COOKIE_NAME, 'demo access granted', salt=get_salt(request)) return response else: form = DemoLoginForm() return render_to_response('demoapp/login.html', {'form': form})
from demoapp.forms import DemoLoginForm from django.shortcuts import render from django.shortcuts import redirect from demoapp import app_settings from demoapp.utils import get_salt def login_view(request): if request.method == 'POST': form = DemoLoginForm(request.POST) if form.is_valid(): response = redirect('/') response.set_signed_cookie(app_settings.COOKIE_NAME, 'demo access granted', salt=get_salt(request)) return response else: form = DemoLoginForm() return render(request, 'demoapp/login.html', {'form': form})
Add context to login view
Add context to login view
Python
unlicense
dboczek/django-demo,dboczek/django-demo
--- +++ @@ -1,5 +1,5 @@ from demoapp.forms import DemoLoginForm -from django.shortcuts import render_to_response +from django.shortcuts import render from django.shortcuts import redirect from demoapp import app_settings from demoapp.utils import get_salt @@ -14,4 +14,4 @@ return response else: form = DemoLoginForm() - return render_to_response('demoapp/login.html', {'form': form}) + return render(request, 'demoapp/login.html', {'form': form})
fc3408e0d8336ca2324b272dbb4aa0e69914a27c
build_chrome_webapp.py
build_chrome_webapp.py
import os.path from shutil import copyfile try: from jinja2 import Template except: print "Could not import Jinja2, run 'easy_install Jinja2'" exit() output_dir = os.path.join('./', 'chrome_webstore') if not os.path.exists(output_dir): os.makedirs(output_dir) def add_background_script(): copyfile('chrome_webstore_background.js', os.path.join(output_dir, 'background.js')) def add_manifest(): copyfile('chrome_webstore_manifest.json', os.path.join(output_dir, 'manifest.json')) def render_main_template(): f = open('./html/index.html') template = Template(f.read().decode('utf-8')) f.close() html = template.render(og_tag='', url='', ON_PRODUCTION=True, ON_DEV=False, USE_PRODUCTION_JAVASCRIPT=True) f = open(os.path.join(output_dir, 'index.html'), 'w') f.write(html.encode('utf-8')) f.close() print "Template rendered" add_manifest() add_background_script() render_main_template()
import os.path from shutil import copyfile from shutil import copytree from shutil import rmtree try: from jinja2 import Template except: print "Could not import Jinja2, run 'easy_install Jinja2'" exit() output_dir = os.path.join('./', 'chrome_webstore') if os.path.exists(output_dir): rmtree(output_dir) os.makedirs(output_dir) def add_background_script(): copyfile('chrome_webstore_background.js', os.path.join(output_dir, 'background.js')) def copy_static_dirs(): copytree('images', os.path.join(output_dir, 'images')) copytree('styles', os.path.join(output_dir, 'styles')) copytree('scripts', os.path.join(output_dir, 'scripts')) def add_manifest(): copyfile('chrome_webstore_manifest.json', os.path.join(output_dir, 'manifest.json')) def render_main_template(): f = open('./html/index.html') template = Template(f.read().decode('utf-8')) f.close() html = template.render(og_tag='', url='', ON_PRODUCTION=True, ON_DEV=False, USE_PRODUCTION_JAVASCRIPT=True) f = open(os.path.join(output_dir, 'index.html'), 'w') f.write(html.encode('utf-8')) f.close() print "Template rendered" add_manifest() add_background_script() render_main_template() copy_static_dirs()
Copy static dirs as well
Copy static dirs as well
Python
mit
youtify/youtify,youtify/youtify,youtify/youtify
--- +++ @@ -1,5 +1,7 @@ import os.path from shutil import copyfile +from shutil import copytree +from shutil import rmtree try: from jinja2 import Template except: @@ -8,11 +10,18 @@ output_dir = os.path.join('./', 'chrome_webstore') -if not os.path.exists(output_dir): - os.makedirs(output_dir) +if os.path.exists(output_dir): + rmtree(output_dir) + +os.makedirs(output_dir) def add_background_script(): copyfile('chrome_webstore_background.js', os.path.join(output_dir, 'background.js')) + +def copy_static_dirs(): + copytree('images', os.path.join(output_dir, 'images')) + copytree('styles', os.path.join(output_dir, 'styles')) + copytree('scripts', os.path.join(output_dir, 'scripts')) def add_manifest(): copyfile('chrome_webstore_manifest.json', os.path.join(output_dir, 'manifest.json')) @@ -32,3 +41,4 @@ add_manifest() add_background_script() render_main_template() +copy_static_dirs()
5c1615d9e3e16c1f48be41733997d8c39355bcb4
django_distributed_queue/utils.py
django_distributed_queue/utils.py
class LazyModel(object): """ It's a helper class that is used in case you have model_class and object primary key. You might need to use only object id. If you try to access other fields of the model class then we will query the database to get that object and provide you with any field and method transparently proxying them. """ def __init__(self, model_class, pk): self._model_class = model_class self.pk = pk self._instance = None def __getattribute__(self, attr_name): # Hiding traces of decoration. if attr_name in ('__init__', '__getattribute__', '_model_class', 'pk', '_instance'): # Stopping recursion. return object.__getattribute__(self, attr_name) # All other attr_names, including auto-defined by system in self, are # searched in decorated self.instance, e.g.: __module__, __class__, etc. if self._instance is None: self._instance = self._model_class.objects.get(pk=self.pk) # Raises correct AttributeError if name is not found in decorated self.func. return getattr(self.instance, attr_name)
class LazyModel(object): """ It's a helper class that is used in case you have model_class and object primary key. You might need to use only object id. If you try to access other fields of the model class then we will query the database to get that object and provide you with any field and method transparently proxying them. """ def __init__(self, model_class, pk): self._model_class = model_class self.pk = pk self._instance = None def __getattribute__(self, attr_name): # Hiding traces of decoration. if attr_name in ('__init__', '__getattribute__', '_model_class', 'pk', '_instance'): # Stopping recursion. return object.__getattribute__(self, attr_name) # All other attr_names, including auto-defined by system in self, are # searched in decorated self.instance, e.g.: __module__, __class__, etc. if self._instance is None: self._instance = self._model_class.objects.get(pk=self.pk) # Raises correct AttributeError if name is not found in decorated self.func. return getattr(self._instance, attr_name)
Fix for LazyModel infinite recursion
Fix for LazyModel infinite recursion
Python
mit
ProstoKSI/django-distributed-queue
--- +++ @@ -23,4 +23,4 @@ if self._instance is None: self._instance = self._model_class.objects.get(pk=self.pk) # Raises correct AttributeError if name is not found in decorated self.func. - return getattr(self.instance, attr_name) + return getattr(self._instance, attr_name)
df1617a7518f66d87470f948e057e4d7d7d8f026
driller/tasks.py
driller/tasks.py
import redis from celery import Celery from .driller import Driller app = Celery('tasks', broker='amqp://guest@localhost//', backend='redis://localhost') redis_pool = redis.ConnectionPool(host='localhost', port=6379, db=1) @app.task def drill(binary, input, fuzz_bitmap, qemu_dir): redis_inst = redis.Redis(connection_pool=redis_pool) driller = Driller(binary, input, fuzz_bitmap, qemu_dir, redis=redis_inst) return driller.drill()
import redis from celery import Celery from .driller import Driller import config backend_url = "redis://%s:%d" % (config.REDIS_HOST, config.REDIS_PORT) app = Celery('tasks', broker=config.BROKER_URL, backend=backend_url) redis_pool = redis.ConnectionPool(host=config.REDIS_HOST, port=config.REDIS_PORT, db=config.REDIS_DB) @app.task def drill(binary, input, fuzz_bitmap, qemu_dir): redis_inst = redis.Redis(connection_pool=redis_pool) driller = Driller(binary, input, fuzz_bitmap, qemu_dir, redis=redis_inst) return driller.drill()
Connect to Celery using config options
Connect to Celery using config options
Python
bsd-2-clause
shellphish/driller
--- +++ @@ -2,9 +2,11 @@ from celery import Celery from .driller import Driller +import config -app = Celery('tasks', broker='amqp://guest@localhost//', backend='redis://localhost') -redis_pool = redis.ConnectionPool(host='localhost', port=6379, db=1) +backend_url = "redis://%s:%d" % (config.REDIS_HOST, config.REDIS_PORT) +app = Celery('tasks', broker=config.BROKER_URL, backend=backend_url) +redis_pool = redis.ConnectionPool(host=config.REDIS_HOST, port=config.REDIS_PORT, db=config.REDIS_DB) @app.task def drill(binary, input, fuzz_bitmap, qemu_dir):
a85d148eb00f83052a97d66da8ff9dd79b40f172
.ycm_extra_conf.py
.ycm_extra_conf.py
import os def FlagsForFile(filename, **kwargs): flags = ['-std=c++14', '-I/usr/local/include'] proj_root = os.path.dirname(os.path.abspath(__file__)) libcanon_include = ''.join(['-I', proj_root, '/deps/libcanon/include']) proj_include = ''.join(['-I', proj_root, '/drudge']) flags.extend([libcanon_include, proj_include]) return {'flags': flags}
import os import subprocess def FlagsForFile(filename, **kwargs): flags = ['-std=c++14', '-I/usr/local/include'] proj_root = os.path.dirname(os.path.abspath(__file__)) libcanon_include = ''.join(['-I', proj_root, '/deps/libcanon/include']) python_include = subprocess.run( ["pkg-config", '--cflags', 'python3'], stdout=subprocess.PIPE ).stdout.decode("utf-8") proj_include = ''.join(['-I', proj_root, '/drudge']) flags.extend([libcanon_include, proj_include]) return {'flags': flags}
Add Python inclusion path to YCM config
Add Python inclusion path to YCM config In the script, the path is read from the result from pkg-config. So it should work in most places.
Python
mit
tschijnmo/drudge,tschijnmo/drudge,tschijnmo/drudge
--- +++ @@ -1,4 +1,5 @@ import os +import subprocess def FlagsForFile(filename, **kwargs): @@ -6,6 +7,9 @@ proj_root = os.path.dirname(os.path.abspath(__file__)) libcanon_include = ''.join(['-I', proj_root, '/deps/libcanon/include']) + python_include = subprocess.run( + ["pkg-config", '--cflags', 'python3'], stdout=subprocess.PIPE + ).stdout.decode("utf-8") proj_include = ''.join(['-I', proj_root, '/drudge']) flags.extend([libcanon_include, proj_include])
238da6f5cb5409409f54980f4ce018fda897a766
API/chat/models.py
API/chat/models.py
from django.db import models class Channel(models.Model): def __str__(self): return self.name name = models.CharField(max_length=20, unique=True) class Message(models.Model): def __str__(self): return self.text def to_dict(self): serializable_fields = ('text', 'datetime_start', 'datetime_sent', 'username') return {key: getattr(self, key) for key in serializable_fields} text = models.TextField(max_length=2000) datetime_start = models.DateTimeField(default=None) datetime_sent = models.DateTimeField(default=None, null=True) typing = models.BooleanField(default=False) username = models.CharField(max_length=20) channel = models.ForeignKey(Channel)
from django.db import models class Channel(models.Model): def __str__(self): return self.name name = models.CharField(max_length=20, unique=True) class Message(models.Model): def __str__(self): return self.text def to_dict(self): serializable_fields = ('text', 'datetime_start', 'datetime_sent', 'username') return {key: getattr(self, key) for key in serializable_fields} TEXT = 'text' IMAGE = 'image' MESSAGE_TYPE = ( (TEXT, 'text'), (IMAGE, 'image'), ) text = models.TextField(max_length=2000) datetime_start = models.DateTimeField(default=None) datetime_sent = models.DateTimeField(default=None, null=True) typing = models.BooleanField(default=False) username = models.CharField(max_length=20) channel = models.ForeignKey(Channel) message_type = models.CharField(max_length=10, choices=MESSAGE_TYPE, default=TEXT)
Add message_type field into message model
Add message_type field into message model
Python
mit
dionyziz/ting,gtklocker/ting,dionyziz/ting,gtklocker/ting,gtklocker/ting,dionyziz/ting,mbalamat/ting,mbalamat/ting,gtklocker/ting,dionyziz/ting,mbalamat/ting,mbalamat/ting
--- +++ @@ -16,9 +16,20 @@ serializable_fields = ('text', 'datetime_start', 'datetime_sent', 'username') return {key: getattr(self, key) for key in serializable_fields} + TEXT = 'text' + IMAGE = 'image' + + MESSAGE_TYPE = ( + (TEXT, 'text'), + (IMAGE, 'image'), + ) + text = models.TextField(max_length=2000) datetime_start = models.DateTimeField(default=None) datetime_sent = models.DateTimeField(default=None, null=True) typing = models.BooleanField(default=False) username = models.CharField(max_length=20) channel = models.ForeignKey(Channel) + message_type = models.CharField(max_length=10, + choices=MESSAGE_TYPE, + default=TEXT)
7d898ec04733d25c1df33c8faf151f2b42a69ec9
base/components/people/constants.py
base/components/people/constants.py
from model_utils import Choices from ohashi.constants import OTHER BLOOD_TYPE = Choices('A', 'B', 'O', 'AB') CLASSIFICATIONS = Choices( (1, 'major', 'Major Unit'), (2, 'minor', 'Minor Unit'), (4, 'temporary', 'Temporary Unit'), (5, 'subunit', 'Sub-Unit'), (7, 'supergroup', 'Supergroup'), ('Special Units', [ (3, 'shuffle', 'Shuffle Unit'), (6, 'revival', 'Revival Unit'), (8, 'satoyama', 'Satoyama Unit'), ]), (OTHER, 'other', 'Other') ) PHOTO_SOURCES = Choices( (1, 'promotional', 'Promotional Photo'), (2, 'blog', 'Blog Photo'), (OTHER, 'other', 'Other') ) SCOPE = Choices( (1, 'hp', 'Hello! Project'), (2, 'ufa', 'Up Front Agency'), (OTHER, 'other', 'Other') ) STATUS = Choices( (1, 'active', 'Active'), (2, 'former', 'Former'), (OTHER, 'other', 'Other') )
from model_utils import Choices from ohashi.constants import OTHER BLOOD_TYPE = Choices('A', 'B', 'O', 'AB') CLASSIFICATIONS = Choices( (1, 'major', 'Major Unit'), (2, 'minor', 'Minor Unit'), (4, 'temporary', 'Temporary Unit'), (5, 'subunit', 'Sub-Unit'), (7, 'supergroup', 'Supergroup'), ('Special Units', [ (3, 'shuffle', 'Shuffle Unit'), (6, 'revival', 'Revival Unit'), (8, 'satoyama', 'Satoyama Unit'), (9, 'satoumi', 'Satoumi Unit'), ]), (OTHER, 'other', 'Other') ) PHOTO_SOURCES = Choices( (1, 'promotional', 'Promotional Photo'), (2, 'blog', 'Blog Photo'), (OTHER, 'other', 'Other') ) SCOPE = Choices( (1, 'hp', 'Hello! Project'), (2, 'ufa', 'Up Front Agency'), (OTHER, 'other', 'Other') ) STATUS = Choices( (1, 'active', 'Active'), (2, 'former', 'Former'), (OTHER, 'other', 'Other') )
Add Satoumi as a classification.
Add Satoumi as a classification.
Python
apache-2.0
hello-base/web,hello-base/web,hello-base/web,hello-base/web
--- +++ @@ -15,6 +15,7 @@ (3, 'shuffle', 'Shuffle Unit'), (6, 'revival', 'Revival Unit'), (8, 'satoyama', 'Satoyama Unit'), + (9, 'satoumi', 'Satoumi Unit'), ]), (OTHER, 'other', 'Other') )
f828ac9ee5082a9a0b5e215c4c814e7f35db11b6
planetstack/core/models/__init__.py
planetstack/core/models/__init__.py
from .plcorebase import PlCoreBase from .planetstack import PlanetStack from .project import Project from .singletonmodel import SingletonModel from .service import Service from .service import ServiceAttribute from .tag import Tag from .role import Role from .site import Site,Deployment, DeploymentRole, DeploymentPrivilege, SiteDeployments from .dashboard import DashboardView from .user import User, UserDashboardView from .serviceclass import ServiceClass from .slice import Slice, SliceDeployments from .site import SitePrivilege, SiteDeployments from .userdeployments import UserDeployments from .image import Image, ImageDeployments from .node import Node from .serviceresource import ServiceResource from .slice import SliceRole from .slice import SlicePrivilege from .site import SiteRole from .site import SitePrivilege from .planetstack import PlanetStackRole from .planetstack import PlanetStackPrivilege from .slicetag import SliceTag from .flavor import Flavor from .sliver import Sliver from .reservation import ReservedResource from .reservation import Reservation from .network import Network, NetworkParameterType, NetworkParameter, NetworkSliver, NetworkTemplate, Router, NetworkSlice, NetworkDeployments from .billing import Account, Invoice, Charge, UsableObject, Payment
from .plcorebase import PlCoreBase from .planetstack import PlanetStack from .project import Project from .singletonmodel import SingletonModel from .service import Service from .service import ServiceAttribute from .tag import Tag from .role import Role from .site import Site,Deployment, DeploymentRole, DeploymentPrivilege, SiteDeployments from .dashboard import DashboardView from .user import User, UserDashboardView from .serviceclass import ServiceClass from .slice import Slice, SliceDeployments from .site import SitePrivilege, SiteDeployments from .userdeployments import UserDeployments from .image import Image, ImageDeployments from .node import Node from .serviceresource import ServiceResource from .slice import SliceRole from .slice import SlicePrivilege from .credential import UserCredential,SiteCredential,SliceCredential from .site import SiteRole from .site import SitePrivilege from .planetstack import PlanetStackRole from .planetstack import PlanetStackPrivilege from .slicetag import SliceTag from .flavor import Flavor from .sliver import Sliver from .reservation import ReservedResource from .reservation import Reservation from .network import Network, NetworkParameterType, NetworkParameter, NetworkSliver, NetworkTemplate, Router, NetworkSlice, NetworkDeployments from .billing import Account, Invoice, Charge, UsableObject, Payment
Add credentials module to core list
Add credentials module to core list
Python
apache-2.0
wathsalav/xos,wathsalav/xos,wathsalav/xos,wathsalav/xos
--- +++ @@ -18,6 +18,7 @@ from .serviceresource import ServiceResource from .slice import SliceRole from .slice import SlicePrivilege +from .credential import UserCredential,SiteCredential,SliceCredential from .site import SiteRole from .site import SitePrivilege from .planetstack import PlanetStackRole
644660b6c41f029f271a0b8866387f358f8fdf54
frappe/patches/v4_0/enable_scheduler_in_system_settings.py
frappe/patches/v4_0/enable_scheduler_in_system_settings.py
# Copyright (c) 2013, Web Notes Technologies Pvt. Ltd. and Contributors # MIT License. See license.txt from __future__ import unicode_literals import frappe from frappe.utils.scheduler import disable_scheduler, enable_scheduler def execute(): frappe.reload_doc("core", "doctype", "system_settings") if frappe.db.get_global("disable_scheduler"): disable_scheduler() else: enable_scheduler()
# Copyright (c) 2013, Web Notes Technologies Pvt. Ltd. and Contributors # MIT License. See license.txt from __future__ import unicode_literals import frappe from frappe.utils.scheduler import disable_scheduler, enable_scheduler from frappe.utils import cint def execute(): frappe.reload_doc("core", "doctype", "system_settings") if cint(frappe.db.get_global("disable_scheduler")): disable_scheduler() else: enable_scheduler()
Fix in enable scheduler patch
Fix in enable scheduler patch
Python
mit
BhupeshGupta/frappe,letzerp/framework,saurabh6790/frappe,rmehta/frappe,elba7r/builder,suyashphadtare/sajil-frappe,rohitw1991/frappe,saguas/frappe,indictranstech/tele-frappe,nerevu/frappe,indictranstech/omnitech-frappe,vCentre/vFRP-6233,gangadharkadam/saloon_frappe,aboganas/frappe,indictranstech/phr-frappe,gangadharkadam/letzfrappe,hatwar/buyback-frappe,mbauskar/tele-frappe,erpletzerp/letzerpcore,Amber-Creative/amber-frappe,indictranstech/ebuy-now-frappe,elba7r/frameworking,vjFaLk/frappe,shitolepriya/test-frappe,mbauskar/Das_frappe,frappe/frappe,gangadharkadam/v4_frappe,gangadhar-kadam/helpdesk-frappe,vCentre/vFRP-6233,RicardoJohann/frappe,rohitw1991/frappe,ESS-LLP/frappe,nerevu/frappe,gangadharkadam/stfrappe,sbktechnology/trufil-frappe,rohitwaghchaure/frappe_smart,aboganas/frappe,StrellaGroup/frappe,manassolanki/frappe,drukhil/frappe,sbktechnology/sap_frappe,rmehta/frappe,indictranstech/Das_frappe,letzerp/framework,erpletzerp/letzerpcore,MaxMorais/frappe,drukhil/frappe,rohitwaghchaure/frappe-alec,ashokrajbathu/secondrep,hernad/frappe,jevonearth/frappe,suyashphadtare/propshikhari-frappe,BhupeshGupta/frappe,gangadharkadam/v4_frappe,indictranstech/fbd_frappe,gangadharkadam/tailorfrappe,hernad/frappe,shitolepriya/test-frappe,bcornwellmott/frappe,gangadhar-kadam/lgnlvefrape,sbkolate/sap_frappe_v6,indautgrp/frappe,indictranstech/frappe,mbauskar/omnitech-frappe,indictranstech/frappe,praba230890/frappe,gangadharkadam/v6_frappe,indictranstech/osmosis-frappe,pawaranand/phr-frappe,saurabh6790/frappe,rohitwaghchaure/frappe_smart,gangadharkadam/v4_frappe,indictranstech/osmosis-frappe,gangadharkadam/v6_frappe,gangadhar-kadam/verve_live_frappe,adityahase/frappe,bcornwellmott/frappe,almeidapaulopt/frappe,indictranstech/frappe-digitales,indictranstech/reciphergroup-frappe,mbauskar/omnitech-frappe,gangadharkadam/saloon_frappe,gangadhar-kadam/laganfrappe,gangadharkadam/vervefrappe,neilLasrado/frappe,mbauskar/frappe,gangadhar-kadam/verve_live_frappe,MaxMorais/frappe,saurabh6790/test-frappe,rohitw1991/smarttailorfrappe,StrellaGroup/frappe,gangadharkadam/vervefrappe,rohitwaghchaure/frappe-digitales,elba7r/builder,mbauskar/helpdesk-frappe,deveninfotech/deven-frappe,paurosello/frappe,mbauskar/frappe,adityahase/frappe,gangadhar-kadam/lgnlvefrape,indictranstech/trufil-frappe,indictranstech/omnitech-frappe,saurabh6790/test-frappe,ESS-LLP/frappe,RicardoJohann/frappe,pombredanne/frappe,mbauskar/phr-frappe,tundebabzy/frappe,tmimori/frappe,paurosello/frappe,anandpdoshi/frappe,neilLasrado/frappe,maxtorete/frappe,gangadharkadam/frappecontribution,gangadharkadam/frappecontribution,nerevu/frappe,maxtorete/frappe,Tejal011089/digitales_frappe,pawaranand/phr-frappe,saguas/frappe,mbauskar/omnitech-demo-frappe,sbkolate/sap_frappe_v6,sbktechnology/sap_frappe,indictranstech/trufil-frappe,gangadhar-kadam/smrterpfrappe,suyashphadtare/sajil-final-frappe,indictranstech/internal-frappe,pawaranand/phr-frappe,saguas/frappe,erpletzerp/letzerpcore,praba230890/frappe,Amber-Creative/amber-frappe,gangadhar-kadam/lgnlvefrape,indictranstech/osmosis-frappe,ShashaQin/frappe,indautgrp/frappe,gangadharkadam/saloon_frappe,indictranstech/phr-frappe,indictranstech/omnitech-frappe,suyashphadtare/propshikhari-frappe,frappe/frappe,nerevu/frappe,hatwar/buyback-frappe,sbktechnology/trufil-frappe,mbauskar/tele-frappe,aboganas/frappe,indictranstech/frappe-digitales,gangadhar-kadam/helpdesk-frappe,vjFaLk/frappe,almeidapaulopt/frappe,StrellaGroup/frappe,mbauskar/phr-frappe,mbauskar/helpdesk-frappe,indictranstech/frappe-digitales,tundebabzy/frappe,rohitwaghchaure/New_Theme_frappe,indictranstech/trufil-frappe,sbkolate/sap_frappe_v6,hernad/frappe,suyashphadtare/propshikhari-frappe,gangadharkadam/vlinkfrappe,geo-poland/frappe,indictranstech/omnitech-frappe,gangadhar-kadam/smrterpfrappe,gangadhar-kadam/verve_test_frappe,gangadhar-kadam/verve_test_frappe,rohitwaghchaure/frappe,gangadharkadam/office_frappe,gangadharkadam/letzfrappe,chdecultot/frappe,gangadharkadam/letzfrappe,gangadharkadam/saloon_frappe_install,indautgrp/frappe,MaxMorais/frappe,maxtorete/frappe,mhbu50/frappe,vCentre/vFRP-6233,chdecultot/frappe,almeidapaulopt/frappe,indictranstech/tele-frappe,jevonearth/frappe,hatwar/buyback-frappe,gangadharkadam/letzfrappe,mbauskar/Das_frappe,gangadharkadam/v6_frappe,saurabh6790/phr-frappe,gangadhar-kadam/verve_live_frappe,gangadharkadam/stfrappe,indictranstech/Das_frappe,gangadhar-kadam/helpdesk-frappe,mbauskar/phr-frappe,indictranstech/phr-frappe,aboganas/frappe,gangadharkadam/saloon_frappe_install,indictranstech/frappe,rohitwaghchaure/frappe-alec,paurosello/frappe,saurabh6790/phr-frappe,mbauskar/tele-frappe,mbauskar/omnitech-demo-frappe,bohlian/frappe,gangadhar-kadam/verve_frappe,paurosello/frappe,BhupeshGupta/frappe,ESS-LLP/frappe,mhbu50/frappe,chdecultot/frappe,rohitwaghchaure/frappe-alec,gangadhar-kadam/helpdesk-frappe,indictranstech/fbd_frappe,pranalik/frappe-bb,geo-poland/frappe,Tejal011089/digitales_frappe,rohitwaghchaure/vestasi-frappe,mhbu50/frappe,gangadharkadam/shfr,pombredanne/frappe,bohlian/frappe,pawaranand/phr_frappe,gangadhar-kadam/laganfrappe,bohlian/frappe,manassolanki/frappe,indictranstech/trufil-frappe,maxtorete/frappe,jevonearth/frappe,mbauskar/Das_frappe,rohitwaghchaure/frappe-digitales,tmimori/frappe,vqw/frappe,vjFaLk/frappe,saurabh6790/test-frappe,indictranstech/osmosis-frappe,mbauskar/helpdesk-frappe,pombredanne/frappe,vqw/frappe,bohlian/frappe,tmimori/frappe,letzerp/framework,ashokrajbathu/secondrep,geo-poland/frappe,RicardoJohann/frappe,manassolanki/frappe,bcornwellmott/frappe,gangadhar-kadam/verve_frappe,hatwar/buyback-frappe,pranalik/frappe-bb,gangadhar-kadam/verve_test_frappe,yashodhank/frappe,neilLasrado/frappe,gangadharkadam/smrtfrappe,rohitwaghchaure/vestasi-frappe,shitolepriya/test-frappe,indictranstech/ebuy-now-frappe,elba7r/builder,gangadharkadam/office_frappe,indictranstech/tele-frappe,gangadharkadam/johnfrappe,ShashaQin/frappe,indictranstech/reciphergroup-frappe,gangadharkadam/smrtfrappe,adityahase/frappe,vqw/frappe,reachalpineswift/frappe-bench,gangadharkadam/tailorfrappe,rohitwaghchaure/New_Theme_frappe,mbauskar/omnitech-demo-frappe,BhupeshGupta/frappe,gangadharkadam/v5_frappe,gangadharkadam/saloon_frappe_install,hernad/frappe,pawaranand/phr_frappe,indictranstech/tele-frappe,indictranstech/frappe-digitales,gangadharkadam/johnfrappe,gangadharkadam/frappecontribution,anandpdoshi/frappe,rohitwaghchaure/frappe-digitales,gangadhar-kadam/laganfrappe,shitolepriya/test-frappe,deveninfotech/deven-frappe,erpletzerp/letzerpcore,rmehta/frappe,pranalik/frappe-bb,chdecultot/frappe,indictranstech/reciphergroup-frappe,saurabh6790/phr-frappe,mbauskar/helpdesk-frappe,mbauskar/omnitech-frappe,indictranstech/ebuy-now-frappe,gangadharkadam/vervefrappe,indictranstech/Das_frappe,gangadhar-kadam/verve_frappe,vjFaLk/frappe,RicardoJohann/frappe,anandpdoshi/frappe,rohitwaghchaure/frappe,gangadharkadam/shfr,indictranstech/internal-frappe,indictranstech/internal-frappe,praba230890/frappe,gangadharkadam/v5_frappe,rohitwaghchaure/vestasi-frappe,mbauskar/frappe,suyashphadtare/propshikhari-frappe,reachalpineswift/frappe-bench,yashodhank/frappe,indictranstech/reciphergroup-frappe,rohitwaghchaure/vestasi-frappe,indautgrp/frappe,drukhil/frappe,gangadhar-kadam/verve_test_frappe,adityahase/frappe,saurabh6790/phr-frappe,praba230890/frappe,sbktechnology/sap_frappe,elba7r/builder,rmehta/frappe,manassolanki/frappe,gangadharkadam/vlinkfrappe,ShashaQin/frappe,mbauskar/omnitech-demo-frappe,pombredanne/frappe,pawaranand/phr_frappe,tmimori/frappe,deveninfotech/deven-frappe,rohitw1991/smartfrappe,gangadharkadam/v4_frappe,gangadharkadam/vervefrappe,rohitwaghchaure/New_Theme_frappe,MaxMorais/frappe,gangadharkadam/vlinkfrappe,almeidapaulopt/frappe,mhbu50/frappe,saurabh6790/frappe,indictranstech/frappe,neilLasrado/frappe,rohitwaghchaure/frappe,pranalik/frappe-bb,gangadharkadam/v6_frappe,drukhil/frappe,suyashphadtare/sajil-frappe,yashodhank/frappe,vCentre/vFRP-6233,saurabh6790/frappe,elba7r/frameworking,indictranstech/ebuy-now-frappe,indictranstech/fbd_frappe,ShashaQin/frappe,sbkolate/sap_frappe_v6,rohitwaghchaure/frappe,reachalpineswift/frappe-bench,suyashphadtare/sajil-final-frappe,mbauskar/omnitech-frappe,ashokrajbathu/secondrep,saguas/frappe,bcornwellmott/frappe,Tejal011089/digitales_frappe,letzerp/framework,gangadharkadam/v5_frappe,ESS-LLP/frappe,deveninfotech/deven-frappe,suyashphadtare/sajil-final-frappe,saurabh6790/test-frappe,anandpdoshi/frappe,frappe/frappe,mbauskar/Das_frappe,gangadhar-kadam/verve_frappe,jevonearth/frappe,gangadhar-kadam/verve_live_frappe,reachalpineswift/frappe-bench,mbauskar/tele-frappe,sbktechnology/sap_frappe,rohitw1991/smartfrappe,elba7r/frameworking,mbauskar/phr-frappe,Tejal011089/digitales_frappe,rohitwaghchaure/frappe-digitales,gangadharkadam/v5_frappe,elba7r/frameworking,gangadharkadam/frappecontribution,indictranstech/fbd_frappe,indictranstech/phr-frappe,tundebabzy/frappe,suyashphadtare/sajil-frappe,gangadharkadam/vlinkfrappe,Amber-Creative/amber-frappe,tundebabzy/frappe,indictranstech/internal-frappe,gangadharkadam/office_frappe,mbauskar/frappe,Amber-Creative/amber-frappe,indictranstech/Das_frappe,rohitw1991/smarttailorfrappe,vqw/frappe,sbktechnology/trufil-frappe,sbktechnology/trufil-frappe,yashodhank/frappe,ashokrajbathu/secondrep,pawaranand/phr_frappe,gangadharkadam/saloon_frappe_install,gangadharkadam/saloon_frappe
--- +++ @@ -4,10 +4,11 @@ from __future__ import unicode_literals import frappe from frappe.utils.scheduler import disable_scheduler, enable_scheduler +from frappe.utils import cint def execute(): frappe.reload_doc("core", "doctype", "system_settings") - if frappe.db.get_global("disable_scheduler"): + if cint(frappe.db.get_global("disable_scheduler")): disable_scheduler() else: enable_scheduler()
fd79823893b9b83a184c2bcd0fbe32fbb51619c9
src/server/convert.py
src/server/convert.py
# midi-beeper-orchestra - program to create an orchestra from PC speakers # Copyright (C) 2015 The Underscores # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as published # by the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. import math def MIDItoHz(MIDIval): """ Converts a MIDI note, MIDIval, value to the equivalent hertz value """ #return 69 + 12 * math.log((MIDIval/440), 2) return (2**((MIDIval-69)/12))*440 def hzToMIDI(hz): """ Converts hertz, hz, to MIDI note equivalent """ midi = 2**((hz-69)/12) * 440 return int(midi + 0.5)
# midi-beeper-orchestra - program to create an orchestra from PC speakers # Copyright (C) 2015 The Underscores # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as published # by the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. import math def MIDItoHz(MIDIval): """ Converts a MIDI note, MIDIval, value to the equivalent hertz value """ return (2**((MIDIval-69)/12))*440 def hzToMIDI(hz): """ Converts hertz, hz, to MIDI note equivalent """ midi = 2**((hz-69)/12) * 440 return int(midi + 0.5)
Remove comment containing incorrect conversion function.
Remove comment containing incorrect conversion function.
Python
agpl-3.0
TheUnderscores/midi-beeper-orchestra
--- +++ @@ -17,7 +17,6 @@ """ Converts a MIDI note, MIDIval, value to the equivalent hertz value """ - #return 69 + 12 * math.log((MIDIval/440), 2) return (2**((MIDIval-69)/12))*440 def hzToMIDI(hz):
d81a1ba12add244cb246efeae5c292a6d995c9b8
deadlinks.py
deadlinks.py
from operator import itemgetter from itertools import chain import os import yaml import requests yaml.load_all directory = "_companies" flat = chain.from_iterable def link_status_company(filename): (name, _) = filename.rsplit(".", 1); print("==== {name} ====".format(name=name)) docs = filter(None, yaml.load_all(open(os.path.join(directory, filename)))) positions = flat(map(itemgetter("positions"), filter(lambda doc: "positions" in doc, docs))) def link_status_position(position): title = position["title"] url = position["url"] print("{title} [ {url} ]".format(title=title, url=url)) response = requests.get(url) status_code_description = requests.status_codes._codes.get(response.status_code, '-') print("{} {} {}".format(response.status_code, status_code_description, response.history)) print() list(map(link_status_position, positions)) list(map(link_status_company, sorted(os.listdir(directory))))
from operator import itemgetter from itertools import chain import os import yaml import requests yaml.load_all directory = "_companies" flat = chain.from_iterable def link_status_company(filename): (name, _) = filename.rsplit(".", 1); print("==== {name} ====".format(name=name)) docs = filter(None, yaml.load_all(open(os.path.join(directory, filename)))) positions = flat(map(itemgetter("positions"), filter(lambda doc: "positions" in doc, docs))) def link_status_position(position): title = position["title"] url = position["url"] print("{title} [ {url} ]".format(title=title, url=url)) try: response = requests.get(url, timeout=10) status_code_description = requests.status_codes._codes.get(response.status_code, '-') print("{} {} {}".format(response.status_code, status_code_description, response.history)) except Exception as e: print(e) print() list(map(link_status_position, positions)) list(map(link_status_company, sorted(os.listdir(directory))))
Add timeout to dead links script
Add timeout to dead links script
Python
apache-2.0
Stockholm-AI/stockholm-ai,Stockholm-AI/stockholm-ai,Stockholm-AI/stockholm-ai,Stockholm-AI/stockholm-ai,Stockholm-AI/stockholm-ai
--- +++ @@ -23,11 +23,14 @@ url = position["url"] print("{title} [ {url} ]".format(title=title, url=url)) - response = requests.get(url) + try: + response = requests.get(url, timeout=10) - status_code_description = requests.status_codes._codes.get(response.status_code, '-') + status_code_description = requests.status_codes._codes.get(response.status_code, '-') - print("{} {} {}".format(response.status_code, status_code_description, response.history)) + print("{} {} {}".format(response.status_code, status_code_description, response.history)) + except Exception as e: + print(e) print()
64cbf90dbb8378752398d15d75ee5af9aa5b32ed
linter.py
linter.py
# # linter.py # Linter for SublimeLinter3, a code checking framework for Sublime Text 3 # # Written by Aparajita Fishman # Copyright (c) 2013 Aparajita Fishman # # Project: https://github.com/SublimeLinter/SublimeLinter-contrib-pep257 # License: MIT # """This module exports the PEP257 plugin linter class.""" import os from SublimeLinter.lint import highlight, PythonLinter class PEP257(PythonLinter): """Provides an interface to the pep257 python module/script.""" language = 'python' cmd = ('pep257@python', '-') regex = r'^.+?:(?P<line>\d+):(?P<col>\d+): (?P<message>.+)' default_type = highlight.WARNING line_col_base = (1, 0) # pep257 uses one-based line and zero-based column numbers module = 'pep257' def check(self, code, filename): """Run pep257 on code and return the output.""" return self.module.check_source(code, os.path.basename(filename))
# # linter.py # Linter for SublimeLinter3, a code checking framework for Sublime Text 3 # # Written by Aparajita Fishman # Copyright (c) 2013 Aparajita Fishman # # Project: https://github.com/SublimeLinter/SublimeLinter-contrib-pep257 # License: MIT # """This module exports the PEP257 plugin linter class.""" import os from SublimeLinter.lint import highlight, PythonLinter class PEP257(PythonLinter): """Provides an interface to the pep257 python module/script.""" syntax = 'python' cmd = ('pep257@python', '-') regex = r'^.+?:(?P<line>\d+):(?P<col>\d+): (?P<message>.+)' default_type = highlight.WARNING line_col_base = (1, 0) # pep257 uses one-based line and zero-based column numbers module = 'pep257' def check(self, code, filename): """Run pep257 on code and return the output.""" return self.module.check_source(code, os.path.basename(filename))
Change 'language' to 'syntax', that is more precise terminology.
Change 'language' to 'syntax', that is more precise terminology.
Python
mit
SublimeLinter/SublimeLinter-pep257
--- +++ @@ -20,7 +20,7 @@ """Provides an interface to the pep257 python module/script.""" - language = 'python' + syntax = 'python' cmd = ('pep257@python', '-') regex = r'^.+?:(?P<line>\d+):(?P<col>\d+): (?P<message>.+)' default_type = highlight.WARNING
96b9c25268e98e9464d8b068aa12de113ad1c66f
joby/spiders/data_science_jobs.py
joby/spiders/data_science_jobs.py
# -*- coding: utf-8 -*- import scrapy class DataScienceJobsSpider(scrapy.Spider): name = "data-science-jobs" allowed_domains = ["www.data-science-jobs.com"] start_urls = ( 'http://www.data-science-jobs.com/', ) def parse(self, response): pass
# -*- coding: utf-8 -*- from logging import getLogger from scrapy.spiders import Rule, CrawlSpider from scrapy.linkextractors import LinkExtractor class DataScienceJobsSpider(CrawlSpider): log = getLogger(__name__) name = 'data-science-jobs' allowed_domains = ['www.data-science-jobs.com', 'fonts.googleapis.com', 'jobs.lever.com'] start_urls = ['http://www.data-science-jobs.com/'] test = Rule(LinkExtractor(allow='family'), callback='parse') test2 = Rule(LinkExtractor(allow='comtravo'), callback='parse') job_links = Rule(LinkExtractor(allow='detail\/'), callback='parse') pagination_links = Rule(LinkExtractor(allow='\?page=\d+'), callback='parse') rules = [job_links, pagination_links, test, test2] response = None def parse(self, response): self.log.info('Parsing %s', response.url)
Add more rules for test purposes.
Add more rules for test purposes.
Python
mit
cyberbikepunk/job-spiders
--- +++ @@ -1,13 +1,21 @@ # -*- coding: utf-8 -*- -import scrapy + +from logging import getLogger +from scrapy.spiders import Rule, CrawlSpider +from scrapy.linkextractors import LinkExtractor -class DataScienceJobsSpider(scrapy.Spider): - name = "data-science-jobs" - allowed_domains = ["www.data-science-jobs.com"] - start_urls = ( - 'http://www.data-science-jobs.com/', - ) +class DataScienceJobsSpider(CrawlSpider): + log = getLogger(__name__) + name = 'data-science-jobs' + allowed_domains = ['www.data-science-jobs.com', 'fonts.googleapis.com', 'jobs.lever.com'] + start_urls = ['http://www.data-science-jobs.com/'] + test = Rule(LinkExtractor(allow='family'), callback='parse') + test2 = Rule(LinkExtractor(allow='comtravo'), callback='parse') + job_links = Rule(LinkExtractor(allow='detail\/'), callback='parse') + pagination_links = Rule(LinkExtractor(allow='\?page=\d+'), callback='parse') + rules = [job_links, pagination_links, test, test2] + response = None def parse(self, response): - pass + self.log.info('Parsing %s', response.url)
8bfb43e6af19e6335a95768289036fb73caac559
pinax/testimonials/templatetags/pinax_testimonials_tags.py
pinax/testimonials/templatetags/pinax_testimonials_tags.py
from django import template from ..models import Testimonial register = template.Library() @register.simple_tag def random_testimonials(number): return Testimonial.objects.filter(active=True).order_by("?")[:number] @register.simple_tag def random_testimonial(): queryset = Testimonial.objects.filter(active=True).order_by("?") return queryset[0] if queryset else None @register.simple_tag def testimonials(number): return Testimonial.objects.filter(active=True).order_by("-added")[:number]
from django import template from ..models import Testimonial register = template.Library() @register.simple_tag def random_testimonials(number): return Testimonial.objects.filter(active=True).order_by("?")[:number] @register.simple_tag def random_testimonial(): queryset = Testimonial.objects.filter(active=True).order_by("?") return queryset[0] if queryset else None @register.simple_tag def testimonials(number=None): return Testimonial.objects.filter(active=True).order_by("-added")[:number]
Make number parameter optional to testimonials tag
Make number parameter optional to testimonials tag
Python
mit
pinax/pinax-testimonials
--- +++ @@ -17,5 +17,5 @@ @register.simple_tag -def testimonials(number): +def testimonials(number=None): return Testimonial.objects.filter(active=True).order_by("-added")[:number]
0692cc324d3759703ee52e117ac19e75d82df6a6
tests/config/tests.py
tests/config/tests.py
from raven.conf import load from unittest2 import TestCase class LoadTest(TestCase): def test_basic(self): dsn = 'https://foo:bar@sentry.local/1' res = {} load(dsn, res) self.assertEquals(res, { 'SENTRY_PROJECT': '1', 'SENTRY_SERVERS': ['https://sentry.local/api/store/'], 'SENTRY_PUBLIC_KEY': 'foo', 'SENTRY_SECRET_KEY': 'bar', }) def test_path(self): dsn = 'https://foo:bar@sentry.local/app/1' res = {} load(dsn, res) self.assertEquals(res, { 'SENTRY_PROJECT': '1', 'SENTRY_SERVERS': ['https://sentry.local/app/api/store/'], 'SENTRY_PUBLIC_KEY': 'foo', 'SENTRY_SECRET_KEY': 'bar', })
import logging import mock from raven.conf import load, setup_logging from unittest2 import TestCase class LoadTest(TestCase): def test_basic(self): dsn = 'https://foo:bar@sentry.local/1' res = {} load(dsn, res) self.assertEquals(res, { 'SENTRY_PROJECT': '1', 'SENTRY_SERVERS': ['https://sentry.local/api/store/'], 'SENTRY_PUBLIC_KEY': 'foo', 'SENTRY_SECRET_KEY': 'bar', }) def test_path(self): dsn = 'https://foo:bar@sentry.local/app/1' res = {} load(dsn, res) self.assertEquals(res, { 'SENTRY_PROJECT': '1', 'SENTRY_SERVERS': ['https://sentry.local/app/api/store/'], 'SENTRY_PUBLIC_KEY': 'foo', 'SENTRY_SECRET_KEY': 'bar', }) class SetupLoggingTest(TestCase): def test_basic_not_configured(self): with mock.patch('logging.getLogger', spec=logging.getLogger) as getLogger: logger = getLogger() logger.handlers = [] handler = mock.Mock() result = setup_logging(handler) self.assertTrue(result) def test_basic_already_configured(self): with mock.patch('logging.getLogger', spec=logging.getLogger) as getLogger: handler = mock.Mock() logger = getLogger() logger.handlers = [handler] result = setup_logging(handler) self.assertFalse(result)
Add basic coverage for the setup_logging method
Add basic coverage for the setup_logging method
Python
bsd-3-clause
inspirehep/raven-python,jmagnusson/raven-python,Goldmund-Wyldebeast-Wunderliebe/raven-python,akalipetis/raven-python,inspirehep/raven-python,nikolas/raven-python,hzy/raven-python,smarkets/raven-python,beniwohli/apm-agent-python,icereval/raven-python,jmagnusson/raven-python,percipient/raven-python,dirtycoder/opbeat_python,someonehan/raven-python,johansteffner/raven-python,Photonomie/raven-python,recht/raven-python,lopter/raven-python-old,icereval/raven-python,arthurlogilab/raven-python,patrys/opbeat_python,getsentry/raven-python,alex/raven,jbarbuto/raven-python,lepture/raven-python,inspirehep/raven-python,jmagnusson/raven-python,Goldmund-Wyldebeast-Wunderliebe/raven-python,tarkatronic/opbeat_python,jmp0xf/raven-python,recht/raven-python,Goldmund-Wyldebeast-Wunderliebe/raven-python,smarkets/raven-python,jbarbuto/raven-python,lepture/raven-python,jbarbuto/raven-python,recht/raven-python,ronaldevers/raven-python,icereval/raven-python,daikeren/opbeat_python,ronaldevers/raven-python,nikolas/raven-python,beniwohli/apm-agent-python,someonehan/raven-python,getsentry/raven-python,beniwohli/apm-agent-python,percipient/raven-python,akalipetis/raven-python,smarkets/raven-python,patrys/opbeat_python,collective/mr.poe,inspirehep/raven-python,ticosax/opbeat_python,jmp0xf/raven-python,akalipetis/raven-python,patrys/opbeat_python,daikeren/opbeat_python,hzy/raven-python,danriti/raven-python,danriti/raven-python,jbarbuto/raven-python,ewdurbin/raven-python,ewdurbin/raven-python,daikeren/opbeat_python,tarkatronic/opbeat_python,nikolas/raven-python,Photonomie/raven-python,percipient/raven-python,openlabs/raven,lepture/raven-python,ticosax/opbeat_python,nikolas/raven-python,jmp0xf/raven-python,danriti/raven-python,akheron/raven-python,ewdurbin/raven-python,beniwohli/apm-agent-python,hzy/raven-python,smarkets/raven-python,someonehan/raven-python,icereval/raven-python,dirtycoder/opbeat_python,getsentry/raven-python,Goldmund-Wyldebeast-Wunderliebe/raven-python,akheron/raven-python,dbravender/raven-python,tarkatronic/opbeat_python,Photonomie/raven-python,arthurlogilab/raven-python,ronaldevers/raven-python,dbravender/raven-python,johansteffner/raven-python,johansteffner/raven-python,ticosax/opbeat_python,arthurlogilab/raven-python,dbravender/raven-python,akheron/raven-python,patrys/opbeat_python,arthurlogilab/raven-python,dirtycoder/opbeat_python
--- +++ @@ -1,4 +1,6 @@ -from raven.conf import load +import logging +import mock +from raven.conf import load, setup_logging from unittest2 import TestCase @@ -24,3 +26,21 @@ 'SENTRY_PUBLIC_KEY': 'foo', 'SENTRY_SECRET_KEY': 'bar', }) + + +class SetupLoggingTest(TestCase): + def test_basic_not_configured(self): + with mock.patch('logging.getLogger', spec=logging.getLogger) as getLogger: + logger = getLogger() + logger.handlers = [] + handler = mock.Mock() + result = setup_logging(handler) + self.assertTrue(result) + + def test_basic_already_configured(self): + with mock.patch('logging.getLogger', spec=logging.getLogger) as getLogger: + handler = mock.Mock() + logger = getLogger() + logger.handlers = [handler] + result = setup_logging(handler) + self.assertFalse(result)
1ec2779f5e4470c6ed19b56d16185c6174ab520c
tests/test_readers.py
tests/test_readers.py
# coding: utf-8 try: import unittest2 except ImportError, e: import unittest as unittest2 import datetime import os from pelican import readers CUR_DIR = os.path.dirname(__file__) CONTENT_PATH = os.path.join(CUR_DIR, 'content') def _filename(*args): return os.path.join(CONTENT_PATH, *args) class RstReaderTest(unittest2.TestCase): def test_article_with_metadata(self): reader = readers.RstReader({}) content, metadata = reader.read(_filename('article_with_metadata.rst')) expected = { 'category': 'yeah', 'author': u'Alexis Métaireau', 'title': 'This is a super article !', 'summary': 'Multi-line metadata should be supported\nas well as <strong>inline markup</strong>.', 'date': datetime.datetime(2010, 12, 2, 10, 14), 'tags': ['foo', 'bar', 'foobar'], } self.assertDictEqual(metadata, expected)
# coding: utf-8 try: import unittest2 as unittest except ImportError, e: import unittest import datetime import os from pelican import readers CUR_DIR = os.path.dirname(__file__) CONTENT_PATH = os.path.join(CUR_DIR, 'content') def _filename(*args): return os.path.join(CONTENT_PATH, *args) class RstReaderTest(unittest.TestCase): def test_article_with_metadata(self): reader = readers.RstReader({}) content, metadata = reader.read(_filename('article_with_metadata.rst')) expected = { 'category': 'yeah', 'author': u'Alexis Métaireau', 'title': 'This is a super article !', 'summary': 'Multi-line metadata should be supported\nas well as <strong>inline markup</strong>.', 'date': datetime.datetime(2010, 12, 2, 10, 14), 'tags': ['foo', 'bar', 'foobar'], } for key, value in expected.items(): self.assertEquals(value, metadata[key], key)
Make the readers tests a bit more verbose.
Make the readers tests a bit more verbose.
Python
agpl-3.0
11craft/pelican,51itclub/pelican,simonjj/pelican,sunzhongwei/pelican,goerz/pelican,51itclub/pelican,Rogdham/pelican,lucasplus/pelican,gymglish/pelican,sunzhongwei/pelican,sunzhongwei/pelican,koobs/pelican,UdeskDeveloper/pelican,ingwinlu/pelican,kennethlyn/pelican,karlcow/pelican,zackw/pelican,alexras/pelican,HyperGroups/pelican,rbarraud/pelican,crmackay/pelican,koobs/pelican,iurisilvio/pelican,jimperio/pelican,zackw/pelican,btnpushnmunky/pelican,ehashman/pelican,douglaskastle/pelican,HyperGroups/pelican,deanishe/pelican,iKevinY/pelican,ls2uper/pelican,karlcow/pelican,Rogdham/pelican,Rogdham/pelican,catdog2/pelican,kernc/pelican,number5/pelican,btnpushnmunky/pelican,crmackay/pelican,iurisilvio/pelican,Natim/pelican,GiovanniMoretti/pelican,avaris/pelican,HyperGroups/pelican,51itclub/pelican,deved69/pelican-1,liyonghelpme/myBlog,abrahamvarricatt/pelican,catdog2/pelican,kernc/pelican,jvehent/pelican,11craft/pelican,douglaskastle/pelican,jvehent/pelican,kennethlyn/pelican,garbas/pelican,gymglish/pelican,levanhien8/pelican,iurisilvio/pelican,lazycoder-ru/pelican,UdeskDeveloper/pelican,douglaskastle/pelican,koobs/pelican,treyhunner/pelican,florianjacob/pelican,Summonee/pelican,JeremyMorgan/pelican,joetboole/pelican,alexras/pelican,kennethlyn/pelican,ehashman/pelican,jo-tham/pelican,Polyconseil/pelican,JeremyMorgan/pelican,deved69/pelican-1,treyhunner/pelican,Scheirle/pelican,jimperio/pelican,deved69/pelican-1,avaris/pelican,Scheirle/pelican,btnpushnmunky/pelican,jo-tham/pelican,joetboole/pelican,fbs/pelican,ls2uper/pelican,eevee/pelican,ingwinlu/pelican,jvehent/pelican,TC01/pelican,talha131/pelican,iKevinY/pelican,kernc/pelican,florianjacob/pelican,getpelican/pelican,karlcow/pelican,goerz/pelican,deanishe/pelican,lazycoder-ru/pelican,levanhien8/pelican,JeremyMorgan/pelican,number5/pelican,liyonghelpme/myBlog,Summonee/pelican,levanhien8/pelican,ionelmc/pelican,janaurka/git-debug-presentiation,joetboole/pelican,ls2uper/pelican,farseerfc/pelican,simonjj/pelican,11craft/pelican,TC01/pelican,alexras/pelican,crmackay/pelican,goerz/pelican,liyonghelpme/myBlog,GiovanniMoretti/pelican,janaurka/git-debug-presentiation,lucasplus/pelican,number5/pelican,simonjj/pelican,abrahamvarricatt/pelican,treyhunner/pelican,eevee/pelican,lazycoder-ru/pelican,rbarraud/pelican,garbas/pelican,janaurka/git-debug-presentiation,arty-name/pelican,liyonghelpme/myBlog,rbarraud/pelican,TC01/pelican,catdog2/pelican,ehashman/pelican,Summonee/pelican,talha131/pelican,Scheirle/pelican,farseerfc/pelican,0xMF/pelican,justinmayer/pelican,florianjacob/pelican,deanishe/pelican,garbas/pelican,gymglish/pelican,abrahamvarricatt/pelican,jimperio/pelican,zackw/pelican,sunzhongwei/pelican,liyonghelpme/myBlog,UdeskDeveloper/pelican,getpelican/pelican,Polyconseil/pelican,lucasplus/pelican,eevee/pelican,GiovanniMoretti/pelican
--- +++ @@ -1,8 +1,8 @@ # coding: utf-8 try: - import unittest2 + import unittest2 as unittest except ImportError, e: - import unittest as unittest2 + import unittest import datetime import os @@ -12,11 +12,12 @@ CUR_DIR = os.path.dirname(__file__) CONTENT_PATH = os.path.join(CUR_DIR, 'content') + def _filename(*args): return os.path.join(CONTENT_PATH, *args) -class RstReaderTest(unittest2.TestCase): +class RstReaderTest(unittest.TestCase): def test_article_with_metadata(self): reader = readers.RstReader({}) @@ -29,4 +30,6 @@ 'date': datetime.datetime(2010, 12, 2, 10, 14), 'tags': ['foo', 'bar', 'foobar'], } - self.assertDictEqual(metadata, expected) + + for key, value in expected.items(): + self.assertEquals(value, metadata[key], key)
f3ea9820a96536e74e6f74f13387140c97ea9f2e
backgroundworker.py
backgroundworker.py
import sys import os sys.path.insert(0, "../financialScraper") import pandas as pd from financialScraper import getqf from sqlalchemy import create_engine running = True # engine = create_engine(os.environ.get('DATABASE_URL')) engine = create_engine('postgres://fkwfcpvbchmxps:VCmxue5WFWCOOHt56aqOm4FD_Z@ec2-54-83-205-46.compute-1.amazonaws.com:5432/d376d3nru8envq') dfdict = getqf.scraper() df = dfdict['nsdqct.csv'] df.to_sql(name='entries', con = engine, if_exists = 'replace')
import sys import os sys.path.insert(0, "../financialScraper") import pandas as pd from financialScraper import getqf from sqlalchemy import create_engine running = True # engine = create_engine(os.environ.get('DATABASE_URL')) engine = create_engine('postgres://fkwfcpvbchmxps:VCmxue5WFWCOOHt56aqOm4FD_Z@ec2-54-83-205-46.compute-1.amazonaws.com:5432/d376d3nru8envq') connection = engine.connect() dfdict = getqf.scraper() df = dfdict['nsdqct.csv'] df.to_sql(name='entries', con = connection, if_exists = 'replace') connection.close()
Add engine connection, and close engine connection to worker dyno
Add engine connection, and close engine connection to worker dyno
Python
mit
caseymacphee/green_quote,caseymacphee/green_quote
--- +++ @@ -9,7 +9,11 @@ # engine = create_engine(os.environ.get('DATABASE_URL')) + engine = create_engine('postgres://fkwfcpvbchmxps:VCmxue5WFWCOOHt56aqOm4FD_Z@ec2-54-83-205-46.compute-1.amazonaws.com:5432/d376d3nru8envq') +connection = engine.connect() dfdict = getqf.scraper() df = dfdict['nsdqct.csv'] -df.to_sql(name='entries', con = engine, if_exists = 'replace') +df.to_sql(name='entries', con = connection, if_exists = 'replace') + +connection.close()
8868cb556851d3caf227281873d619ec3ddc726a
matador/commands/deploy_ticket.py
matador/commands/deploy_ticket.py
#!/usr/bin/env python from .command import Command from matador import utils class DeployTicket(Command): def _add_arguments(self, parser): parser.add_argument( '-e', '--environment', type=str, required=True, help='Agresso environment name') def _execute(self): project_folder = utils.project_folder() self._logger.info(project_folder) working_folder = utils.working_folder('uog01', self.args.environment) self._logger.info(working_folder) project = utils.project() self._logger.info(project) self._logger.info(utils.is_git_repository())
#!/usr/bin/env python from .command import Command from matador import utils class DeployTicket(Command): def _add_arguments(self, parser): parser.prog = 'matador deploy-ticket' parser.add_argument( '-e', '--environment', type=str, required=True, help='Agresso environment name') def _execute(self): project_folder = utils.project_folder() self._logger.info(project_folder) working_folder = utils.working_folder('uog01', self.args.environment) self._logger.info(working_folder) project = utils.project() self._logger.info(project) self._logger.info(utils.is_git_repository())
Add program name to parser
Add program name to parser
Python
mit
Empiria/matador
--- +++ @@ -6,6 +6,7 @@ class DeployTicket(Command): def _add_arguments(self, parser): + parser.prog = 'matador deploy-ticket' parser.add_argument( '-e', '--environment', type=str,
8f82336aed62a18b2c6f824fcf0e6b1a1d00b8d3
tests/test_astroid.py
tests/test_astroid.py
# -*- coding: utf-8 -*- from __future__ import unicode_literals, print_function import re import astroid from . import tools, test_mark_tokens class TestAstroid(test_mark_tokens.TestMarkTokens): is_astroid_test = True module = astroid nodes_classes = astroid.ALL_NODE_CLASSES context_classes = [ (astroid.Name, astroid.DelName, astroid.AssignName), (astroid.Attribute, astroid.DelAttr, astroid.AssignAttr), ] @staticmethod def iter_fields(node): """ Yield a tuple of ``(fieldname, value)`` for each field that is present on *node*. Similar to ast.iter_fields, but for astroid and ignores context """ for field in node._astroid_fields + node._other_fields: if field == 'ctx': continue yield field, getattr(node, field) @classmethod def create_mark_checker(cls, source): builder = astroid.builder.AstroidBuilder() tree = builder.string_build(source) return tools.MarkChecker(source, tree=tree)
# -*- coding: utf-8 -*- from __future__ import unicode_literals, print_function import astroid from astroid.node_classes import NodeNG from . import tools, test_mark_tokens class TestAstroid(test_mark_tokens.TestMarkTokens): is_astroid_test = True module = astroid nodes_classes = NodeNG context_classes = [ (astroid.Name, astroid.DelName, astroid.AssignName), (astroid.Attribute, astroid.DelAttr, astroid.AssignAttr), ] @staticmethod def iter_fields(node): """ Yield a tuple of ``(fieldname, value)`` for each field that is present on *node*. Similar to ast.iter_fields, but for astroid and ignores context """ for field in node._astroid_fields + node._other_fields: if field == 'ctx': continue yield field, getattr(node, field) @classmethod def create_mark_checker(cls, source): builder = astroid.builder.AstroidBuilder() tree = builder.string_build(source) return tools.MarkChecker(source, tree=tree)
Use NodeNG instead of astroid.ALL_NODE_CLASSES
Use NodeNG instead of astroid.ALL_NODE_CLASSES
Python
apache-2.0
gristlabs/asttokens
--- +++ @@ -1,9 +1,8 @@ # -*- coding: utf-8 -*- from __future__ import unicode_literals, print_function -import re - import astroid +from astroid.node_classes import NodeNG from . import tools, test_mark_tokens @@ -13,7 +12,7 @@ is_astroid_test = True module = astroid - nodes_classes = astroid.ALL_NODE_CLASSES + nodes_classes = NodeNG context_classes = [ (astroid.Name, astroid.DelName, astroid.AssignName), (astroid.Attribute, astroid.DelAttr, astroid.AssignAttr),
59a345f29d77026c9bd8b2ec27df71253dc3d1e6
tests/test_helpers.py
tests/test_helpers.py
from scrapi.base import helpers class TestHelpers(object): def test_format_one_tag(self): single_tag = ' A single tag ' single_output = helpers.format_tags(single_tag) assert single_output == ['a single tag'] assert isinstance(single_output, list) def test_format_many_tags(self): many_tags = [' A', 'Bunch', ' oftags '] many_output = helpers.format_tags(many_tags) assert set(many_output) == set(['a', 'bunch', 'oftags']) def test_format_sep_tags(self): sep_tags = ['These, we know', 'should be many'] sep_output = helpers.format_tags(sep_tags, sep=',') assert set(sep_output) == set(['these', 'we know', 'should be many']) def test_extract_doi(self): identifiers = 'doi: THIS_IS_A_DOI!' valid_doi = helpers.oai_extract_doi(identifiers) assert valid_doi == 'THIS_IS_A_DOI!' def test_oai_extract_url(self): identifiers = 'I might be a url but rly I am naaaahhttt' extraction_attempt = helpers.oai_extract_url(identifiers) extraction_attempt def test_process_contributors(self): args = ['Stardust Rhodes', 'Golddust Rhodes', 'Dusty Rhodes'] response = helpers.oai_process_contributors(args) assert isinstance(response, list)
from scrapi.base import helpers class TestHelpers(object): def test_format_one_tag(self): single_tag = ' A single tag ' single_output = helpers.format_tags(single_tag) assert single_output == ['a single tag'] assert isinstance(single_output, list) def test_format_many_tags(self): many_tags = [' A', 'Bunch', ' oftags '] many_output = helpers.format_tags(many_tags) assert set(many_output) == set(['a', 'bunch', 'oftags']) def test_format_sep_tags(self): sep_tags = ['These, we know', 'should be many'] sep_output = helpers.format_tags(sep_tags, sep=',') assert set(sep_output) == set(['these', 'we know', 'should be many']) def test_extract_dois(self): identifiers = 'doi: THIS_IS_A_DOI!' valid_doi = helpers.oai_extract_dois(identifiers) assert valid_doi == 'THIS_IS_A_DOI!' def test_oai_extract_url(self): identifiers = 'I might be a url but rly I am naaaahhttt' extraction_attempt = helpers.oai_extract_url(identifiers) extraction_attempt def test_process_contributors(self): args = ['Stardust Rhodes', 'Golddust Rhodes', 'Dusty Rhodes'] response = helpers.oai_process_contributors(args) assert isinstance(response, list)
Update test to use new name of helper for dois
Update test to use new name of helper for dois
Python
apache-2.0
fabianvf/scrapi,CenterForOpenScience/scrapi,CenterForOpenScience/scrapi,mehanig/scrapi,icereval/scrapi,jeffreyliu3230/scrapi,ostwald/scrapi,felliott/scrapi,erinspace/scrapi,fabianvf/scrapi,felliott/scrapi,mehanig/scrapi,alexgarciac/scrapi,erinspace/scrapi
--- +++ @@ -19,9 +19,9 @@ sep_output = helpers.format_tags(sep_tags, sep=',') assert set(sep_output) == set(['these', 'we know', 'should be many']) - def test_extract_doi(self): + def test_extract_dois(self): identifiers = 'doi: THIS_IS_A_DOI!' - valid_doi = helpers.oai_extract_doi(identifiers) + valid_doi = helpers.oai_extract_dois(identifiers) assert valid_doi == 'THIS_IS_A_DOI!' def test_oai_extract_url(self):
71f007559ca2adf69d09f3c6f720383bfd784897
students/forms.py
students/forms.py
from django.contrib.auth import get_user_model from django.core.exceptions import ObjectDoesNotExist, ValidationError from django.utils.translation import ugettext_lazy as _ from registration.forms import RegistrationForm from .models import WhitelistedUsername User = get_user_model() class ExclusiveRegistrationForm(RegistrationForm): def __init__(self, *args, **kwargs): super(ExclusiveRegistrationForm, self).__init__(*args, **kwargs) self.fields['username'].label = 'Student number' def clean(self): # TODO: try catch KeyError here to avoid empty form error form_username = self.cleaned_data['username'] try: # If this runs without raising an exception, then the username is in # our database of whitelisted usernames. WhitelistedUsername.objects.get(username=form_username.lower()) except ObjectDoesNotExist: err = ValidationError(_('Unrecognised student number. Are you a CS1 student at UCT?s'), code='invalid') self.add_error(User.USERNAME_FIELD, err) super(ExclusiveRegistrationForm, self).clean()
from django.contrib.auth import get_user_model from django.contrib.auth.forms import AuthenticationForm from django.core.exceptions import ObjectDoesNotExist, ValidationError from django.utils.translation import ugettext_lazy as _ from registration.forms import RegistrationForm from .models import WhitelistedUsername User = get_user_model() class CustomAuthenticationForm(AuthenticationForm): def __init__(self, *args, **kwargs): super(CustomAuthenticationForm, self).__init__(*args, **kwargs) self.fields['username'].label = 'Student number' class ExclusiveRegistrationForm(RegistrationForm): def __init__(self, *args, **kwargs): super(ExclusiveRegistrationForm, self).__init__(*args, **kwargs) self.fields['username'].label = 'Student number' def clean(self): # TODO: try catch KeyError here to avoid empty form error form_username = self.cleaned_data['username'] try: # If this runs without raising an exception, then the username is in # our database of whitelisted usernames. WhitelistedUsername.objects.get(username=form_username.lower()) except ObjectDoesNotExist: err = ValidationError(_('Unrecognised student number. Are you a CS1 student at UCT?s'), code='invalid') self.add_error(User.USERNAME_FIELD, err) super(ExclusiveRegistrationForm, self).clean()
Add custom authentication form to set the label 'username' label to 'student number' for the login form.
Add custom authentication form to set the label 'username' label to 'student number' for the login form.
Python
mit
muhummadPatel/raspied,muhummadPatel/raspied,muhummadPatel/raspied
--- +++ @@ -1,4 +1,5 @@ from django.contrib.auth import get_user_model +from django.contrib.auth.forms import AuthenticationForm from django.core.exceptions import ObjectDoesNotExist, ValidationError from django.utils.translation import ugettext_lazy as _ @@ -8,6 +9,12 @@ User = get_user_model() + + +class CustomAuthenticationForm(AuthenticationForm): + def __init__(self, *args, **kwargs): + super(CustomAuthenticationForm, self).__init__(*args, **kwargs) + self.fields['username'].label = 'Student number' class ExclusiveRegistrationForm(RegistrationForm):
48075a16190bbcc3d260dfa242a5553b129de8a8
tests/test_see.py
tests/test_see.py
#!/usr/bin/env python # encoding: utf-8 """ Unit tests for see.py """ from __future__ import print_function, unicode_literals try: import unittest2 as unittest except ImportError: import unittest import see class TestSee(unittest.TestCase): def test_line_width(self): # Arrange default_width = 1 max_width = 1 # Act width = see.line_width(default_width, max_width) # Assert self.assertIsInstance(width, int) self.assertEqual(width, 1) def test_regex_filter(self): # Arrange names = ["george", "helen"] pat = "or*" # Act out = see.regex_filter(names, pat) # Assert self.assertIsInstance(out, tuple) self.assertEqual(out, ("george",)) def test_fn_filter(self): # Arrange names = ["george", "helen"] pat = "*or*" # Act out = see.fn_filter(names, pat) # Assert self.assertIsInstance(out, tuple) self.assertEqual(out, ("george",)) def test_see_with_no_args(self): # Act out = see.see() # Assert self.assertIsInstance(out, see._SeeOutput) if __name__ == '__main__': unittest.main() # End of file
#!/usr/bin/env python # encoding: utf-8 """ Unit tests for see.py """ from __future__ import print_function, unicode_literals try: import unittest2 as unittest except ImportError: import unittest import os import sys sys.path.insert(0, os.path.dirname(__file__)) import see class TestSee(unittest.TestCase): def test_line_width(self): # Arrange default_width = 1 max_width = 1 # Act width = see.line_width(default_width, max_width) # Assert self.assertIsInstance(width, int) self.assertEqual(width, 1) def test_regex_filter(self): # Arrange names = ["george", "helen"] pat = "or*" # Act out = see.regex_filter(names, pat) # Assert self.assertIsInstance(out, tuple) self.assertEqual(out, ("george",)) def test_fn_filter(self): # Arrange names = ["george", "helen"] pat = "*or*" # Act out = see.fn_filter(names, pat) # Assert self.assertIsInstance(out, tuple) self.assertEqual(out, ("george",)) def test_see_with_no_args(self): # Act out = see.see() # Assert self.assertIsInstance(out, see._SeeOutput) if __name__ == '__main__': unittest.main() # End of file
Update tests to import see
Update tests to import see
Python
bsd-3-clause
araile/see
--- +++ @@ -8,6 +8,11 @@ import unittest2 as unittest except ImportError: import unittest + +import os +import sys + +sys.path.insert(0, os.path.dirname(__file__)) import see
053a8f9deb8bfc0bb93cddcd48c8a7817bfe8c48
loom/analysis.py
loom/analysis.py
import six from datetime import datetime, date from collections import Mapping, Iterable from jsonmapping.transforms import transliterate IGNORE_FIELDS = ['$schema', '$sources', '$latin', '$text', '$attrcount', '$linkcount', 'id'] def latinize(text): """ Transliterate text to latin. """ if text is None or not len(text): return text return transliterate(text).lower() def extract_text(data, sep=' : '): """ Get all the instances of text from a given object, recursively. """ if isinstance(data, Mapping): values = [] for k, v in data.items(): if k in IGNORE_FIELDS: continue values.append(v) data = values if isinstance(data, (date, datetime)): data = data.isoformat() elif isinstance(data, (int, float)): data = six.text_type(data) if isinstance(data, six.string_types): return data if isinstance(data, Iterable): text = [extract_text(d, sep=sep) for d in data] return sep.join([t for t in text if t is not None])
import six from datetime import datetime, date from collections import Mapping, Iterable from jsonmapping.transforms import transliterate IGNORE_FIELDS = ['$schema', '$sources', '$latin', '$text', '$attrcount', '$linkcount', 'id'] def latinize(text): """ Transliterate text to latin. """ if text is None or not len(text): return text return transliterate(text).lower() def extract_text(data): """ Get all the instances of text from a given object, recursively. """ if isinstance(data, Mapping): values = [] for k, v in data.items(): if k in IGNORE_FIELDS: continue values.append(v) data = values if isinstance(data, (date, datetime)): data = data.isoformat() elif isinstance(data, (int, float)): data = six.text_type(data) if isinstance(data, six.string_types): return [data] if isinstance(data, Iterable): values = [] for d in data: values.extend(extract_text(d)) return values
Make text a list in the index.
Make text a list in the index.
Python
agpl-3.0
occrp/loom,occrp/datamapper
--- +++ @@ -15,7 +15,7 @@ return transliterate(text).lower() -def extract_text(data, sep=' : '): +def extract_text(data): """ Get all the instances of text from a given object, recursively. """ if isinstance(data, Mapping): values = [] @@ -29,7 +29,9 @@ elif isinstance(data, (int, float)): data = six.text_type(data) if isinstance(data, six.string_types): - return data + return [data] if isinstance(data, Iterable): - text = [extract_text(d, sep=sep) for d in data] - return sep.join([t for t in text if t is not None]) + values = [] + for d in data: + values.extend(extract_text(d)) + return values
af79bd6dce28e8147994a8fe2afb4df742dcd3eb
client/test_server_proxy.py
client/test_server_proxy.py
"""Code snippet to test the Java ServerProxy interface to the Pings server.""" import ServerProxy, ClientInfo sp = ServerProxy('localhost', 6543) if False: # Need to change permissions on ServerProxy Java class for this to work. print 'Calling doJsonRequest directly...' r = sp.doJsonRequest('/get_pings', None) print r print info = ClientInfo() pings = sp.getPings(info) print pings.token print pings.addresses[0] print pings.geoip_info[0] # Fill in results. for i in range(len(pings.addresses)): pings.results[i] = 'FOO %d' % i print print 'Submitting results' sp.submitResults(info, pings)
"""Code snippet to test the Java ServerProxy interface to the Pings server.""" import ServerProxy, ClientInfo sp = ServerProxy('localhost', 6543) if False: # Need to change permissions on ServerProxy Java class for this to work. print 'Calling doJsonRequest directly...' r = sp.doJsonRequest('/get_pings', None) print r print info = ClientInfo() pings = sp.getPings(info) print 'Token', pings.token print 'First address', pings.addresses[0] print 'Geoip for first address', pings.geoip_info[0] print 'Client Geoip', pings.client_geoip # Fill in results. for i in range(len(pings.addresses)): pings.results[i] = 'FOO %d' % i print print 'Submitting results' sp.submitResults(info, pings)
Add display of client geoip info. And print description of what we are printing.
Add display of client geoip info. And print description of what we are printing.
Python
bsd-3-clause
lisa-lab/pings,lisa-lab/pings,lisa-lab/pings,lisa-lab/pings
--- +++ @@ -13,9 +13,10 @@ print info = ClientInfo() pings = sp.getPings(info) -print pings.token -print pings.addresses[0] -print pings.geoip_info[0] +print 'Token', pings.token +print 'First address', pings.addresses[0] +print 'Geoip for first address', pings.geoip_info[0] +print 'Client Geoip', pings.client_geoip # Fill in results. for i in range(len(pings.addresses)):
fe5edfe737a774aa86cce578321fbb7fb4c8795e
tagcache/utils.py
tagcache/utils.py
# -*- encoding: utf-8 -*- import os import errno def ensure_intermediate_dir(path): """ Basiclly equivalent to command `mkdir -p` """ try: os.makedirs(os.path.dirname(path)) except OSError, e: if e.errno != errno.EEXIST: raise e def open_file(filename, flag, mode=0777): """ Wrapper of `os.open` which ensure intermediate dirs are created as well. """ try: return os.open(filename, flag, mode=mode) except OSError, e: if e.errno != errno.ENOENT or not (flag & os.O_CREAT): raise e # a directory component not exists ensure_intermediate_dir(filename) # second try return os.open(filename, flag, mode=mode) def link_file(src, dst): """ Wrapper of `os.link` which ensure intermediate dirs are created as well. """ try: return os.link(src, dst) except OSError, e: if e.errno != errno.ENOENT: raise e ensure_intermediate_dir(dst) return os.link(src, dst) def rename_file(old, new): """ Wrapper of `os.rename` which ensure intermediate dirs are created as well. """ try: return os.rename(old, new) except OSError, e: if e.errno != errno.ENOENT: raise e ensure_intermediate_dir(new) return os.rename(old, new)
# -*- encoding: utf-8 -*- import os import errno def ensure_intermediate_dir(path): """ Basiclly equivalent to command `mkdir -p` """ try: os.makedirs(os.path.dirname(path)) except OSError, e: if e.errno != errno.EEXIST: raise e def open_file(filename, flag, mode=0777): """ Wrapper of `os.open` which ensure intermediate dirs are created as well. """ try: return os.open(filename, flag, mode) except OSError, e: if e.errno != errno.ENOENT or not (flag & os.O_CREAT): raise e # a directory component not exists ensure_intermediate_dir(filename) # second try return os.open(filename, flag, mode) def link_file(src, dst): """ Wrapper of `os.link` which ensure intermediate dirs are created as well. """ try: return os.link(src, dst) except OSError, e: if e.errno != errno.ENOENT: raise e ensure_intermediate_dir(dst) return os.link(src, dst) def rename_file(old, new): """ Wrapper of `os.rename` which ensure intermediate dirs are created as well. """ try: return os.rename(old, new) except OSError, e: if e.errno != errno.ENOENT: raise e ensure_intermediate_dir(new) return os.rename(old, new)
Fix a bug in file_open (os.open does not take keyword argument).
Fix a bug in file_open (os.open does not take keyword argument).
Python
mit
huangjunwen/tagcache
--- +++ @@ -28,7 +28,7 @@ """ try: - return os.open(filename, flag, mode=mode) + return os.open(filename, flag, mode) except OSError, e: @@ -40,7 +40,7 @@ ensure_intermediate_dir(filename) # second try - return os.open(filename, flag, mode=mode) + return os.open(filename, flag, mode) def link_file(src, dst):
4ef8681f9dcd0f92be524925d3cacdae68c45616
tests/conftest.py
tests/conftest.py
# -*- coding: utf-8 -*- from pytest import fixture from iamport import Iamport DEFAULT_TEST_IMP_KEY = 'imp_apikey' DEFAULT_TEST_IMP_SECRET = ('ekKoeW8RyKuT0zgaZsUtXXTLQ4AhPFW3ZGseDA6bkA5lamv9O' 'qDMnxyeB9wqOsuO9W3Mx9YSJ4dTqJ3f') def pytest_addoption(parser): parser.addoption('--imp-key', default=DEFAULT_TEST_IMP_KEY, help='iamport client key for testing ' '[default: %default]') parser.addoption('--imp-secret', default=DEFAULT_TEST_IMP_SECRET, help='iamport secret key for testing ' '[default: %default]') @fixture def iamport(request): imp_key = request.config.getoption('--imp-key') imp_secret = request.config.getoption('--imp-secret') return Iamport(imp_key=imp_key, imp_secret=imp_secret)
# -*- coding: utf-8 -*- from pytest import fixture from iamport import Iamport DEFAULT_TEST_IMP_KEY = 'imp_apikey' DEFAULT_TEST_IMP_SECRET = ( 'ekKoeW8RyKuT0zgaZsUtXXTLQ4AhPFW3ZGseDA6b' 'kA5lamv9OqDMnxyeB9wqOsuO9W3Mx9YSJ4dTqJ3f' ) def pytest_addoption(parser): parser.addoption( '--imp-key', default=DEFAULT_TEST_IMP_KEY, help='iamport client key for testing ' '[default: %(default)s]' ) parser.addoption( '--imp-secret', default=DEFAULT_TEST_IMP_SECRET, help='iamport secret key for testing ' '[default: %(default)s]' ) @fixture def iamport(request): imp_key = request.config.getoption('--imp-key') imp_secret = request.config.getoption('--imp-secret') return Iamport(imp_key=imp_key, imp_secret=imp_secret)
Change %default to %(default)s for removing warning
Change %default to %(default)s for removing warning
Python
mit
iamport/iamport-rest-client-python
--- +++ @@ -3,18 +3,27 @@ from iamport import Iamport + DEFAULT_TEST_IMP_KEY = 'imp_apikey' -DEFAULT_TEST_IMP_SECRET = ('ekKoeW8RyKuT0zgaZsUtXXTLQ4AhPFW3ZGseDA6bkA5lamv9O' - 'qDMnxyeB9wqOsuO9W3Mx9YSJ4dTqJ3f') +DEFAULT_TEST_IMP_SECRET = ( + 'ekKoeW8RyKuT0zgaZsUtXXTLQ4AhPFW3ZGseDA6b' + 'kA5lamv9OqDMnxyeB9wqOsuO9W3Mx9YSJ4dTqJ3f' +) def pytest_addoption(parser): - parser.addoption('--imp-key', default=DEFAULT_TEST_IMP_KEY, - help='iamport client key for testing ' - '[default: %default]') - parser.addoption('--imp-secret', default=DEFAULT_TEST_IMP_SECRET, - help='iamport secret key for testing ' - '[default: %default]') + parser.addoption( + '--imp-key', + default=DEFAULT_TEST_IMP_KEY, + help='iamport client key for testing ' + '[default: %(default)s]' + ) + parser.addoption( + '--imp-secret', + default=DEFAULT_TEST_IMP_SECRET, + help='iamport secret key for testing ' + '[default: %(default)s]' + ) @fixture
3a1615238d4500f0fa7b9eea9ee2bfe460bc21f9
cax/tasks/purity.py
cax/tasks/purity.py
"""Add electron lifetime """ from sympy.parsing.sympy_parser import parse_expr from pax import units from cax import config from cax.task import Task class AddElectronLifetime(Task): "Add electron lifetime to dataset" def __init__(self): self.collection_purity = config.mongo_collection('purity') Task.__init__(self) def each_run(self): if 'processor' in self.run_doc: return # Fetch the latest electron lifetime fit doc = self.collection_purity.find_one(sort=(('calculation_time', -1),)) function = parse_expr(doc['electron_lifetime_function']) # Compute value from this function on this dataset lifetime = function.evalf(subs={"t" : self.run_doc['start'].timestamp()}) run_number = self.run_doc['number'] self.log.info("Run %d: calculated lifetime of %d us" % (run_number, lifetime)) if not config.DATABASE_LOG: return # Update run database key = 'processor.DEFAULT.electron_lifetime_liquid' self.collection.find_and_modify({'_id': self.run_doc['_id']}, {'$set': {key: lifetime * units.us}})
"""Add electron lifetime """ from sympy.parsing.sympy_parser import parse_expr from pax import units from cax import config from cax.task import Task class AddElectronLifetime(Task): "Add electron lifetime to dataset" def __init__(self): self.collection_purity = config.mongo_collection('purity') Task.__init__(self) def each_run(self): if 'processor' in self.run_doc: return # Fetch the latest electron lifetime fit doc = self.collection_purity.find_one(sort=(('calculation_time', -1),)) function = parse_expr(doc['electron_lifetime_function']) # Compute value from this function on this dataset lifetime = function.evalf(subs={"t" : self.run_doc['start'].timestamp()}) lifetime = float(lifetime) # Convert away from Sympy type. run_number = self.run_doc['number'] self.log.info("Run %d: calculated lifetime of %d us" % (run_number, lifetime)) if not config.DATABASE_LOG: return # Update run database key = 'processor.DEFAULT.electron_lifetime_liquid' self.collection.find_and_modify({'_id': self.run_doc['_id']}, {'$set': {key: lifetime * units.us}})
Convert lifetime to float instead of sympy type.
Convert lifetime to float instead of sympy type.
Python
isc
XENON1T/cax,XENON1T/cax
--- +++ @@ -28,6 +28,7 @@ # Compute value from this function on this dataset lifetime = function.evalf(subs={"t" : self.run_doc['start'].timestamp()}) + lifetime = float(lifetime) # Convert away from Sympy type. run_number = self.run_doc['number'] self.log.info("Run %d: calculated lifetime of %d us" % (run_number,
46511322dc8d738cc43561025bca3298946da2e6
server.py
server.py
from swiftdav.swiftdav import SwiftProvider, WsgiDAVDomainController from waitress import serve from wsgidav.wsgidav_app import DEFAULT_CONFIG, WsgiDAVApp proxy = 'http://127.0.0.1:8080/auth/v1.0' insecure = False # Set to True to disable SSL certificate validation config = DEFAULT_CONFIG.copy() config.update({ "provider_mapping": {"": SwiftProvider()}, "verbose": 1, "propsmanager": True, "locksmanager": True, "acceptbasic": True, "acceptdigest": False, "defaultdigest": False, "domaincontroller": WsgiDAVDomainController(proxy, insecure) }) app = WsgiDAVApp(config) serve(app, host="0.0.0.0", port=8000)
from swiftdav.swiftdav import SwiftProvider, WsgiDAVDomainController from waitress import serve from wsgidav.wsgidav_app import DEFAULT_CONFIG, WsgiDAVApp proxy = 'http://127.0.0.1:8080/auth/v1.0' insecure = False # Set to True to disable SSL certificate validation config = DEFAULT_CONFIG.copy() config.update({ "provider_mapping": {"": SwiftProvider()}, "verbose": 1, "propsmanager": True, "locksmanager": True, "acceptbasic": True, "acceptdigest": False, "defaultdigest": False, "domaincontroller": WsgiDAVDomainController(proxy, insecure) }) app = WsgiDAVApp(config) serve(app, host="0.0.0.0", port=8000, max_request_body_size=5*1024*1024*1024)
Increase waitress setting max_request_body_size to 5GiB
Increase waitress setting max_request_body_size to 5GiB Python waitress limits the body size to 1GiB by default, thus uploading of larger objects will fail if this value is not increased. Please note that this value should be increased if your Swift cluster supports uploading of objects larger than 5GiB.
Python
apache-2.0
cschwede/swiftdav,cschwede/swiftdav
--- +++ @@ -18,4 +18,4 @@ }) app = WsgiDAVApp(config) -serve(app, host="0.0.0.0", port=8000) +serve(app, host="0.0.0.0", port=8000, max_request_body_size=5*1024*1024*1024)
5ef05f6c75e3dd0febffd527aebfdf8c2ed6bb51
server.py
server.py
import argparse import json from flask import Flask, request parser = argparse.ArgumentParser(description="Start a Blindstore server.") parser.add_argument('-d', '--debug', action='store_true', help="enable Flask debug mode. DO NOT use in production.") args = parser.parse_args() NUM_RECORDS = 5 RECORD_SIZE = 64 app = Flask(__name__) @app.route('/db_size') def get_db_size(): return json.dumps({'num_records': NUM_RECORDS, 'record_size': RECORD_SIZE}) @app.route('/retrieve', methods=['POST']) def get(): public_key = request.form['PUBLIC_KEY'] enc_index = request.form['ENC_INDEX'] return "/retrieve index '{index}' with key '{key}'".format(index=enc_index, key=public_key) @app.route('/set', methods=['POST']) def put(): enc_index = request.form['ENC_INDEX'] enc_data = request.form['ENC_DATA'] return "/set '{index}' to '{data}'".format(data=enc_data, index=enc_index) if __name__ == '__main__': app.run(debug=args.debug)
import argparse import json from flask import Flask, request parser = argparse.ArgumentParser(description="Start a Blindstore server.") parser.add_argument('-d', '--debug', action='store_true', help="enable Flask debug mode. DO NOT use in production.") args = parser.parse_args() NUM_RECORDS = 5 RECORD_SIZE = 64 app = Flask(__name__) @app.route('/db_size') def get_db_size(): return json.dumps({'num_records': NUM_RECORDS, 'record_size': RECORD_SIZE}), \ 200, {'Content-Type': 'text/json'} @app.route('/retrieve', methods=['POST']) def get(): public_key = request.form['PUBLIC_KEY'] enc_index = request.form['ENC_INDEX'] return "/retrieve index '{index}' with key '{key}'".format(index=enc_index, key=public_key) @app.route('/set', methods=['POST']) def put(): enc_index = request.form['ENC_INDEX'] enc_data = request.form['ENC_DATA'] return "/set '{index}' to '{data}'".format(data=enc_data, index=enc_index) if __name__ == '__main__': app.run(debug=args.debug)
Return Content-Type header with JSON size info
Return Content-Type header with JSON size info
Python
mit
blindstore/blindstore-old-scarab
--- +++ @@ -15,7 +15,8 @@ @app.route('/db_size') def get_db_size(): - return json.dumps({'num_records': NUM_RECORDS, 'record_size': RECORD_SIZE}) + return json.dumps({'num_records': NUM_RECORDS, 'record_size': RECORD_SIZE}), \ + 200, {'Content-Type': 'text/json'} @app.route('/retrieve', methods=['POST']) def get():
0f7ebf148ab3f88fc983e60f689a9c740ae64e47
outgoing_mail.py
outgoing_mail.py
#!/usr/bin/env python # # Copyright 2010 Eric Entzel <eric@ubermac.net> # from google.appengine.api import mail from google.appengine.ext.webapp import template import os from_address = 'admin@' + os.environ['APPLICATION_ID'] + '.appspotmail.com' def send(to, template_name, values): path = os.path.join(os.path.dirname(__file__), 'email_templates', template_name) message = mail.EmailMessage(sender=from_address, to=to) message.subject = template.render(path + '.subject', values) message.body = template.render(path + '.body', values) message.send()
#!/usr/bin/env python # # Copyright 2010 Eric Entzel <eric@ubermac.net> # from google.appengine.api import mail from google.appengine.ext.webapp import template import os from_address = 'EventBot <admin@' + os.environ['APPLICATION_ID'] + '.appspotmail.com>' def send(to, template_name, values): path = os.path.join(os.path.dirname(__file__), 'email_templates', template_name) message = mail.EmailMessage(sender=from_address, to=to) message.subject = template.render(path + '.subject', values) message.body = template.render(path + '.body', values) message.send()
Add display name for from address
Add display name for from address
Python
mit
eentzel/myeventbot,eentzel/myeventbot,eentzel/myeventbot,eentzel/myeventbot,eentzel/myeventbot
--- +++ @@ -8,7 +8,7 @@ import os -from_address = 'admin@' + os.environ['APPLICATION_ID'] + '.appspotmail.com' +from_address = 'EventBot <admin@' + os.environ['APPLICATION_ID'] + '.appspotmail.com>' def send(to, template_name, values):
3822b5b142d54f83aadf7e366f2b5b925f557e1a
test/testUtils/__init__.py
test/testUtils/__init__.py
import ibmiotf.application import os class AbstractTest(object): WIOTP_API_KEY=os.getenv("WIOTP_API_KEY") WIOTP_API_TOKEN=os.getenv("WIOTP_API_TOKEN") ORG_ID = os.getenv("WIOTP_ORG_ID") appOptions = {'auth-key': WIOTP_API_KEY, 'auth-token': WIOTP_API_TOKEN} setupAppClient = ibmiotf.application.Client(appOptions)
import ibmiotf.application import os class AbstractTest(object): WIOTP_API_KEY=os.getenv("WIOTP_API_KEY") WIOTP_API_TOKEN=os.getenv("WIOTP_API_TOKEN") ORG_ID = os.getenv("WIOTP_ORG_ID") if WIOTP_API_KEY is None: raise Exception("WIOTP_API_KEY environment variable is not set") if WIOTP_API_TOKEN is None: raise Exception("WIOTP_API_TOKEN environment variable is not set") if ORG_ID is None: raise Exception("WIOTP_ORG_ID environment variable is not set") appOptions = {'auth-key': WIOTP_API_KEY, 'auth-token': WIOTP_API_TOKEN} setupAppClient = ibmiotf.application.Client(appOptions)
Make tests throw better error if env vars are missing
Make tests throw better error if env vars are missing
Python
epl-1.0
ibm-watson-iot/iot-python,ibm-watson-iot/iot-python,ibm-messaging/iot-python
--- +++ @@ -7,5 +7,12 @@ WIOTP_API_TOKEN=os.getenv("WIOTP_API_TOKEN") ORG_ID = os.getenv("WIOTP_ORG_ID") + if WIOTP_API_KEY is None: + raise Exception("WIOTP_API_KEY environment variable is not set") + if WIOTP_API_TOKEN is None: + raise Exception("WIOTP_API_TOKEN environment variable is not set") + if ORG_ID is None: + raise Exception("WIOTP_ORG_ID environment variable is not set") + appOptions = {'auth-key': WIOTP_API_KEY, 'auth-token': WIOTP_API_TOKEN} setupAppClient = ibmiotf.application.Client(appOptions)
b853abc579f5dfaab896cf57c39268a36c109a83
tests/test_address_book.py
tests/test_address_book.py
from unittest import TestCase class AddressBookTestCase(TestCase): def test_add_person(self): pass def test_add_group(self): pass def test_find_person_by_first_name(self): pass def test_find_person_by_last_name(self): pass def test_find_person_by_email(self): passjjj
from unittest import TestCase class AddressBookTestCase(TestCase): def test_add_person(self): person = Person( 'John', 'Doe', ['Russian Federation, Kemerovo region, Kemerovo, Kirova street 23, apt. 42'], ['+79834772053'] ) self.address_book.add_person(person) self.assertIn(person, self.address_book) def test_add_group(self): pass def test_find_person_by_first_name(self): pass def test_find_person_by_last_name(self): pass def test_find_person_by_email(self): passjjj
Test person adding to addressbook
Test person adding to addressbook
Python
mit
dizpers/python-address-book-assignment
--- +++ @@ -4,7 +4,14 @@ class AddressBookTestCase(TestCase): def test_add_person(self): - pass + person = Person( + 'John', + 'Doe', + ['Russian Federation, Kemerovo region, Kemerovo, Kirova street 23, apt. 42'], + ['+79834772053'] + ) + self.address_book.add_person(person) + self.assertIn(person, self.address_book) def test_add_group(self): pass
097eae49564a8eefd66d903d8e8cd900054ef147
characters/views.py
characters/views.py
from django.shortcuts import get_object_or_404, redirect, render from django.views import generic from characters.forms import CharacterForm from characters.models import Character, Class, Race class CharacterIndexView(generic.ListView): template_name = 'characters/index.html' context_object_name = 'all_characters' # better than 'object_list' def get_queryset(self): return Character.objects.all() class CharacterDetailView(generic.DetailView): model = Character template_name = 'characters/view_character.html' def create_character(request): form = CharacterForm(request.POST or None) if request.method == 'POST' and form.is_valid(): character = Character( name=request.POST['name'], background=request.POST['background'], race_id=1, cclass_id=1 ) character.save() return redirect('characters:view', character_id=character.id) context = {'form': form} return render(request, 'characters/create_character.html', context)
from django.shortcuts import get_object_or_404, redirect, render from django.views import generic from characters.forms import CharacterForm from characters.models import Character, Class, Race class CharacterIndexView(generic.ListView): template_name = 'characters/index.html' context_object_name = 'all_characters' # better than 'object_list' def get_queryset(self): return Character.objects.all().order_by('name') class CharacterDetailView(generic.DetailView): model = Character template_name = 'characters/view_character.html' def create_character(request): form = CharacterForm(request.POST or None) if request.method == 'POST' and form.is_valid(): character = Character( name=request.POST['name'], background=request.POST['background'], race_id=1, cclass_id=1 ) character.save() return redirect('characters:view', character_id=character.id) context = {'form': form} return render(request, 'characters/create_character.html', context)
Order character listing by name
Order character listing by name
Python
mit
mpirnat/django-tutorial-v2
--- +++ @@ -11,7 +11,7 @@ context_object_name = 'all_characters' # better than 'object_list' def get_queryset(self): - return Character.objects.all() + return Character.objects.all().order_by('name') class CharacterDetailView(generic.DetailView):
fefde8aef88cbfb13cb1f0bfcd3ac476ad7a903c
spacy/download.py
spacy/download.py
from __future__ import print_function import sys import sputnik from sputnik.package_list import (PackageNotFoundException, CompatiblePackageNotFoundException) from . import about def download(lang, force=False, fail_on_exist=True): if force: sputnik.purge(about.__title__, about.__version__) try: sputnik.package(about.__title__, about.__version__, about.__models__.get(lang, lang)) if fail_on_exist: print("Model already installed. Please run 'python -m " "spacy.%s.download --force' to reinstall." % lang, file=sys.stderr) sys.exit(0) except (PackageNotFoundException, CompatiblePackageNotFoundException): pass package = sputnik.install(about.__title__, about.__version__, about.__models__.get(lang, lang)) try: sputnik.package(about.__title__, about.__version__, about.__models__.get(lang, lang)) except (PackageNotFoundException, CompatiblePackageNotFoundException): print("Model failed to install. Please run 'python -m " "spacy.%s.download --force'." % lang, file=sys.stderr) sys.exit(1) print("Model successfully installed.", file=sys.stderr)
from __future__ import print_function import sys import sputnik from sputnik.package_list import (PackageNotFoundException, CompatiblePackageNotFoundException) from . import about from . import util def download(lang, force=False, fail_on_exist=True): if force: sputnik.purge(about.__title__, about.__version__) try: sputnik.package(about.__title__, about.__version__, about.__models__.get(lang, lang)) if fail_on_exist: print("Model already installed. Please run 'python -m " "spacy.%s.download --force' to reinstall." % lang, file=sys.stderr) sys.exit(0) except (PackageNotFoundException, CompatiblePackageNotFoundException): pass package = sputnik.install(about.__title__, about.__version__, about.__models__.get(lang, lang)) try: sputnik.package(about.__title__, about.__version__, about.__models__.get(lang, lang)) except (PackageNotFoundException, CompatiblePackageNotFoundException): print("Model failed to install. Please run 'python -m " "spacy.%s.download --force'." % lang, file=sys.stderr) sys.exit(1) data_path = util.get_data_path() print("Model successfully installed to %s" % data_path, file=sys.stderr)
Make installation print data path.
Make installation print data path.
Python
mit
explosion/spaCy,explosion/spaCy,aikramer2/spaCy,recognai/spaCy,aikramer2/spaCy,recognai/spaCy,oroszgy/spaCy.hu,Gregory-Howard/spaCy,recognai/spaCy,Gregory-Howard/spaCy,recognai/spaCy,Gregory-Howard/spaCy,banglakit/spaCy,oroszgy/spaCy.hu,oroszgy/spaCy.hu,spacy-io/spaCy,spacy-io/spaCy,raphael0202/spaCy,banglakit/spaCy,explosion/spaCy,honnibal/spaCy,recognai/spaCy,oroszgy/spaCy.hu,raphael0202/spaCy,Gregory-Howard/spaCy,aikramer2/spaCy,aikramer2/spaCy,aikramer2/spaCy,honnibal/spaCy,spacy-io/spaCy,explosion/spaCy,banglakit/spaCy,Gregory-Howard/spaCy,oroszgy/spaCy.hu,honnibal/spaCy,banglakit/spaCy,raphael0202/spaCy,raphael0202/spaCy,Gregory-Howard/spaCy,spacy-io/spaCy,recognai/spaCy,aikramer2/spaCy,explosion/spaCy,raphael0202/spaCy,banglakit/spaCy,explosion/spaCy,oroszgy/spaCy.hu,spacy-io/spaCy,banglakit/spaCy,spacy-io/spaCy,honnibal/spaCy,raphael0202/spaCy
--- +++ @@ -7,6 +7,7 @@ CompatiblePackageNotFoundException) from . import about +from . import util def download(lang, force=False, fail_on_exist=True): @@ -34,4 +35,5 @@ "spacy.%s.download --force'." % lang, file=sys.stderr) sys.exit(1) - print("Model successfully installed.", file=sys.stderr) + data_path = util.get_data_path() + print("Model successfully installed to %s" % data_path, file=sys.stderr)
267a768bd1ccc87c3c1f54c4ac520a0e12e5fd5f
moksha/tests/test_clientsockets.py
moksha/tests/test_clientsockets.py
import webtest import moksha.tests.utils as testutils from moksha.api.widgets.live import get_moksha_socket from moksha.middleware import make_moksha_middleware from tw2.core import make_middleware as make_tw2_middleware class TestClientSocketDumb: def _setUp(self): def kernel(config): def app(environ, start_response): start_response('200 OK', [('Content-Type', 'text/html')]) socket = get_moksha_socket(config) return map(str, [socket.display()]) app = make_moksha_middleware(app, config) app = make_tw2_middleware(app, config) app = webtest.TestApp(app) self.app = app for _setup, name in testutils.make_setup_functions(kernel): yield _setup, name def _tearDown(self): pass @testutils.crosstest def test_middleware_wrap(self): targets = ['moksha_websocket', 'TCPSocket'] response = self.app.get('/') assert(any([target in response for target in targets]))
import webtest import moksha.tests.utils as testutils from moksha.api.widgets.live import get_moksha_socket from moksha.middleware import make_moksha_middleware from tw2.core import make_middleware as make_tw2_middleware class TestClientSocketDumb: def _setUp(self): def kernel(config): def app(environ, start_response): start_response('200 OK', [('Content-Type', 'text/html')]) socket = get_moksha_socket(config) return map(str, [socket.display()]) app = make_moksha_middleware(app, config) app = make_tw2_middleware(app, config) app = webtest.TestApp(app) self.app = app for _setup, name in testutils.make_setup_functions(kernel): yield _setup, name def _tearDown(self): pass @testutils.crosstest def test_has_socket_str(self): targets = ['moksha_websocket', 'TCPSocket'] response = self.app.get('/') assert(any([target in response for target in targets]))
Rename test. Fix copy/pasta forgetfulness.
Rename test. Fix copy/pasta forgetfulness.
Python
apache-2.0
pombredanne/moksha,mokshaproject/moksha,mokshaproject/moksha,mokshaproject/moksha,pombredanne/moksha,pombredanne/moksha,mokshaproject/moksha,pombredanne/moksha
--- +++ @@ -27,7 +27,7 @@ pass @testutils.crosstest - def test_middleware_wrap(self): + def test_has_socket_str(self): targets = ['moksha_websocket', 'TCPSocket'] response = self.app.get('/') assert(any([target in response for target in targets]))
68cec560ad108e0e6a081ed92aab2f06a7545821
tests/conftest.py
tests/conftest.py
import pytest @pytest.fixture(autouse=True) def tagschecker(request): tags = set(request.config.getini('TAGS')) tags_marker = request.node.get_marker('tags') xfailtags_marker = request.node.get_marker('xfailtags') skiptags_marker = request.node.get_marker('skiptags') if xfailtags_marker and not tags.isdisjoint(set(xfailtags_marker.args)): request.node.add_marker(pytest.mark.xfail()) elif ( tags_marker and tags.isdisjoint(set(tags_marker.args)) or skiptags_marker and not tags.isdisjoint(set(skiptags_marker.args)) ): pytest.skip('skipped for this tags: {}'.format(tags))
import pytest from docker import Client @pytest.fixture(scope="session") def docker_client(): client = Client(base_url='unix://var/run/docker.sock', timeout=180) return client @pytest.fixture(autouse=True) def tagschecker(request): tags = set(request.config.getini('TAGS')) tags_marker = request.node.get_marker('tags') xfailtags_marker = request.node.get_marker('xfailtags') skiptags_marker = request.node.get_marker('skiptags') if xfailtags_marker and not tags.isdisjoint(set(xfailtags_marker.args)): request.node.add_marker(pytest.mark.xfail()) elif ( tags_marker and tags.isdisjoint(set(tags_marker.args)) or skiptags_marker and not tags.isdisjoint(set(skiptags_marker.args)) ): pytest.skip('skipped for this tags: {}'.format(tags))
Increase docker-py timeout to 180
Increase docker-py timeout to 180
Python
mit
dincamihai/salt-toaster,dincamihai/salt-toaster
--- +++ @@ -1,4 +1,11 @@ import pytest +from docker import Client + + +@pytest.fixture(scope="session") +def docker_client(): + client = Client(base_url='unix://var/run/docker.sock', timeout=180) + return client @pytest.fixture(autouse=True)