commit
stringlengths 40
40
| old_file
stringlengths 4
150
| new_file
stringlengths 4
150
| old_contents
stringlengths 0
3.26k
| new_contents
stringlengths 1
4.43k
| subject
stringlengths 15
501
| message
stringlengths 15
4.06k
| lang
stringclasses 4
values | license
stringclasses 13
values | repos
stringlengths 5
91.5k
| diff
stringlengths 0
4.35k
|
|---|---|---|---|---|---|---|---|---|---|---|
91ad56ea892d2f2fdb2af97f81ec70a7b9f9305c
|
analysis/sanity-check-velocity.py
|
analysis/sanity-check-velocity.py
|
#!/usr/bin/env python
import climate
import joblib
import lmj.cubes
import numpy as np
def _check(t):
t.load()
t.add_velocities(smooth=0)
vel = abs(t.df[t.marker_velocity_columns].values).flatten()
vel = vel[np.isfinite(vel)]
pct = np.percentile(vel, [1, 2, 5, 10, 20, 50, 80, 90, 95, 98, 99])
print(t.subject.key, t.block.key, t.key, *pct)
def main(root):
trials = lmj.cubes.Experiment(root).trials_matching('*')
check = joblib.delayed(_check)
joblib.Parallel(-1)(check(t) for t in trials)
if __name__ == '__main__':
climate.call(main)
|
#!/usr/bin/env python
import climate
import joblib
import lmj.cubes
import numpy as np
def _check(t):
t.load()
t.add_velocities(smooth=0)
t.add_accelerations(smooth=0)
vel = abs(t.df[t.marker_velocity_columns].values).flatten()
vel = vel[np.isfinite(vel)]
pct = np.percentile(vel, [1, 2, 5, 10, 20, 50, 80, 90, 95, 98, 99])
np.set_printoptions(suppress=True, linewidth=1000, precision=2)
t.log('%s', pct)
def main(root):
trials = lmj.cubes.Experiment(root).trials_matching('*')
check = joblib.delayed(_check)
joblib.Parallel(-1)(check(t) for t in trials)
if __name__ == '__main__':
climate.call(main)
|
Use trial logging. Tweak numpy logging output.
|
Use trial logging. Tweak numpy logging output.
|
Python
|
mit
|
lmjohns3/cube-experiment,lmjohns3/cube-experiment,lmjohns3/cube-experiment
|
---
+++
@@ -9,10 +9,12 @@
def _check(t):
t.load()
t.add_velocities(smooth=0)
+ t.add_accelerations(smooth=0)
vel = abs(t.df[t.marker_velocity_columns].values).flatten()
vel = vel[np.isfinite(vel)]
pct = np.percentile(vel, [1, 2, 5, 10, 20, 50, 80, 90, 95, 98, 99])
- print(t.subject.key, t.block.key, t.key, *pct)
+ np.set_printoptions(suppress=True, linewidth=1000, precision=2)
+ t.log('%s', pct)
def main(root):
|
0cccd467ac4c0bd8b8110fcfe47f81d73a238aa9
|
plugins/uptime.py
|
plugins/uptime.py
|
import time
class Plugin:
def __init__(self, vk_bot):
self.vk_bot = vk_bot
self.vk_bot.add_command('uptime', self.uptime)
async def uptime(self, vk_api, sender, message):
await self.vk_bot.send_message(sender, 'Total uptime: {} seconds'.format(round(time.time() - self.vk_bot.start_time)))
|
import time
class Plugin:
def __init__(self, vk_bot):
self.vk_bot = vk_bot
self.vk_bot.add_command('uptime', self.uptime)
async def uptime(self, vk_api, sender, message):
await self.vk_bot.send_message(sender,
'Total uptime: {} seconds'.format(round(time.time() - self.vk_bot.start_time)))
|
Fix code string length (PEP8)
|
Fix code string length (PEP8)
|
Python
|
mit
|
roman901/vk_bot
|
---
+++
@@ -7,4 +7,5 @@
self.vk_bot.add_command('uptime', self.uptime)
async def uptime(self, vk_api, sender, message):
- await self.vk_bot.send_message(sender, 'Total uptime: {} seconds'.format(round(time.time() - self.vk_bot.start_time)))
+ await self.vk_bot.send_message(sender,
+ 'Total uptime: {} seconds'.format(round(time.time() - self.vk_bot.start_time)))
|
e26434ee69545b8c16b62ebd78e5bec0c95d579a
|
lib/rapidsms/webui/urls.py
|
lib/rapidsms/webui/urls.py
|
#!/usr/bin/env python
# vim: ai ts=4 sts=4 et sw=4
import os
urlpatterns = []
# load the rapidsms configuration
from rapidsms.config import Config
conf = Config(os.environ["RAPIDSMS_INI"])
# iterate each of the active rapidsms apps (from the ini),
# and (attempt to) import the urls.py from each. it's okay
# if this fails, since not all apps have a webui
for rs_app in conf["rapidsms"]["apps"]:
try:
package_name = "apps.%s.urls" % (rs_app["type"])
module = __import__(package_name, {}, {}, ["urlpatterns"])
urlpatterns += module.urlpatterns
except Exception, e:
continue
|
#!/usr/bin/env python
# vim: ai ts=4 sts=4 et sw=4
import os, sys
urlpatterns = []
loaded = []
# load the rapidsms configuration
from rapidsms.config import Config
conf = Config(os.environ["RAPIDSMS_INI"])
# iterate each of the active rapidsms apps (from the ini),
# and (attempt to) import the urls.py from each. it's okay
# if this fails, since not all apps have a webui
for rs_app in conf["rapidsms"]["apps"]:
try:
package_name = "apps.%s.urls" % (rs_app["type"])
module = __import__(package_name, {}, {}, ["urlpatterns"])
urlpatterns += module.urlpatterns
loaded += [rs_app["type"]]
except Exception, e:
continue
print >>sys.stderr, "Loaded url patterns from %s" % ", ".join(loaded)
|
Print a list of which URLs got loaded. This doesn't help that much when trying to debug errors that keep URLs from getting loaded. But it's a start.
|
Print a list of which URLs got loaded. This doesn't help that much when
trying to debug errors that keep URLs from getting loaded. But it's a
start.
|
Python
|
bsd-3-clause
|
rapidsms/rapidsms-legacy,rapidsms/rapidsms-legacy,rapidsms/rapidsms-legacy
|
---
+++
@@ -1,16 +1,14 @@
#!/usr/bin/env python
# vim: ai ts=4 sts=4 et sw=4
-import os
-
+import os, sys
urlpatterns = []
-
+loaded = []
# load the rapidsms configuration
from rapidsms.config import Config
conf = Config(os.environ["RAPIDSMS_INI"])
-
# iterate each of the active rapidsms apps (from the ini),
# and (attempt to) import the urls.py from each. it's okay
@@ -21,6 +19,8 @@
package_name = "apps.%s.urls" % (rs_app["type"])
module = __import__(package_name, {}, {}, ["urlpatterns"])
urlpatterns += module.urlpatterns
+ loaded += [rs_app["type"]]
except Exception, e:
continue
+print >>sys.stderr, "Loaded url patterns from %s" % ", ".join(loaded)
|
21e02ee04ef359abfcaacf35375176b296ae7ea1
|
tests/__init__.py
|
tests/__init__.py
|
from unittest import TestCase
from redash import settings, db, app
import redash.models
# TODO: this isn't pretty...
settings.DATABASE_CONFIG = {
'name': 'circle_test',
'engine': 'peewee.PostgresqlDatabase',
'threadlocals': True
}
app.config['DATABASE'] = settings.DATABASE_CONFIG
db.load_database()
for model in redash.models.all_models:
model._meta.database = db.database
class BaseTestCase(TestCase):
def setUp(self):
redash.models.create_db(True, True)
def tearDown(self):
db.close_db(None)
redash.models.create_db(False, True)
|
import logging
from unittest import TestCase
from redash import settings, db, app
import redash.models
# TODO: this isn't pretty...
settings.DATABASE_CONFIG = {
'name': 'circle_test',
'engine': 'peewee.PostgresqlDatabase',
'threadlocals': True
}
app.config['DATABASE'] = settings.DATABASE_CONFIG
db.load_database()
logging.getLogger('peewee').setLevel(logging.INFO)
for model in redash.models.all_models:
model._meta.database = db.database
class BaseTestCase(TestCase):
def setUp(self):
redash.models.create_db(True, True)
def tearDown(self):
db.close_db(None)
redash.models.create_db(False, True)
|
Reduce Peewee's logging level to INFO in tests.
|
Reduce Peewee's logging level to INFO in tests.
|
Python
|
bsd-2-clause
|
rockwotj/redash,crowdworks/redash,pubnative/redash,stefanseifert/redash,pubnative/redash,akariv/redash,amino-data/redash,jmvasquez/redashtest,ninneko/redash,useabode/redash,stefanseifert/redash,crowdworks/redash,pubnative/redash,stefanseifert/redash,stefanseifert/redash,getredash/redash,amino-data/redash,M32Media/redash,guaguadev/redash,akariv/redash,useabode/redash,rockwotj/redash,ninneko/redash,jmvasquez/redashtest,M32Media/redash,imsally/redash,guaguadev/redash,easytaxibr/redash,denisov-vlad/redash,guaguadev/redash,ninneko/redash,moritz9/redash,getredash/redash,chriszs/redash,easytaxibr/redash,hudl/redash,ninneko/redash,akariv/redash,jmvasquez/redashtest,jmvasquez/redashtest,imsally/redash,useabode/redash,crowdworks/redash,denisov-vlad/redash,ninneko/redash,getredash/redash,amino-data/redash,alexanderlz/redash,moritz9/redash,denisov-vlad/redash,akariv/redash,denisov-vlad/redash,EverlyWell/redash,chriszs/redash,denisov-vlad/redash,guaguadev/redash,pubnative/redash,EverlyWell/redash,44px/redash,chriszs/redash,M32Media/redash,hudl/redash,amino-data/redash,moritz9/redash,imsally/redash,vishesh92/redash,hudl/redash,44px/redash,vishesh92/redash,alexanderlz/redash,44px/redash,stefanseifert/redash,moritz9/redash,easytaxibr/redash,guaguadev/redash,jmvasquez/redashtest,pubnative/redash,useabode/redash,rockwotj/redash,44px/redash,EverlyWell/redash,easytaxibr/redash,vishesh92/redash,vishesh92/redash,rockwotj/redash,akariv/redash,EverlyWell/redash,chriszs/redash,alexanderlz/redash,crowdworks/redash,easytaxibr/redash,hudl/redash,getredash/redash,imsally/redash,alexanderlz/redash,M32Media/redash,getredash/redash
|
---
+++
@@ -1,3 +1,4 @@
+import logging
from unittest import TestCase
from redash import settings, db, app
import redash.models
@@ -11,6 +12,8 @@
app.config['DATABASE'] = settings.DATABASE_CONFIG
db.load_database()
+logging.getLogger('peewee').setLevel(logging.INFO)
+
for model in redash.models.all_models:
model._meta.database = db.database
|
48af7d169bac32898763af671f3a30170b85d2cd
|
tests/__main__.py
|
tests/__main__.py
|
import unittest
if __name__ == '__main__':
all_tests = unittest.TestLoader().discover('./', pattern='*_tests.py')
unittest.TextTestRunner().run(all_tests)
|
import sys
import unittest
if __name__ == '__main__':
all_tests = unittest.TestLoader().discover('./', pattern='*_tests.py')
ret = unittest.TextTestRunner().run(all_tests)
sys.exit(not ret.wasSuccessful())
|
Fix an issue when unit tests always return 0 status.
|
Fix an issue when unit tests always return 0 status.
|
Python
|
mit
|
sergeymironov0001/twitch-chat-bot
|
---
+++
@@ -1,5 +1,7 @@
+import sys
import unittest
if __name__ == '__main__':
all_tests = unittest.TestLoader().discover('./', pattern='*_tests.py')
- unittest.TextTestRunner().run(all_tests)
+ ret = unittest.TextTestRunner().run(all_tests)
+ sys.exit(not ret.wasSuccessful())
|
9dc35ebafb3e33c3736c8d58a8cb2353695ddedb
|
tests/settings.py
|
tests/settings.py
|
DATABASES = {"default": {"ENGINE": "django.db.backends.sqlite3"}}
SECRET_KEY = "secrekey"
INSTALLED_APPS = ["phonenumber_field", "tests"]
|
DATABASES = {"default": {"ENGINE": "django.db.backends.sqlite3"}}
DEFAULT_AUTO_FIELD = "django.db.models.BigAutoField"
SECRET_KEY = "secrekey"
INSTALLED_APPS = ["phonenumber_field", "tests"]
|
Set DEFAULT_AUTO_FIELD for the test project
|
Set DEFAULT_AUTO_FIELD for the test project
https://docs.djangoproject.com/en/dev/releases/3.2/#customizing-type-of-auto-created-primary-keys
Avoid warnings on Django master:
```
tests.TestModelPhoneNU: (models.W042) Auto-created primary key used when not defining a primary key type, by default
'django.db.models.AutoField'.
HINT: Configure the DEFAULT_AUTO_FIELD setting or the AppConfig.default_auto_field attribute to point to a s
ubclass of AutoField, e.g. 'django.db.models.BigAutoField
```
|
Python
|
mit
|
stefanfoulis/django-phonenumber-field
|
---
+++
@@ -1,4 +1,5 @@
DATABASES = {"default": {"ENGINE": "django.db.backends.sqlite3"}}
+DEFAULT_AUTO_FIELD = "django.db.models.BigAutoField"
SECRET_KEY = "secrekey"
|
25478444e1ec5b4b1c9f811fea7fe0b401f14514
|
lingcod/bookmarks/forms.py
|
lingcod/bookmarks/forms.py
|
from lingcod.features.forms import FeatureForm
from lingcod.bookmarks.models import Bookmark
from django import forms
class BookmarkForm(FeatureForm):
name = forms.CharField(label='Bookmark Name')
latitude = forms.FloatField(widget=forms.HiddenInput())
longitude = forms.FloatField(widget=forms.HiddenInput())
altitude = forms.FloatField(widget=forms.HiddenInput())
heading = forms.FloatField(widget=forms.HiddenInput())
tilt = forms.FloatField(widget=forms.HiddenInput())
roll = forms.FloatField(widget=forms.HiddenInput())
altitudeMode = forms.FloatField(widget=forms.HiddenInput())
publicstate = forms.CharField(widget=forms.HiddenInput())
ip = forms.CharField(widget=forms.HiddenInput())
class Meta(FeatureForm.Meta):
model = Bookmark
|
from lingcod.features.forms import FeatureForm
from lingcod.bookmarks.models import Bookmark
from django import forms
class BookmarkForm(FeatureForm):
name = forms.CharField(label='Bookmark Name')
latitude = forms.FloatField(widget=forms.HiddenInput())
longitude = forms.FloatField(widget=forms.HiddenInput())
altitude = forms.FloatField(widget=forms.HiddenInput())
heading = forms.FloatField(widget=forms.HiddenInput())
tilt = forms.FloatField(widget=forms.HiddenInput())
roll = forms.FloatField(widget=forms.HiddenInput())
altitudeMode = forms.FloatField(widget=forms.HiddenInput())
publicstate = forms.CharField(widget=forms.HiddenInput())
ip = forms.CharField(widget=forms.HiddenInput(), required=False)
class Meta(FeatureForm.Meta):
model = Bookmark
|
Allow IP to be blank in form
|
Allow IP to be blank in form
|
Python
|
bsd-3-clause
|
Alwnikrotikz/marinemap,google-code-export/marinemap,google-code-export/marinemap,Alwnikrotikz/marinemap,google-code-export/marinemap,google-code-export/marinemap,Alwnikrotikz/marinemap,Alwnikrotikz/marinemap
|
---
+++
@@ -12,6 +12,6 @@
roll = forms.FloatField(widget=forms.HiddenInput())
altitudeMode = forms.FloatField(widget=forms.HiddenInput())
publicstate = forms.CharField(widget=forms.HiddenInput())
- ip = forms.CharField(widget=forms.HiddenInput())
+ ip = forms.CharField(widget=forms.HiddenInput(), required=False)
class Meta(FeatureForm.Meta):
model = Bookmark
|
9217bfc6bab0d152e33d9fda60218c404b61d064
|
cmd2/__init__.py
|
cmd2/__init__.py
|
#
# -*- coding: utf-8 -*-
from .cmd2 import __version__, Cmd, CmdResult, Statement, categorize
from .cmd2 import with_argument_list, with_argparser, with_argparser_and_unknown_args, with_category
|
#
# -*- coding: utf-8 -*-
from .cmd2 import __version__, Cmd, CmdResult, Statement, EmptyStatement, categorize
from .cmd2 import with_argument_list, with_argparser, with_argparser_and_unknown_args, with_category
|
Add EmptyStatement exception to default imports
|
Add EmptyStatement exception to default imports
|
Python
|
mit
|
python-cmd2/cmd2,python-cmd2/cmd2
|
---
+++
@@ -1,4 +1,4 @@
#
# -*- coding: utf-8 -*-
-from .cmd2 import __version__, Cmd, CmdResult, Statement, categorize
+from .cmd2 import __version__, Cmd, CmdResult, Statement, EmptyStatement, categorize
from .cmd2 import with_argument_list, with_argparser, with_argparser_and_unknown_args, with_category
|
e2126518957d0e3e360af3f80b1657bde9053b23
|
capstone/game/players/alphabeta.py
|
capstone/game/players/alphabeta.py
|
import random
from ..player import Player
from ..utils import utility
class AlphaBeta(Player):
name = 'Alpha-Beta'
def __init__(self, eval_func=utility, max_depth=1000):
self._eval = eval_func
self._max_depth = max_depth
def __str__(self):
return self.name
def __repr__(self):
return self.name
def _ab(self, game, cur_depth, alpha, beta):
if game.is_over() or cur_depth == self._max_depth:
return None, self._eval(game, game.cur_player())
best_move = None
best_score = -100000000
for move in game.legal_moves():
_, score = self._ab(game=game.copy().make_move(move),
cur_depth=cur_depth + 1,
alpha=-beta,
beta=-max(alpha, best_score))
score = -score
if score > best_score:
best_score = score
best_move = move
if best_score >= beta:
return best_move, best_score
return best_move, best_score
##########
# Player #
##########
def choose_move(self, game):
move, _ = self._ab(game, cur_depth=0, alpha=-100000000, beta=100000000)
return move
|
import random
import numpy as np
from ..player import Player
from ..utils import utility
class AlphaBeta(Player):
name = 'Alpha-Beta'
def __init__(self, eval_func=utility, max_depth=np.inf):
self._eval = eval_func
self._max_depth = max_depth
def __str__(self):
return self.name
def __repr__(self):
return self.name
def _ab(self, game, cur_depth, alpha, beta):
if game.is_over() or cur_depth == self._max_depth:
return None, self._eval(game, game.cur_player())
best_move = None
best_score = -np.inf
for move in game.legal_moves():
_, score = self._ab(game=game.copy().make_move(move),
cur_depth=cur_depth + 1,
alpha=-beta,
beta=-max(alpha, best_score))
score = -score
if score > best_score:
best_score = score
best_move = move
if best_score >= beta:
return best_move, best_score
return best_move, best_score
##########
# Player #
##########
def choose_move(self, game):
move, _ = self._ab(game, cur_depth=0, alpha=-np.inf, beta=np.inf)
return move
|
Use np.inf for max/min limit values
|
Use np.inf for max/min limit values
|
Python
|
mit
|
davidrobles/mlnd-capstone-code
|
---
+++
@@ -1,4 +1,5 @@
import random
+import numpy as np
from ..player import Player
from ..utils import utility
@@ -7,7 +8,7 @@
name = 'Alpha-Beta'
- def __init__(self, eval_func=utility, max_depth=1000):
+ def __init__(self, eval_func=utility, max_depth=np.inf):
self._eval = eval_func
self._max_depth = max_depth
@@ -21,7 +22,7 @@
if game.is_over() or cur_depth == self._max_depth:
return None, self._eval(game, game.cur_player())
best_move = None
- best_score = -100000000
+ best_score = -np.inf
for move in game.legal_moves():
_, score = self._ab(game=game.copy().make_move(move),
cur_depth=cur_depth + 1,
@@ -40,5 +41,5 @@
##########
def choose_move(self, game):
- move, _ = self._ab(game, cur_depth=0, alpha=-100000000, beta=100000000)
+ move, _ = self._ab(game, cur_depth=0, alpha=-np.inf, beta=np.inf)
return move
|
a17c2ce30f30d0441b1475457b0bc9d04da9f143
|
coil/__init__.py
|
coil/__init__.py
|
"""Coil: A Configuration Library."""
__version__ = "0.2.2"
|
"""Coil: A Configuration Library."""
__version__ = "0.3.0"
from coil.parser import Parser
def parse_file(file_name):
"""Open and parse a coil file.
Returns the root Struct.
"""
coil = open(file_name)
return Parser(coil, file_name).root()
def parse(string):
"""Parse a coil string.
Returns the root Struct.
"""
return Parser(string.splitlines()).root()
|
Add helpers for parsing files and strings
|
Add helpers for parsing files and strings
|
Python
|
mit
|
tectronics/coil,marineam/coil,kovacsbalu/coil,kovacsbalu/coil,marineam/coil,tectronics/coil
|
---
+++
@@ -1,3 +1,20 @@
"""Coil: A Configuration Library."""
-__version__ = "0.2.2"
+__version__ = "0.3.0"
+
+from coil.parser import Parser
+
+def parse_file(file_name):
+ """Open and parse a coil file.
+
+ Returns the root Struct.
+ """
+ coil = open(file_name)
+ return Parser(coil, file_name).root()
+
+def parse(string):
+ """Parse a coil string.
+
+ Returns the root Struct.
+ """
+ return Parser(string.splitlines()).root()
|
f664609d579e7b709945756def90092f0814998e
|
libpb/__init__.py
|
libpb/__init__.py
|
"""FreeBSD port building infrastructure."""
from __future__ import absolute_import
from . import event
def stop(kill=False, kill_clean=False):
"""Stop building ports and cleanup."""
from os import killpg
from signal import SIGTERM, SIGKILL
from .builder import builders
from .env import cpus, flags
from .queue import attr_queue, clean_queue, queues
if flags["no_op"]:
raise SystemExit(254)
flags["mode"] = "clean"
# Stop all queues
attr_queue.load = 0
for queue in queues:
queue.load = 0
# Make cleaning go a bit faster
if kill_clean:
clean_queue.load = 0
return
else:
clean_queue.load = cpus
# Wait for all active ports to finish so that they may be cleaned
active = set()
for queue in queues:
for job in queue.active:
port = job.port
active.add(port)
port.stage_completed.connect(lambda x: x.clean())
# Clean all other outstanding ports
for builder in builders:
for port in builder.ports:
if port not in active:
port.clean()
|
"""FreeBSD port building infrastructure."""
from __future__ import absolute_import
from . import event
def stop(kill=False, kill_clean=False):
"""Stop building ports and cleanup."""
from os import kill, killpg
from signal import SIGTERM, SIGKILL
from .builder import builders
from .env import cpus, flags
from .queue import attr_queue, clean_queue, queues
if flags["no_op"]:
raise SystemExit(254)
flags["mode"] = "clean"
kill_queues = (attr_queue,) + queues
if kill_clean:
kill_queues += (clean_queue,)
# Kill all active jobs
for queue in kill_queues:
for pid in (job.pid for job in queue.active if job.pid):
try:
if kill:
killpg(pid, SIGKILL)
else:
kill(pid, SIGTERM)
except OSError:
pass
# Stop all queues
attr_queue.load = 0
for queue in queues:
queue.load = 0
# Make cleaning go a bit faster
if kill_clean:
clean_queue.load = 0
return
else:
clean_queue.load = cpus
# Wait for all active ports to finish so that they may be cleaned
active = set()
for queue in queues:
for job in queue.active:
port = job.port
active.add(port)
port.stage_completed.connect(lambda x: x.clean())
# Clean all other outstanding ports
for builder in builders:
for port in builder.ports:
if port not in active:
port.clean()
|
Send SIGTERM and SIGKILL to child processes.
|
Send SIGTERM and SIGKILL to child processes.
With the removal of subprocess there was no way to known what were the
subprocesses, however after the introduction of Jobs tracking the PIDs
it is now possible. Use those PIDs.
|
Python
|
bsd-2-clause
|
DragonSA/portbuilder,DragonSA/portbuilder
|
---
+++
@@ -6,7 +6,7 @@
def stop(kill=False, kill_clean=False):
"""Stop building ports and cleanup."""
- from os import killpg
+ from os import kill, killpg
from signal import SIGTERM, SIGKILL
from .builder import builders
from .env import cpus, flags
@@ -16,6 +16,21 @@
raise SystemExit(254)
flags["mode"] = "clean"
+
+ kill_queues = (attr_queue,) + queues
+ if kill_clean:
+ kill_queues += (clean_queue,)
+
+ # Kill all active jobs
+ for queue in kill_queues:
+ for pid in (job.pid for job in queue.active if job.pid):
+ try:
+ if kill:
+ killpg(pid, SIGKILL)
+ else:
+ kill(pid, SIGTERM)
+ except OSError:
+ pass
# Stop all queues
attr_queue.load = 0
|
605339144c61c4860f1dc7dec5fc5a0ff959600f
|
company/forms.py
|
company/forms.py
|
from django import forms
from . import models
from pola.forms import (CommitDescriptionMixin,
FormHorizontalMixin, SaveButtonMixin,
ReadOnlyFieldsMixin)
class CompanyForm(ReadOnlyFieldsMixin, SaveButtonMixin, FormHorizontalMixin,
CommitDescriptionMixin, forms.ModelForm):
readonly_fields = [
'name'
]
class Meta:
model = models.Company
fields = [
'nip',
'name',
'official_name',
'address',
'plCapital',
'plCapital_notes',
'plTaxes',
'plTaxes_notes',
'plRnD',
'plRnD_notes',
'plWorkers',
'plWorkers_notes',
'plBrand',
'plBrand_notes',
'verified',
]
|
from django import forms
from . import models
from pola.forms import (CommitDescriptionMixin,
FormHorizontalMixin, SaveButtonMixin,
ReadOnlyFieldsMixin)
class CompanyForm(ReadOnlyFieldsMixin, SaveButtonMixin, FormHorizontalMixin,
CommitDescriptionMixin, forms.ModelForm):
readonly_fields = [
'name'
]
class Meta:
model = models.Company
fields = [
'nip',
'name',
'official_name',
'common_name',
'address',
'plCapital',
'plCapital_notes',
'plTaxes',
'plTaxes_notes',
'plRnD',
'plRnD_notes',
'plWorkers',
'plWorkers_notes',
'plBrand',
'plBrand_notes',
'verified',
]
|
Add 'common_name' to company's form
|
Add 'common_name' to company's form
|
Python
|
bsd-3-clause
|
KlubJagiellonski/pola-backend,KlubJagiellonski/pola-backend,KlubJagiellonski/pola-backend,KlubJagiellonski/pola-backend
|
---
+++
@@ -18,6 +18,7 @@
'nip',
'name',
'official_name',
+ 'common_name',
'address',
'plCapital',
'plCapital_notes',
|
559fa4bf1982de6dd4a8943939b535972731bd08
|
comrade/core/context_processors.py
|
comrade/core/context_processors.py
|
from django.conf import settings
from django.contrib.sites.models import Site
from settings import DeploymentType
def default(request):
context = {}
context['DEPLOYMENT'] = settings.DEPLOYMENT
context['site'] = Site.objects.get_current()
if settings.DEPLOYMENT != DeploymentType.PRODUCTION:
context['GIT_COMMIT'] = settings.GIT_COMMIT
context['site_email'] = settings.CONTACT_EMAIL
return context
def ssl_media(request):
if request.is_secure():
ssl_media_url = settings.MEDIA_URL.replace('http://','https://')
else:
ssl_media_url = settings.MEDIA_URL
return {'MEDIA_URL': ssl_media_url}
|
from django.conf import settings
from django.contrib.sites.models import Site
from settings import DeploymentType
def default(request):
context = {}
context['DEPLOYMENT'] = settings.DEPLOYMENT
context['current_site'] = Site.objects.get_current()
if settings.DEPLOYMENT != DeploymentType.PRODUCTION:
context['GIT_COMMIT'] = settings.GIT_COMMIT
context['site_email'] = settings.CONTACT_EMAIL
if request.is_secure():
context['protocol'] = 'https://'
else:
context['protocol'] = 'http://'
context['current_site_url'] = (context['protocol'] +
context['current_site'].domain)
return context
def ssl_media(request):
if request.is_secure():
ssl_media_url = settings.MEDIA_URL.replace('http://','https://')
else:
ssl_media_url = settings.MEDIA_URL
return {'MEDIA_URL': ssl_media_url}
|
Add full base URL for site to default context.
|
Add full base URL for site to default context.
|
Python
|
mit
|
bueda/django-comrade
|
---
+++
@@ -5,10 +5,16 @@
def default(request):
context = {}
context['DEPLOYMENT'] = settings.DEPLOYMENT
- context['site'] = Site.objects.get_current()
+ context['current_site'] = Site.objects.get_current()
if settings.DEPLOYMENT != DeploymentType.PRODUCTION:
context['GIT_COMMIT'] = settings.GIT_COMMIT
context['site_email'] = settings.CONTACT_EMAIL
+ if request.is_secure():
+ context['protocol'] = 'https://'
+ else:
+ context['protocol'] = 'http://'
+ context['current_site_url'] = (context['protocol'] +
+ context['current_site'].domain)
return context
def ssl_media(request):
|
1b97aa2dae43a8988802ca532a3200f444f85db3
|
markups/common.py
|
markups/common.py
|
# This file is part of python-markups module
# License: BSD
# Copyright: (C) Dmitry Shachnev, 2012
import os.path
# Some common constants and functions
(LANGUAGE_HOME_PAGE, MODULE_HOME_PAGE, SYNTAX_DOCUMENTATION) = range(3)
CONFIGURATION_DIR = (os.environ.get('XDG_CONFIG_HOME') or
os.path.expanduser('~/.config'))
MATHJAX_LOCAL_URL = 'file:///usr/share/javascript/mathjax/MathJax.js'
MATHJAX_WEB_URL = 'http://cdn.mathjax.org/mathjax/latest/MathJax.js'
def get_pygments_stylesheet(selector):
try:
from pygments.formatters import HtmlFormatter
except ImportError:
return ''
else:
return HtmlFormatter().get_style_defs(selector) + '\n'
def get_mathjax_url(webenv):
if os.path.exists(MATHJAX_LOCAL_URL[7:]) and not webenv:
return MATHJAX_LOCAL_URL
else:
return MATHJAX_WEB_URL
|
# This file is part of python-markups module
# License: BSD
# Copyright: (C) Dmitry Shachnev, 2012
import os.path
# Some common constants and functions
(LANGUAGE_HOME_PAGE, MODULE_HOME_PAGE, SYNTAX_DOCUMENTATION) = range(3)
CONFIGURATION_DIR = (os.environ.get('XDG_CONFIG_HOME') or
os.path.expanduser('~/.config'))
MATHJAX_LOCAL_URL = 'file:///usr/share/javascript/mathjax/MathJax.js'
MATHJAX_WEB_URL = 'http://cdn.mathjax.org/mathjax/latest/MathJax.js'
PYGMENTS_STYLE = 'default'
def get_pygments_stylesheet(selector, style=None):
try:
from pygments.formatters import HtmlFormatter
except ImportError:
return ''
else:
return HtmlFormatter(style=(style or PYGMENTS_STYLE)).get_style_defs(selector) + '\n'
def get_mathjax_url(webenv):
if os.path.exists(MATHJAX_LOCAL_URL[7:]) and not webenv:
return MATHJAX_LOCAL_URL
else:
return MATHJAX_WEB_URL
|
Add initial support for pygments styles
|
Add initial support for pygments styles
|
Python
|
bsd-3-clause
|
retext-project/pymarkups,mitya57/pymarkups
|
---
+++
@@ -11,13 +11,15 @@
MATHJAX_LOCAL_URL = 'file:///usr/share/javascript/mathjax/MathJax.js'
MATHJAX_WEB_URL = 'http://cdn.mathjax.org/mathjax/latest/MathJax.js'
-def get_pygments_stylesheet(selector):
+PYGMENTS_STYLE = 'default'
+
+def get_pygments_stylesheet(selector, style=None):
try:
from pygments.formatters import HtmlFormatter
except ImportError:
return ''
else:
- return HtmlFormatter().get_style_defs(selector) + '\n'
+ return HtmlFormatter(style=(style or PYGMENTS_STYLE)).get_style_defs(selector) + '\n'
def get_mathjax_url(webenv):
if os.path.exists(MATHJAX_LOCAL_URL[7:]) and not webenv:
|
05cb698d45ce4e33e2f4bfdc38f9633083a284a7
|
test_project/project_specific/generic_channel_example.py
|
test_project/project_specific/generic_channel_example.py
|
import autocomplete_light
from models import Contact, Address
class MyGenericChannel(autocomplete_light.GenericChannelBase):
def get_querysets(self):
return {
Contact: Contact.objects.all(),
Address: Address.objects.all(),
}
def order_results(self, results):
if results.model == Address:
return results.order_by('street')
elif results.model == Contact:
return results.order_by('name')
autocomplete_light.register(MyGenericChannel)
|
import autocomplete_light
from models import Contact, Address
class MyGenericChannel(autocomplete_light.GenericChannelBase):
def get_querysets(self):
return {
Contact: Contact.objects.all(),
Address: Address.objects.all(),
}
def order_results(self, results):
if results.model == Address:
return results.order_by('street')
elif results.model == Contact:
return results.order_by('name')
def query_filter(self, results):
q = self.request.GET.get('q', None)
if q:
if results.model == Address:
results = results.filter(street__icontains=q)
elif results.model == Contact:
results = results.filter(name__icontains=q)
return results
autocomplete_light.register(MyGenericChannel)
|
Implement query_filter for MyGenericChannel, because it should search by something other than search_name in the case of Address
|
Implement query_filter for MyGenericChannel, because it should search by something other than search_name in the case of Address
|
Python
|
mit
|
Eraldo/django-autocomplete-light,spookylukey/django-autocomplete-light,Perkville/django-autocomplete-light,jonashaag/django-autocomplete-light,yourlabs/django-autocomplete-light,shubhamdipt/django-autocomplete-light,Perkville/django-autocomplete-light,Visgean/django-autocomplete-light,dsanders11/django-autocomplete-light,Visgean/django-autocomplete-light,jonashaag/django-autocomplete-light,Eraldo/django-autocomplete-light,spookylukey/django-autocomplete-light,dsanders11/django-autocomplete-light,yourlabs/django-autocomplete-light,blueyed/django-autocomplete-light,shubhamdipt/django-autocomplete-light,luzfcb/django-autocomplete-light,dsanders11/django-autocomplete-light,Visgean/django-autocomplete-light,shubhamdipt/django-autocomplete-light,yourlabs/django-autocomplete-light,jonashaag/django-autocomplete-light,Perkville/django-autocomplete-light,Perkville/django-autocomplete-light,yourlabs/django-autocomplete-light,shubhamdipt/django-autocomplete-light,luzfcb/django-autocomplete-light,Visgean/django-autocomplete-light,luzfcb/django-autocomplete-light,luzfcb/django-autocomplete-light,Eraldo/django-autocomplete-light,blueyed/django-autocomplete-light,blueyed/django-autocomplete-light,dsanders11/django-autocomplete-light,spookylukey/django-autocomplete-light,Eraldo/django-autocomplete-light
|
---
+++
@@ -15,4 +15,15 @@
elif results.model == Contact:
return results.order_by('name')
+ def query_filter(self, results):
+ q = self.request.GET.get('q', None)
+
+ if q:
+ if results.model == Address:
+ results = results.filter(street__icontains=q)
+ elif results.model == Contact:
+ results = results.filter(name__icontains=q)
+
+ return results
+
autocomplete_light.register(MyGenericChannel)
|
d8c75104acb68ca648c5a3b30d6791775272e5c1
|
authentic2/idp/idp_openid/admin.py
|
authentic2/idp/idp_openid/admin.py
|
# -*- coding: utf-8 -*-
# vim: set ts=4 sw=4 : */
from django.contrib import admin
from models import TrustedRoot, Association, Nonce
admin.site.register(TrustedRoot)
admin.site.register(Association)
admin.site.register(Nonce)
|
# -*- coding: utf-8 -*-
from django.contrib import admin
from models import TrustedRoot, Association, Nonce
admin.site.register(TrustedRoot)
admin.site.register(Association)
admin.site.register(Nonce)
|
Remove vim instruction in prologue.
|
[idp/idp_openid] Remove vim instruction in prologue.
|
Python
|
agpl-3.0
|
incuna/authentic,incuna/authentic,adieu/authentic2,incuna/authentic,adieu/authentic2,BryceLohr/authentic,BryceLohr/authentic,incuna/authentic,BryceLohr/authentic,pu239ppy/authentic2,BryceLohr/authentic,pu239ppy/authentic2,pu239ppy/authentic2,adieu/authentic2,adieu/authentic2,incuna/authentic,pu239ppy/authentic2
|
---
+++
@@ -1,5 +1,4 @@
# -*- coding: utf-8 -*-
-# vim: set ts=4 sw=4 : */
from django.contrib import admin
from models import TrustedRoot, Association, Nonce
|
15c596ea224fb0a624404d425a2580ab3503807d
|
bluebottle/utils/middleware.py
|
bluebottle/utils/middleware.py
|
from importlib import import_module
from django.conf import settings
from django.contrib.sessions import middleware
from django.db import connection
from django.utils import translation
from tenant_extras.middleware import tenant_translation
from bluebottle.utils.models import get_languages, get_default_language
class SubDomainSessionMiddleware(middleware.SessionMiddleware):
def process_request(self, request):
engine = import_module(settings.SESSION_ENGINE)
session_key = request.COOKIES.get(settings.SESSION_COOKIE_NAME, None)
if session_key is None:
# Look for old cookie in request for auth purposes.
session_key = request.COOKIES.get('sessionid', None)
request.session = engine.SessionStore(session_key)
class APILanguageMiddleware(middleware.SessionMiddleware):
def process_request(self, request):
if request.path.startswith('/api'):
try:
language = request.META['HTTP_X_APPLICATION_LANGUAGE']
if language not in [lang.code for lang in get_languages()]:
language = get_default_language()
except KeyError:
language = get_default_language()
translation.activate(language)
translation._trans._active.value = tenant_translation(
language, connection.tenant.client_name
)
request.LANGUAGE_CODE = translation.get_language()
|
from importlib import import_module
from django.conf import settings
from django.contrib.sessions import middleware
from django.db import connection
from django.utils import translation
from tenant_extras.middleware import tenant_translation
from bluebottle.utils.models import get_languages, get_default_language
class SubDomainSessionMiddleware(middleware.SessionMiddleware):
def process_request(self, request):
engine = import_module(settings.SESSION_ENGINE)
session_key = request.COOKIES.get(settings.SESSION_COOKIE_NAME, None)
if session_key is None:
# Look for old cookie in request for auth purposes.
session_key = request.COOKIES.get('sessionid', None)
request.session = engine.SessionStore(session_key)
class APILanguageMiddleware(middleware.SessionMiddleware):
def process_request(self, request):
if request.path.startswith('/api'):
try:
language = request.META['HTTP_X_APPLICATION_LANGUAGE']
if language not in [lang.full_code for lang in get_languages()]:
language = get_default_language()
except KeyError:
language = get_default_language()
translation.activate(language)
translation._trans._active.value = tenant_translation(
language, connection.tenant.client_name
)
request.LANGUAGE_CODE = translation.get_language()
|
Use full code to check if api language actually exists
|
Use full code to check if api language actually exists
|
Python
|
bsd-3-clause
|
onepercentclub/bluebottle,onepercentclub/bluebottle,onepercentclub/bluebottle,onepercentclub/bluebottle,onepercentclub/bluebottle
|
---
+++
@@ -26,7 +26,7 @@
if request.path.startswith('/api'):
try:
language = request.META['HTTP_X_APPLICATION_LANGUAGE']
- if language not in [lang.code for lang in get_languages()]:
+ if language not in [lang.full_code for lang in get_languages()]:
language = get_default_language()
except KeyError:
language = get_default_language()
|
0ad7be235135303cb9d902df2a89b17da8aac918
|
syntacticframes_project/syntacticframes/migrations/0012_auto_20150220_1836.py
|
syntacticframes_project/syntacticframes/migrations/0012_auto_20150220_1836.py
|
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
from distutils.version import LooseVersion
from django.db import models, migrations
def set_position_value_for_levin_classes(apps, schema_editor):
i = 0
LevinClass = apps.get_model('syntacticframes', 'LevinClass')
levin_class_list = sorted(LevinClass.objects.all(), key=lambda l: int(l.number))
for levin_class in levin_class_list:
verbnet_classes = sorted(
levin_class.verbnetclass_set.all(),
key=lambda v: LooseVersion(v.name.split('-')[1]))
for v in verbnet_classes:
v.position = i
v.save()
i += 10
class Migration(migrations.Migration):
dependencies = [
('syntacticframes', '0011_auto_20150121_1600'),
]
operations = [
migrations.AlterModelOptions(
name='verbnetclass',
options={'ordering': ['position']},
),
migrations.AddField(
model_name='verbnetclass',
name='position',
field=models.PositiveSmallIntegerField(default=0),
preserve_default=False,
),
migrations.RunPython(set_position_value_for_levin_classes),
]
|
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
from distutils.version import LooseVersion
from django.db import models, migrations
def set_position_value_for_levin_classes(apps, schema_editor):
i = 0
LevinClass = apps.get_model('syntacticframes', 'LevinClass')
levin_class_list = sorted(LevinClass.objects.all(), key=lambda l: int(l.number))
for levin_class in levin_class_list:
verbnet_classes = sorted(
levin_class.verbnetclass_set.all(),
key=lambda v: LooseVersion(v.name.split('-')[1]))
for v in verbnet_classes:
v.position = i
v.save()
i += 10
class Migration(migrations.Migration):
dependencies = [
('syntacticframes', '0011_auto_20150121_1600'),
]
operations = [
migrations.AlterModelOptions(
name='verbnetclass',
options={'ordering': ['position']},
),
migrations.AddField(
model_name='verbnetclass',
name='position',
field=models.PositiveSmallIntegerField(default=0),
preserve_default=False,
),
migrations.RunPython(
code=set_position_value_for_levin_classes,
# Nothing to reverse since this is about a new field
reverse_code=lambda apps, schema_editor: None),
]
|
Make the 0012 migration reversible
|
Make the 0012 migration reversible
|
Python
|
mit
|
aymara/verbenet-editor,aymara/verbenet-editor,aymara/verbenet-editor
|
---
+++
@@ -36,5 +36,8 @@
field=models.PositiveSmallIntegerField(default=0),
preserve_default=False,
),
- migrations.RunPython(set_position_value_for_levin_classes),
+ migrations.RunPython(
+ code=set_position_value_for_levin_classes,
+ # Nothing to reverse since this is about a new field
+ reverse_code=lambda apps, schema_editor: None),
]
|
31d973be4f50e2e26d8ff65dd86d96d55e6b6250
|
contrib/zmq/chromecast-snoop.py
|
contrib/zmq/chromecast-snoop.py
|
#!/usr/bin/env python3
class Listener(object):
def __init__(self):
import zmqclient
self.artist = None
self.title = None
pub = zmqclient.pub()
def new_media_status(self, status):
if status.artist != self.artist or status.title != self.title:
self.artist = status.artist
self.title = status.title
print("{} -- {}".format(status.artist, status.title))
if status.artist != None or status.title != None:
pub.send(b"CHROMECAST", zmq.SNDMORE)
pub.send_json(status.media_metadata)
if __name__ == '__main__':
import pychromecast, time
cast = pychromecast.get_chromecast()
cast.wait()
print("Connected to {}".format(cast.device.friendly_name))
zmq = Listener()
cast.media_controller.register_status_listener(zmq)
while True:
time.sleep(30)
|
#!/usr/bin/env python3
class Listener(object):
def __init__(self):
import zmqclient
self.artist = None
self.title = None
self.pub = zmqclient.pub()
def new_media_status(self, status):
if status.artist != self.artist or status.title != self.title:
self.artist = status.artist
self.title = status.title
print("{} -- {}".format(status.artist, status.title))
if status.artist != None or status.title != None:
try:
self.pub.send(b"CHROMECAST", 2) # 2 == zmq.SNDMORE FIXME
self.pub.send_json(status.media_metadata)
except Exception as e:
print(e)
if __name__ == '__main__':
import pychromecast, time
cast = pychromecast.get_chromecast()
cast.wait()
print("Connected to {}".format(cast.device.friendly_name))
zmq = Listener()
cast.media_controller.register_status_listener(zmq)
while True:
time.sleep(30)
|
Print exceptions since pychromecast just ignores them
|
Print exceptions since pychromecast just ignores them
|
Python
|
apache-2.0
|
hackeriet/nfcd,hackeriet/nfcd,hackeriet/pyhackeriet,hackeriet/pyhackeriet,hackeriet/nfcd,hackeriet/pyhackeriet
|
---
+++
@@ -6,7 +6,7 @@
self.artist = None
self.title = None
- pub = zmqclient.pub()
+ self.pub = zmqclient.pub()
def new_media_status(self, status):
if status.artist != self.artist or status.title != self.title:
@@ -14,8 +14,11 @@
self.title = status.title
print("{} -- {}".format(status.artist, status.title))
if status.artist != None or status.title != None:
- pub.send(b"CHROMECAST", zmq.SNDMORE)
- pub.send_json(status.media_metadata)
+ try:
+ self.pub.send(b"CHROMECAST", 2) # 2 == zmq.SNDMORE FIXME
+ self.pub.send_json(status.media_metadata)
+ except Exception as e:
+ print(e)
if __name__ == '__main__':
import pychromecast, time
|
fdbf12d560a18724b5293e4cee8aa3a03a48c90b
|
tcconfig/tcdel.py
|
tcconfig/tcdel.py
|
#!/usr/bin/env python
# encoding: utf-8
"""
.. codeauthor:: Tsuyoshi Hombashi <gogogo.vm@gmail.com>
"""
from __future__ import absolute_import
import sys
import logbook
import subprocrunner
import tcconfig
from .traffic_control import TrafficControl
from ._argparse_wrapper import ArgparseWrapper
from ._common import verify_network_interface
from ._error import NetworkInterfaceNotFoundError
handler = logbook.StderrHandler()
handler.push_application()
def parse_option():
parser = ArgparseWrapper(tcconfig.VERSION)
group = parser.parser.add_argument_group("Traffic Control")
group.add_argument(
"--device", required=True,
help="network device name (e.g. eth0)")
return parser.parser.parse_args()
def main():
options = parse_option()
logger = logbook.Logger("tcdel")
logger.level = options.log_level
subprocrunner.logger.level = options.log_level
if options.quiet:
subprocrunner.logger.disable()
else:
subprocrunner.logger.enable()
subprocrunner.Which("tc").verify()
try:
verify_network_interface(options.device)
except NetworkInterfaceNotFoundError as e:
logger.error(e)
return 1
tc = TrafficControl(options.device)
return tc.delete_tc()
if __name__ == '__main__':
sys.exit(main())
|
#!/usr/bin/env python
# encoding: utf-8
"""
.. codeauthor:: Tsuyoshi Hombashi <gogogo.vm@gmail.com>
"""
from __future__ import absolute_import
import sys
import logbook
import subprocrunner
import tcconfig
from .traffic_control import TrafficControl
from ._argparse_wrapper import ArgparseWrapper
from ._common import verify_network_interface
from ._error import NetworkInterfaceNotFoundError
handler = logbook.StderrHandler()
handler.push_application()
def parse_option():
parser = ArgparseWrapper(tcconfig.VERSION)
group = parser.parser.add_argument_group("Traffic Control")
group.add_argument(
"--device", required=True,
help="network device name (e.g. eth0)")
return parser.parser.parse_args()
def main():
options = parse_option()
logger = logbook.Logger("tcdel")
logger.level = options.log_level
subprocrunner.logger.level = options.log_level
if options.quiet:
subprocrunner.logger.disable()
else:
subprocrunner.logger.enable()
subprocrunner.Which("tc").verify()
try:
verify_network_interface(options.device)
except NetworkInterfaceNotFoundError as e:
logger.error(e)
return 1
tc = TrafficControl(options.device)
try:
return tc.delete_tc()
except NetworkInterfaceNotFoundError as e:
logger.debug(e)
return 0
return 1
if __name__ == '__main__':
sys.exit(main())
|
Modify to handle exception properly
|
Modify to handle exception properly
|
Python
|
mit
|
thombashi/tcconfig,thombashi/tcconfig
|
---
+++
@@ -54,7 +54,13 @@
tc = TrafficControl(options.device)
- return tc.delete_tc()
+ try:
+ return tc.delete_tc()
+ except NetworkInterfaceNotFoundError as e:
+ logger.debug(e)
+ return 0
+
+ return 1
if __name__ == '__main__':
|
f6dd7d0ca966856325adc50f4c5ca2cc48dda0a5
|
cogbot/cog_bot_server_state.py
|
cogbot/cog_bot_server_state.py
|
import json
import logging
import typing
from datetime import datetime
import discord
from cogbot.types import ServerId, ChannelId
log = logging.getLogger(__name__)
class CogBotServerState:
def __init__(self, bot, server: discord.Server, log_channel: ChannelId = None):
self.bot = bot
self.server: discord.Server = server
# resolve log channel
self.log_channel: discord.Channel = None
if log_channel:
self.log_channel = self.bot.get_channel(log_channel)
if not self.log_channel:
log.warning(
f"[{self.server}] Failed to resolve log channel <{log_channel}>"
)
async def mod_log(
self, member: discord.Member, content: str, channel: discord.Channel = None
):
if self.log_channel:
now = datetime.utcnow()
quote_name = f"{member.display_name} ({member.name}#{member.discriminator})"
em = discord.Embed(description=content, timestamp=now)
em.set_author(name=quote_name, icon_url=member.avatar_url)
em.set_footer(text=f"#{channel}" if channel else None)
await self.bot.send_message(self.log_channel, embed=em)
|
import json
import logging
import typing
from datetime import datetime
import discord
from cogbot.types import ServerId, ChannelId
log = logging.getLogger(__name__)
class CogBotServerState:
def __init__(self, bot, server: discord.Server, log_channel: ChannelId = None):
self.bot = bot
self.server: discord.Server = server
# resolve log channel
self.log_channel: discord.Channel = None
if log_channel:
self.log_channel = self.bot.get_channel(log_channel)
if not self.log_channel:
log.warning(
f"[{self.server}] Failed to resolve log channel <{log_channel}>"
)
async def mod_log(
self, member: discord.Member, content: str, channel: discord.Channel = None
):
if self.log_channel:
now = datetime.utcnow()
quote_name = f"{member.display_name} ({member.name}#{member.discriminator})"
em = discord.Embed(description=content, timestamp=now)
em.set_author(name=quote_name, icon_url=member.avatar_url)
if channel:
em.set_footer(text=f"#{channel}")
await self.bot.send_message(self.log_channel, embed=em)
|
Fix mod log optional channel
|
Fix mod log optional channel
|
Python
|
mit
|
Arcensoth/cogbot
|
---
+++
@@ -33,5 +33,6 @@
quote_name = f"{member.display_name} ({member.name}#{member.discriminator})"
em = discord.Embed(description=content, timestamp=now)
em.set_author(name=quote_name, icon_url=member.avatar_url)
- em.set_footer(text=f"#{channel}" if channel else None)
+ if channel:
+ em.set_footer(text=f"#{channel}")
await self.bot.send_message(self.log_channel, embed=em)
|
67c671260858cc2c3d3041188cebda63cac1c4eb
|
prequ/__init__.py
|
prequ/__init__.py
|
import pkg_resources
try:
__version__ = pkg_resources.get_distribution(__name__).version
except pkg_resources.DistributionNotFound:
__version__ = None
|
import pkg_resources
try:
__version__ = pkg_resources.get_distribution(__name__).version
except pkg_resources.DistributionNotFound: # pragma: no cover
__version__ = None
|
Add "no cover" pragma to version setting code
|
Add "no cover" pragma to version setting code
|
Python
|
bsd-2-clause
|
suutari-ai/prequ,suutari/prequ,suutari/prequ
|
---
+++
@@ -2,5 +2,5 @@
try:
__version__ = pkg_resources.get_distribution(__name__).version
-except pkg_resources.DistributionNotFound:
+except pkg_resources.DistributionNotFound: # pragma: no cover
__version__ = None
|
6171b8111359cc54a4af2c3444ce0e0e2db5ba80
|
froide/helper/context_processors.py
|
froide/helper/context_processors.py
|
from django.conf import settings
def froide(request):
return {"froide": settings.FROIDE_CONFIG}
def site_settings(request):
return {"SITE_NAME": settings.SITE_NAME,
"SITE_URL": settings.SITE_URL,
"FROIDE_DRYRUN": settings.FROIDE_DRYRUN,
"FROIDE_DRYRUN_DOMAIN": settings.FROIDE_DRYRUN_DOMAIN}
|
from django.conf import settings
def froide(request):
return {"froide": settings.FROIDE_CONFIG}
def site_settings(request):
return {"SITE_NAME": settings.SITE_NAME,
"SITE_URL": settings.SITE_URL,
"FROIDE_DRYRUN": settings.FROIDE_DRYRUN,
"FROIDE_DRYRUN_DOMAIN": settings.FROIDE_DRYRUN_DOMAIN,
"LANGUAGE_CODE": settings.LANGUAGE_CODE}
|
Add Froide Dry Run Domain and Language Code to context_processor
|
Add Froide Dry Run Domain and Language Code to context_processor
|
Python
|
mit
|
okfse/froide,ryankanno/froide,fin/froide,LilithWittmann/froide,okfse/froide,fin/froide,ryankanno/froide,LilithWittmann/froide,catcosmo/froide,CodeforHawaii/froide,CodeforHawaii/froide,stefanw/froide,catcosmo/froide,ryankanno/froide,catcosmo/froide,catcosmo/froide,okfse/froide,fin/froide,ryankanno/froide,stefanw/froide,ryankanno/froide,CodeforHawaii/froide,LilithWittmann/froide,okfse/froide,stefanw/froide,fin/froide,stefanw/froide,catcosmo/froide,stefanw/froide,LilithWittmann/froide,okfse/froide,CodeforHawaii/froide,LilithWittmann/froide,CodeforHawaii/froide
|
---
+++
@@ -7,4 +7,5 @@
return {"SITE_NAME": settings.SITE_NAME,
"SITE_URL": settings.SITE_URL,
"FROIDE_DRYRUN": settings.FROIDE_DRYRUN,
- "FROIDE_DRYRUN_DOMAIN": settings.FROIDE_DRYRUN_DOMAIN}
+ "FROIDE_DRYRUN_DOMAIN": settings.FROIDE_DRYRUN_DOMAIN,
+ "LANGUAGE_CODE": settings.LANGUAGE_CODE}
|
50451c69d337228c2016851258ff7249bf906440
|
profiling/plot.py
|
profiling/plot.py
|
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
import glob
import re
import numpy as np
import matplotlib.pyplot as plt
csv_files = glob.glob('*.csv')
fig = plt.figure()
ax = fig.add_subplot(111)
colors = iter(plt.cm.rainbow(np.linspace(0,1,len(csv_files))))
p = re.compile(r'profiling_(.*?)_(.*?)\.csv')
ms_to_s = 1.0 / 1000.0
for csv_file in csv_files:
data = np.genfromtxt(csv_file, delimiter=',', skip_header=1).transpose()
j = data[0]
N = data[1]
avg = data[2]
std = data[3]
m = p.search(csv_file)
name = m.group(2)
name = name.replace('_', ' ')
ax.errorbar(N, avg*ms_to_s, yerr=std*ms_to_s,
label=name, color=next(colors), marker='o')
ax.grid(True)
ax.set_xlabel('N')
ax.set_ylabel('Timing [s]')
ax.set_xscale('log')
ax.set_yscale('log')
ax.legend(loc='best')
plt.show()
|
#!/usr/bin/env python3
# -*- coding: utf-8 -*-
import glob
import re
import os
import sys
import numpy as np
import matplotlib.pyplot as plt
if len(sys.argv) == 1:
print('Usage: plot.py path/to/build/profiling')
sys.exit(0)
csv_files = glob.glob(os.path.join(sys.argv[1], '*.csv'))
fig = plt.figure()
ax = fig.add_subplot(111)
colors = iter(plt.cm.rainbow(np.linspace(0,1,len(csv_files))))
p = re.compile(r'profiling_(.*?)_(.*?)\.csv')
ms_to_s = 1.0 / 1000.0
for csv_file in csv_files:
data = np.genfromtxt(csv_file, delimiter=',', skip_header=1).transpose()
j = data[0]
N = data[1]
avg = data[2]
std = data[3]
m = p.search(csv_file)
name = m.group(2)
name = name.replace('_', ' ')
ax.errorbar(N, avg*ms_to_s, yerr=std*ms_to_s,
label=name, color=next(colors), marker='o')
ax.grid(True)
ax.set_xlabel('N')
ax.set_ylabel('Timing [s]')
ax.set_xscale('log')
ax.set_yscale('log')
ax.legend(loc='best')
plt.show()
|
Use path given as argument
|
Use path given as argument
|
Python
|
bsd-3-clause
|
nbigaouette/sorting,nbigaouette/sorting,nbigaouette/sorting,nbigaouette/sorting
|
---
+++
@@ -3,12 +3,18 @@
import glob
import re
+import os
+import sys
import numpy as np
import matplotlib.pyplot as plt
-csv_files = glob.glob('*.csv')
+if len(sys.argv) == 1:
+ print('Usage: plot.py path/to/build/profiling')
+ sys.exit(0)
+
+csv_files = glob.glob(os.path.join(sys.argv[1], '*.csv'))
fig = plt.figure()
ax = fig.add_subplot(111)
|
d8ce56feada64d287306d7f439ec12a42acda0d6
|
bot.py
|
bot.py
|
#!/usr/bin/env python3
# -*- coding: utf8 -*-
import tweepy
consumer_key = ""
consumer_secret = ""
access_token = ""
access_token_secret = ""
auth = tweepy.OAuthHandler(consumer_key, consumer_secret)
auth.set_access_token(access_token, access_token_secret)
api = tweepy.API(auth)
userid = str(input("Please input id who you want fav attack\n"))
count = input("input number you want to fav!\n")
fav = api.user_timeline(id = userid, count = count)
try:
for status in fav:
api.create_favorite(status.id_str)
except tweepy.error.TweepError as e:
if e.args[0][0]['code'] == 139:
print("You have already favorited this status! \n")
else:
print(e.reason)
finally:
print("Done!")
|
#!/usr/bin/env python3
# -*- coding: utf8 -*-
import tweepy
consumer_key = ""
consumer_secret = ""
access_token = ""
access_token_secret = ""
auth = tweepy.OAuthHandler(consumer_key, consumer_secret)
auth.set_access_token(access_token, access_token_secret)
api = tweepy.API(auth)
def getdata():
userid = str(input("Please input id who you want fav attack\n"))
count = input("input number you want to fav!\n")
fav = api.user_timeline(id = userid, count = count)
def main():
getdata()
try:
for status in fav:
api.create_favorite(status.id_str)
except tweepy.error.TweepError as e:
if e.args[0][0]['code'] == 139:
print("You have already favorited this status! \n")
else:
print(e.reason)
finally:
print("Done!")
if __name__ == "__main__":
main()
|
Make it more complex (((
|
Make it more complex (((
|
Python
|
mit
|
zhangyubaka/tweepy_favbot
|
---
+++
@@ -15,18 +15,23 @@
api = tweepy.API(auth)
-userid = str(input("Please input id who you want fav attack\n"))
-count = input("input number you want to fav!\n")
+def getdata():
+ userid = str(input("Please input id who you want fav attack\n"))
+ count = input("input number you want to fav!\n")
+ fav = api.user_timeline(id = userid, count = count)
-fav = api.user_timeline(id = userid, count = count)
+def main():
+ getdata()
+ try:
+ for status in fav:
+ api.create_favorite(status.id_str)
+ except tweepy.error.TweepError as e:
+ if e.args[0][0]['code'] == 139:
+ print("You have already favorited this status! \n")
+ else:
+ print(e.reason)
+ finally:
+ print("Done!")
-try:
- for status in fav:
- api.create_favorite(status.id_str)
-except tweepy.error.TweepError as e:
- if e.args[0][0]['code'] == 139:
- print("You have already favorited this status! \n")
- else:
- print(e.reason)
-finally:
- print("Done!")
+if __name__ == "__main__":
+ main()
|
f511af4fc89a170914a86de1704e8e842ffd6b6d
|
test/test_configuration.py
|
test/test_configuration.py
|
#!/usr/bin/env python
"""Test coordinate classes."""
import sys
try:
import unittest2 as unittest # Python 2.6
except ImportError:
import unittest
import heatmap as hm
class Tests(unittest.TestCase):
# To remove Python 3's
# "DeprecationWarning: Please use assertRaisesRegex instead"
if sys.version_info[0] == 2:
assertRaisesRegex = unittest.TestCase.assertRaisesRegexp
def test_basic(self):
'''Test Configuration class.'''
# Act
config = hm.Configuration(use_defaults=True)
# Assert
self.assertEqual(config.margin, 0)
self.assertEqual(config.frequency, 1)
def test_fill_missing_no_input(self):
'''Test Configuration class.'''
# Arrange
config = hm.Configuration(use_defaults=True)
# Act / Assert
with self.assertRaisesRegex(ValueError, "no input specified"):
config.fill_missing()
if __name__ == '__main__':
unittest.main()
|
#!/usr/bin/env python
"""Test coordinate classes."""
import sys
try:
import unittest2 as unittest # Python 2.6
except ImportError:
import unittest
ROOT_DIR = os.path.split(os.path.abspath(os.path.dirname(__file__)))[0]
sys.path.append(ROOT_DIR)
import heatmap as hm
class Tests(unittest.TestCase):
# To remove Python 3's
# "DeprecationWarning: Please use assertRaisesRegex instead"
if sys.version_info[0] == 2:
assertRaisesRegex = unittest.TestCase.assertRaisesRegexp
def test_basic(self):
'''Test Configuration class.'''
# Act
config = hm.Configuration(use_defaults=True)
# Assert
self.assertEqual(config.margin, 0)
self.assertEqual(config.frequency, 1)
def test_fill_missing_no_input(self):
'''Test Configuration class.'''
# Arrange
config = hm.Configuration(use_defaults=True)
# Act / Assert
with self.assertRaisesRegex(ValueError, "no input specified"):
config.fill_missing()
if __name__ == '__main__':
unittest.main()
|
Update sys.path to import heatmap
|
Update sys.path to import heatmap
|
Python
|
agpl-3.0
|
hugovk/heatmap,hugovk/heatmap,sethoscope/heatmap,sethoscope/heatmap
|
---
+++
@@ -8,6 +8,8 @@
except ImportError:
import unittest
+ROOT_DIR = os.path.split(os.path.abspath(os.path.dirname(__file__)))[0]
+sys.path.append(ROOT_DIR)
import heatmap as hm
@@ -20,7 +22,6 @@
def test_basic(self):
'''Test Configuration class.'''
-
# Act
config = hm.Configuration(use_defaults=True)
|
c3883ad36139db4592c259a94e7b254702ec2f00
|
run.py
|
run.py
|
import glpi.server;glpi.server.app.run(debug = True)
|
import glpi.server;glpi.server.app.run(debug = True, host='0.0.0.0')
|
Add option to rlisten on al interfaces
|
Add option to rlisten on al interfaces
|
Python
|
agpl-3.0
|
ddurieux/poc_glpi
|
---
+++
@@ -1 +1 @@
-import glpi.server;glpi.server.app.run(debug = True)
+import glpi.server;glpi.server.app.run(debug = True, host='0.0.0.0')
|
f62ec2304bb42b0bf0eba121c1b88adb35cec6d3
|
run.py
|
run.py
|
#!/usr/bin/env python
import os
import argparse
def run():
os.system("gunicorn server:app --reload --config gunicorn_config.py")
def deploy():
os.system("git push heroku master")
def main():
parser = argparse.ArgumentParser()
parser.add_argument('--deploy', action="store_true", required=False)
args = parser.parse_args()
if args.deploy:
deploy()
else:
run()
if __name__ == '__main__':
main()
|
#!/usr/bin/env python
import os
import argparse
def run():
os.system("gunicorn server:app --reload --config gunicorn_config.py")
def deploy():
os.system("git push dokku master")
def main():
parser = argparse.ArgumentParser()
parser.add_argument('--deploy', action="store_true", required=False)
args = parser.parse_args()
if args.deploy:
deploy()
else:
run()
if __name__ == '__main__':
main()
|
Switch to dokku for deployment.
|
Switch to dokku for deployment.
|
Python
|
mit
|
EmilStenstrom/json-tagger,EmilStenstrom/json-tagger,EmilStenstrom/json-tagger,EmilStenstrom/json-tagger,EmilStenstrom/json-tagger
|
---
+++
@@ -6,7 +6,7 @@
os.system("gunicorn server:app --reload --config gunicorn_config.py")
def deploy():
- os.system("git push heroku master")
+ os.system("git push dokku master")
def main():
parser = argparse.ArgumentParser()
|
a24faf712d8dfba0f6ac9fc295807552dca37ae9
|
custom/inddex/reports/utils.py
|
custom/inddex/reports/utils.py
|
from corehq.apps.reports.datatables import DataTablesColumn, DataTablesHeader
from corehq.apps.reports.generic import GenericTabularReport
from corehq.apps.reports.standard import CustomProjectReport, DatespanMixin
class MultiTabularReport(DatespanMixin, CustomProjectReport, GenericTabularReport):
report_template_path = 'inddex/multi_report.html'
exportable = True
export_only = False
@property
def data_providers(self):
# data providers should supply a title, slug, headers, and rows
return []
@property
def report_context(self):
context = {
'name': self.name,
'export_only': self.export_only
}
if not self.needs_filters:
context['data_providers'] = [{
'title': data_provider.title,
'slug': data_provider.slug,
'headers': DataTablesHeader(
*(DataTablesColumn(header) for header in data_provider.headers),
),
'rows': data_provider.rows,
} for data_provider in self.data_providers]
return context
@property
def export_table(self):
return [
[dp.slug, [dp.headers] + dp.rows]
for dp in self.data_providers
]
|
from itertools import chain
from corehq.apps.reports.datatables import DataTablesColumn, DataTablesHeader
from corehq.apps.reports.generic import GenericTabularReport
from corehq.apps.reports.standard import CustomProjectReport, DatespanMixin
class MultiTabularReport(DatespanMixin, CustomProjectReport, GenericTabularReport):
report_template_path = 'inddex/multi_report.html'
exportable = True
exportable_all = True
export_only = False
@property
def data_providers(self):
# data providers should supply a title, slug, headers, and rows
return []
@property
def report_context(self):
context = {
'name': self.name,
'export_only': self.export_only
}
if not self.needs_filters:
context['data_providers'] = [{
'title': data_provider.title,
'slug': data_provider.slug,
'headers': DataTablesHeader(
*(DataTablesColumn(header) for header in data_provider.headers),
),
'rows': data_provider.rows,
} for data_provider in self.data_providers]
return context
@property
def export_table(self):
return [
[dp.slug, chain([dp.headers], dp.rows)]
for dp in self.data_providers
]
|
Move export to a background process
|
Move export to a background process
|
Python
|
bsd-3-clause
|
dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq
|
---
+++
@@ -1,3 +1,5 @@
+from itertools import chain
+
from corehq.apps.reports.datatables import DataTablesColumn, DataTablesHeader
from corehq.apps.reports.generic import GenericTabularReport
from corehq.apps.reports.standard import CustomProjectReport, DatespanMixin
@@ -6,6 +8,7 @@
class MultiTabularReport(DatespanMixin, CustomProjectReport, GenericTabularReport):
report_template_path = 'inddex/multi_report.html'
exportable = True
+ exportable_all = True
export_only = False
@property
@@ -33,6 +36,6 @@
@property
def export_table(self):
return [
- [dp.slug, [dp.headers] + dp.rows]
+ [dp.slug, chain([dp.headers], dp.rows)]
for dp in self.data_providers
]
|
46c33ca68c1124fb06c4ba62306cb00ba61d7e5c
|
tests/__init__.py
|
tests/__init__.py
|
from flexmock import flexmock
from flask.ext.storage import MockStorage
from flask_uploads import init
class TestCase(object):
added_objects = []
committed_objects = []
created_objects = []
deleted_objects = []
def setup_method(self, method, resizer=None):
init(db_mock, MockStorage, resizer)
self.db = db_mock
self.Storage = MockStorage
self.storage = MockStorage()
self.resizer = resizer
def teardown_method(self, method):
# Empty the stacks.
TestCase.added_objects[:] = []
TestCase.committed_objects[:] = []
TestCase.created_objects[:] = []
TestCase.deleted_objects[:] = []
class MockModel(object):
def __init__(self, **kw):
TestCase.created_objects.append(self)
for key, val in kw.iteritems():
setattr(self, key, val)
db_mock = flexmock(
Column=lambda *a, **kw: ('column', a, kw),
Integer=('integer', [], {}),
Unicode=lambda *a, **kw: ('unicode', a, kw),
Model=MockModel,
session=flexmock(
add=TestCase.added_objects.append,
commit=lambda: TestCase.committed_objects.extend(
TestCase.added_objects + TestCase.deleted_objects
),
delete=TestCase.deleted_objects.append,
),
)
|
from flexmock import flexmock
from flask.ext.storage import MockStorage
from flask_uploads import init
class TestCase(object):
added_objects = []
committed_objects = []
created_objects = []
deleted_objects = []
def setup_method(self, method, resizer=None):
init(db_mock, MockStorage, resizer)
self.db = db_mock
self.Storage = MockStorage
self.storage = MockStorage()
self.resizer = resizer
def teardown_method(self, method):
# Empty the stacks.
TestCase.added_objects[:] = []
TestCase.committed_objects[:] = []
TestCase.created_objects[:] = []
TestCase.deleted_objects[:] = []
class MockModel(object):
def __init__(self, **kw):
TestCase.created_objects.append(self)
for key, val in kw.iteritems():
setattr(self, key, val)
db_mock = flexmock(
Column=lambda *a, **kw: ('column', a, kw),
Integer=('integer', [], {}),
Unicode=lambda *a, **kw: ('unicode', a, kw),
Model=MockModel,
metadata=flexmock(tables={}),
session=flexmock(
add=TestCase.added_objects.append,
commit=lambda: TestCase.committed_objects.extend(
TestCase.added_objects + TestCase.deleted_objects
),
delete=TestCase.deleted_objects.append,
),
)
|
Add metadata.tables to mock db.
|
Add metadata.tables to mock db.
|
Python
|
mit
|
FelixLoether/flask-uploads,FelixLoether/flask-image-upload-thing
|
---
+++
@@ -36,6 +36,7 @@
Integer=('integer', [], {}),
Unicode=lambda *a, **kw: ('unicode', a, kw),
Model=MockModel,
+ metadata=flexmock(tables={}),
session=flexmock(
add=TestCase.added_objects.append,
commit=lambda: TestCase.committed_objects.extend(
|
3037c15a658ee05a4483c5b1c8c137a06e9d7707
|
tests/conftest.py
|
tests/conftest.py
|
'''
General-purpose fixtures for vdirsyncer's testsuite.
'''
import logging
import os
import click_log
from hypothesis import HealthCheck, Verbosity, settings
import pytest
@pytest.fixture(autouse=True)
def setup_logging():
click_log.basic_config('vdirsyncer').setLevel(logging.DEBUG)
try:
import pytest_benchmark
except ImportError:
@pytest.fixture
def benchmark():
return lambda x: x()
else:
del pytest_benchmark
settings.register_profile("ci", settings(
max_examples=1000,
verbosity=Verbosity.verbose,
suppress_health_check=[HealthCheck.too_slow],
))
settings.register_profile("deterministic", settings(
derandomize=True,
suppress_health_check=HealthCheck.all(),
))
if os.environ.get('DETERMINISTIC_TESTS', 'false').lower() == 'true':
settings.load_profile("deterministic")
elif os.environ.get('CI', 'false').lower() == 'true':
settings.load_profile("ci")
|
'''
General-purpose fixtures for vdirsyncer's testsuite.
'''
import logging
import os
import click_log
from hypothesis import HealthCheck, Verbosity, settings
import pytest
@pytest.fixture(autouse=True)
def setup_logging():
click_log.basic_config('vdirsyncer').setLevel(logging.DEBUG)
try:
import pytest_benchmark
except ImportError:
@pytest.fixture
def benchmark():
return lambda x: x()
else:
del pytest_benchmark
settings.register_profile("ci", settings(
max_examples=1000,
verbosity=Verbosity.verbose,
suppress_health_check=[HealthCheck.too_slow],
))
settings.register_profile("deterministic", settings(
derandomize=True,
suppress_health_check=HealthCheck.all(),
))
settings.register_profile("dev", settings(
suppress_health_check=[HealthCheck.too_slow]
))
if os.environ.get('DETERMINISTIC_TESTS', 'false').lower() == 'true':
settings.load_profile("deterministic")
elif os.environ.get('CI', 'false').lower() == 'true':
settings.load_profile("ci")
else:
settings.load_profile("dev")
|
Use hypothesis setting load_profile to setup health check
|
Use hypothesis setting load_profile to setup health check
Fixes #779
|
Python
|
mit
|
untitaker/vdirsyncer,untitaker/vdirsyncer,untitaker/vdirsyncer
|
---
+++
@@ -35,8 +35,13 @@
derandomize=True,
suppress_health_check=HealthCheck.all(),
))
+settings.register_profile("dev", settings(
+ suppress_health_check=[HealthCheck.too_slow]
+))
if os.environ.get('DETERMINISTIC_TESTS', 'false').lower() == 'true':
settings.load_profile("deterministic")
elif os.environ.get('CI', 'false').lower() == 'true':
settings.load_profile("ci")
+else:
+ settings.load_profile("dev")
|
24b78a4d510606563106da24d568d5fb79ddca2b
|
IPython/__main__.py
|
IPython/__main__.py
|
# encoding: utf-8
"""Terminal-based IPython entry point.
"""
#-----------------------------------------------------------------------------
# Copyright (c) 2012, IPython Development Team.
#
# Distributed under the terms of the Modified BSD License.
#
# The full license is in the file COPYING.txt, distributed with this software.
#-----------------------------------------------------------------------------
from IPython.terminal.ipapp import launch_new_instance
launch_new_instance()
|
# encoding: utf-8
"""Terminal-based IPython entry point.
"""
#-----------------------------------------------------------------------------
# Copyright (c) 2012, IPython Development Team.
#
# Distributed under the terms of the Modified BSD License.
#
# The full license is in the file COPYING.txt, distributed with this software.
#-----------------------------------------------------------------------------
from IPython import start_ipython
start_ipython()
|
Use new entry point for python -m IPython
|
Use new entry point for python -m IPython
|
Python
|
bsd-3-clause
|
ipython/ipython,ipython/ipython
|
---
+++
@@ -9,6 +9,6 @@
# The full license is in the file COPYING.txt, distributed with this software.
#-----------------------------------------------------------------------------
-from IPython.terminal.ipapp import launch_new_instance
+from IPython import start_ipython
-launch_new_instance()
+start_ipython()
|
8c8b668ba3684c3e756bf9fccafbd1bd8e1a7cfe
|
mediapipe/__init__.py
|
mediapipe/__init__.py
|
"""Copyright 2019 - 2020 The MediaPipe Authors.
Licensed under the Apache License, Version 2.0 (the "License");
you may not use this file except in compliance with the License.
You may obtain a copy of the License at
http://www.apache.org/licenses/LICENSE-2.0
Unless required by applicable law or agreed to in writing, software
distributed under the License is distributed on an "AS IS" BASIS,
WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
See the License for the specific language governing permissions and
limitations under the License.
"""
|
# Copyright 2019 - 2022 The MediaPipe Authors.
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
|
Fix comment for `mediapipe` license.
|
Fix comment for `mediapipe` license.
The `"""` comment indicates a public docstring for the module, and will end up in the generated docs. By using a "private" comment (`#`) we will not document the license as part of the API.
The Apache license is noted in the footer of generated docs, and this is sufficient.
PiperOrigin-RevId: 487688719
|
Python
|
apache-2.0
|
google/mediapipe,google/mediapipe,google/mediapipe,google/mediapipe,google/mediapipe,google/mediapipe,google/mediapipe,google/mediapipe
|
---
+++
@@ -1,14 +1,13 @@
-"""Copyright 2019 - 2020 The MediaPipe Authors.
-
-Licensed under the Apache License, Version 2.0 (the "License");
-you may not use this file except in compliance with the License.
-You may obtain a copy of the License at
-
- http://www.apache.org/licenses/LICENSE-2.0
-
-Unless required by applicable law or agreed to in writing, software
-distributed under the License is distributed on an "AS IS" BASIS,
-WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
-See the License for the specific language governing permissions and
-limitations under the License.
-"""
+# Copyright 2019 - 2022 The MediaPipe Authors.
+#
+# Licensed under the Apache License, Version 2.0 (the "License");
+# you may not use this file except in compliance with the License.
+# You may obtain a copy of the License at
+#
+# http://www.apache.org/licenses/LICENSE-2.0
+#
+# Unless required by applicable law or agreed to in writing, software
+# distributed under the License is distributed on an "AS IS" BASIS,
+# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+# See the License for the specific language governing permissions and
+# limitations under the License.
|
189847ffdca0264ddd6248faa9974ba35eaea373
|
tests/test_aur.py
|
tests/test_aur.py
|
# MIT licensed
# Copyright (c) 2013-2017 lilydjwg <lilydjwg@gmail.com>, et al.
from flaky import flaky
import pytest
pytestmark = pytest.mark.asyncio
@flaky(max_runs=5)
async def test_aur(get_version):
assert await get_version("asciidoc-fake", {"aur": None}) == "1.0-1"
@flaky(max_runs=5)
async def test_aur_strip_release(get_version):
assert await get_version("asciidoc-fake", {"aur": None, "strip-release": 1}) == "1.0"
|
# MIT licensed
# Copyright (c) 2013-2017 lilydjwg <lilydjwg@gmail.com>, et al.
from flaky import flaky
import pytest
pytestmark = pytest.mark.asyncio
@flaky
async def test_aur(get_version):
assert await get_version("asciidoc-fake", {"aur": None}) == "1.0-1"
@flaky
async def test_aur_strip_release(get_version):
assert await get_version("asciidoc-fake", {"aur": None, "strip-release": 1}) == "1.0"
|
Revert "make AUR tests more flaky"
|
Revert "make AUR tests more flaky"
This reverts commit 61df628bd8bc97acbed40a4af67b124c47584f5f.
It doesn't help :-(
|
Python
|
mit
|
lilydjwg/nvchecker
|
---
+++
@@ -5,10 +5,10 @@
import pytest
pytestmark = pytest.mark.asyncio
-@flaky(max_runs=5)
+@flaky
async def test_aur(get_version):
assert await get_version("asciidoc-fake", {"aur": None}) == "1.0-1"
-@flaky(max_runs=5)
+@flaky
async def test_aur_strip_release(get_version):
assert await get_version("asciidoc-fake", {"aur": None, "strip-release": 1}) == "1.0"
|
6d22cc47174139b56fad7d94696b08d9830a7ea4
|
lettuce_webdriver/tests/__init__.py
|
lettuce_webdriver/tests/__init__.py
|
from __future__ import print_function
import os
from contextlib import contextmanager
from selenium import webdriver
from aloe import around, world
here = os.path.dirname(__file__)
html_pages = os.path.join(here, 'html_pages')
@around.each_feature
@contextmanager
def with_browser(feature):
world.browser = webdriver.Firefox()
world.browser.get('')
yield
world.browser.quit()
delattr(world, 'browser')
@around.each_step
@contextmanager
def print_source(step):
try:
yield
except:
print(world.browser.page_source)
print(world.browser.get_screenshot_as_base64())
raise
|
from __future__ import print_function
import os
from contextlib import contextmanager
from selenium import webdriver
from aloe import around, world
here = os.path.dirname(__file__)
html_pages = os.path.join(here, 'html_pages')
@around.each_feature
@contextmanager
def with_browser(feature):
world.browser = webdriver.Firefox()
world.browser.get('')
yield
world.browser.quit()
delattr(world, 'browser')
@around.each_step
@contextmanager
def print_source(step):
try:
yield
except:
try:
step_container = step.scenario
except AttributeError:
step_container = step.background
print(step_container.feature.name)
print(step_container.name)
print(step.sentence)
print(world.browser.page_source)
print(world.browser.get_screenshot_as_base64())
raise
|
Print scenario/step names on failure
|
Print scenario/step names on failure
|
Python
|
mit
|
koterpillar/aloe_webdriver,infoxchange/aloe_webdriver,infoxchange/aloe_webdriver,aloetesting/aloe_webdriver,aloetesting/aloe_webdriver,aloetesting/aloe_webdriver,koterpillar/aloe_webdriver
|
---
+++
@@ -27,6 +27,14 @@
try:
yield
except:
+ try:
+ step_container = step.scenario
+ except AttributeError:
+ step_container = step.background
+
+ print(step_container.feature.name)
+ print(step_container.name)
+ print(step.sentence)
print(world.browser.page_source)
print(world.browser.get_screenshot_as_base64())
raise
|
73c1900a05fa3e4f68224f4e0d5dce2c08687254
|
opwen_email_server/backend/email_sender.py
|
opwen_email_server/backend/email_sender.py
|
from typing import Tuple
from opwen_email_server import azure_constants as constants
from opwen_email_server import config
from opwen_email_server.services.queue import AzureQueue
from opwen_email_server.services.sendgrid import SendgridEmailSender
QUEUE = AzureQueue(account=config.QUEUES_ACCOUNT, key=config.QUEUES_KEY,
name=constants.QUEUE_EMAIL_SEND)
EMAIL = SendgridEmailSender(key=config.EMAIL_SENDER_KEY)
def send(email: dict) -> Tuple[str, int]:
success = EMAIL.send_email(email)
if not success:
return 'error', 500
return 'sent', 200
if __name__ == '__main__':
from argparse import ArgumentParser
from json import loads
from uuid import uuid4
parser = ArgumentParser()
parser.add_argument('email')
args = parser.parse_args()
email = loads(args.email)
email.setdefault('_uid', str(uuid4()))
send(email)
|
from typing import Tuple
from opwen_email_server import azure_constants as constants
from opwen_email_server import config
from opwen_email_server.services.queue import AzureQueue
from opwen_email_server.services.sendgrid import SendgridEmailSender
QUEUE = AzureQueue(account=config.QUEUES_ACCOUNT, key=config.QUEUES_KEY,
name=constants.QUEUE_EMAIL_SEND)
EMAIL = SendgridEmailSender(key=config.EMAIL_SENDER_KEY)
def send(email: dict) -> Tuple[str, int]:
success = EMAIL.send_email(email)
if not success:
return 'error', 500
return 'sent', 200
if __name__ == '__main__':
from argparse import ArgumentParser
from argparse import FileType
from base64 import b64encode
from json import loads
from os.path import basename
from uuid import uuid4
parser = ArgumentParser()
parser.add_argument('email')
parser.add_argument('--attachment', type=FileType('rb'))
args = parser.parse_args()
email = loads(args.email)
email.setdefault('_uid', str(uuid4()))
if args.attachment:
email.setdefault('attachments', []).append({
'filename': basename(args.attachment.name),
'content': b64encode(args.attachment.read()).decode('ascii')
})
args.attachment.close()
send(email)
|
Add attachment support to email sender CLI
|
Add attachment support to email sender CLI
|
Python
|
apache-2.0
|
ascoderu/opwen-cloudserver,ascoderu/opwen-cloudserver
|
---
+++
@@ -22,14 +22,25 @@
if __name__ == '__main__':
from argparse import ArgumentParser
+ from argparse import FileType
+ from base64 import b64encode
from json import loads
+ from os.path import basename
from uuid import uuid4
parser = ArgumentParser()
parser.add_argument('email')
+ parser.add_argument('--attachment', type=FileType('rb'))
args = parser.parse_args()
email = loads(args.email)
email.setdefault('_uid', str(uuid4()))
+ if args.attachment:
+ email.setdefault('attachments', []).append({
+ 'filename': basename(args.attachment.name),
+ 'content': b64encode(args.attachment.read()).decode('ascii')
+ })
+ args.attachment.close()
+
send(email)
|
dd5d81263842066bf1faa3699c12da47f9453401
|
packages/flake8-config-4catalyzer/setup.py
|
packages/flake8-config-4catalyzer/setup.py
|
from setuptools import setup
setup(
name='flake8-config-4catalyzer',
version='0.2.1',
url='https://github.com/4Catalyzer/python/tree/packages/flake8-config-4catalyzer',
author="Alex Rothberg",
author_email='arothberg@4catalyzer.com',
license='MIT',
classifiers=(
'Development Status :: 3 - Alpha',
'Environment :: Console',
'Framework :: Flake8',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 3',
'Topic :: Software Development :: Libraries :: Python Modules',
'Topic :: Software Development :: Quality Assurance',
),
keywords='flake8',
install_requires=(
'flake8',
'flake8-commas',
'flake8-import-order',
),
extras_require={
':python_version>="3.5"': ('flake8-bugbear',),
},
)
|
from setuptools import setup
setup(
name='flake8-config-4catalyzer',
version='0.2.1',
url='https://github.com/4Catalyzer/python/tree/packages/flake8-config-4catalyzer',
author="Alex Rothberg",
author_email='arothberg@4catalyzer.com',
license='MIT',
classifiers=(
'Development Status :: 3 - Alpha',
'Environment :: Console',
'Framework :: Flake8',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 3',
'Topic :: Software Development :: Libraries :: Python Modules',
'Topic :: Software Development :: Quality Assurance',
),
keywords='flake8',
install_requires=(
'flake8',
'flake8-commas',
'flake8-debugger',
'flake8-import-order',
),
extras_require={
':python_version>="3.5"': ('flake8-bugbear',),
},
)
|
Add flake8-debugger to list of flake8 checks
|
Add flake8-debugger to list of flake8 checks
|
Python
|
mit
|
4Catalyzer/python
|
---
+++
@@ -25,6 +25,7 @@
install_requires=(
'flake8',
'flake8-commas',
+ 'flake8-debugger',
'flake8-import-order',
),
extras_require={
|
dcb1fc943ec4fe39bd752b1015ba11f6d8145c27
|
modules/status.py
|
modules/status.py
|
import discord
from modules.botModule import BotModule
from modules.help import *
import time
import datetime
class Status(BotModule):
name = 'status'
description = 'Allow for the assignment and removal of roles.'
help_text = 'Usage: `!status` shows information about this instance of scubot.'
trigger_string = '!status'
init_time = 0
def __init__(self):
init_time = time.time()
def uptime_convert(self,seconds):
minutes, seconds = divmod(seconds, 60)
hours, minutes = divmod(minutes, 60)
days, hours = divmod(hours, 24)
return days, hours, minutes, seconds
async def parse_command(self, message, client):
uptime = self.init_time - time.time()
uptime_string = self.uptime_convert(uptime)
uptime_string = [str(round(x,0))[:-2] for x in uptime_string]
uptime_string = uptime_string[0] + 'd ' + uptime_string[1] + 'h ' + uptime_string[2] + 'm ' + uptime_string[3] + 's'
print(self.loaded_modules)
module_string = ''
for botModule in self.loaded_modules:
module_string += botModule.name + ', '
module_string = module_string[:-2]
msg = '```\n Uptime: ' + uptime_string + '\n Loaded modules: ' + module_string + '\n```'
await client.send_message(message.channel, msg)
|
import discord
from modules.botModule import BotModule
from modules.help import *
import time
import datetime
class Status(BotModule):
name = 'status'
description = 'Allow for the assignment and removal of roles.'
help_text = 'Usage: `!status` shows information about this instance of scubot.'
trigger_string = '!status'
init_time = time.time()
def __init__(self):
init_time = time.time()
def uptime_convert(self,seconds):
minutes, seconds = divmod(seconds, 60)
hours, minutes = divmod(minutes, 60)
days, hours = divmod(hours, 24)
return days, hours, minutes, seconds
async def parse_command(self, message, client):
uptime = time.time() - self.init_time
uptime_string = self.uptime_convert(uptime)
uptime_string = [str(round(x,0))[:-2] for x in uptime_string]
uptime_string = uptime_string[0] + 'd ' + uptime_string[1] + 'h ' + uptime_string[2] + 'm ' + uptime_string[3] + 's'
module_string = ''
for botModule in self.loaded_modules:
module_string += botModule.name + ', '
module_string = module_string[:-2]
msg = '```\n Uptime: ' + uptime_string + '\n Loaded modules: ' + module_string + '\n```'
await client.send_message(message.channel, msg)
|
Fix uptime and uptime timing
|
Fix uptime and uptime timing
|
Python
|
mit
|
suclearnub/scubot
|
---
+++
@@ -13,7 +13,7 @@
trigger_string = '!status'
- init_time = 0
+ init_time = time.time()
def __init__(self):
init_time = time.time()
@@ -25,11 +25,10 @@
return days, hours, minutes, seconds
async def parse_command(self, message, client):
- uptime = self.init_time - time.time()
+ uptime = time.time() - self.init_time
uptime_string = self.uptime_convert(uptime)
uptime_string = [str(round(x,0))[:-2] for x in uptime_string]
uptime_string = uptime_string[0] + 'd ' + uptime_string[1] + 'h ' + uptime_string[2] + 'm ' + uptime_string[3] + 's'
- print(self.loaded_modules)
module_string = ''
for botModule in self.loaded_modules:
module_string += botModule.name + ', '
|
a9e80e81fe2e6ad1325047cb3045ab12640f984f
|
cla_backend/apps/cla_eventlog/management/commands/find_and_delete_old_cases.py
|
cla_backend/apps/cla_eventlog/management/commands/find_and_delete_old_cases.py
|
from django.core.management.base import BaseCommand
from dateutil.relativedelta import relativedelta
from legalaid.models import Case
from cla_butler.tasks import DeleteOldData
class FindAndDeleteCasesUsingCreationTime(DeleteOldData):
def get_eligible_cases(self):
self._setup()
two_years = self.now - relativedelta(years=2)
return Case.objects.filter(created__lte=two_years).exclude(log__created__gte=two_years)
class Command(BaseCommand):
help = (
"Find or delete cases that are 2 years old or over that were not deleted prior to the task command being fixed"
)
def handle(self, *args, **kwargs):
instance = FindAndDeleteCasesUsingCreationTime()
if len(args) == 0:
cases = instance.get_eligible_cases()
print(cases.count())
elif args[0] == "delete":
instance.run()
|
from django.core.management.base import BaseCommand
from dateutil.relativedelta import relativedelta
from legalaid.models import Case
from cla_butler.tasks import DeleteOldData
class FindAndDeleteCasesUsingCreationTime(DeleteOldData):
def get_eligible_cases(self):
self._setup()
two_years = self.now - relativedelta(years=2)
return Case.objects.filter(created__lte=two_years).exclude(log__created__gte=two_years)
class Command(BaseCommand):
help = (
"Find or delete cases that are 2 years old or over that were not deleted prior to the task command being fixed"
)
def handle(self, *args, **kwargs):
instance = FindAndDeleteCasesUsingCreationTime()
cases = instance.get_eligible_cases()
if len(args) == 0:
print(cases.count())
elif args[0] == "test_find":
return cases
elif args[0] == "delete":
instance.run()
|
Refactor code so we can use command in tests
|
Refactor code so we can use command in tests
|
Python
|
mit
|
ministryofjustice/cla_backend,ministryofjustice/cla_backend,ministryofjustice/cla_backend,ministryofjustice/cla_backend
|
---
+++
@@ -20,8 +20,10 @@
def handle(self, *args, **kwargs):
instance = FindAndDeleteCasesUsingCreationTime()
+ cases = instance.get_eligible_cases()
if len(args) == 0:
- cases = instance.get_eligible_cases()
print(cases.count())
+ elif args[0] == "test_find":
+ return cases
elif args[0] == "delete":
instance.run()
|
cb2a6c44fcff3453bbf20dbecbd807ee8e640e8e
|
database/queries/select_queries.py
|
database/queries/select_queries.py
|
SELECT_MOVIES_ORDERED_BY_RATING = '''
SELECT *
FROM MOVIE
ORDER BY RATING;
'''
SELECT_PROJECTION_FOR_MOVIE = '''
SELECT p.*, COUNT(r.ROW * r.COL)
FROM PROJECTION as p
LEFT JOIN RESERVATION as r
ON r.PROJECTION_ID = p.ID
WHERE p.MOVIE_ID = ?
GROUP BY p.ID
ORDER BY p.DATE;
'''
SELECT_PROJECTION_ORDERED_BY_DATE = '''
SELECT PROJECTION.*, COUNT(RESERVATION.ROW * RESERVATION.COL)
FROM PROJECTION
LEFT JOIN RESERVATION
ON RESERVATION.PROJECTION_ID = PROJECTION.ID
WHERE PROJECTION.MOVIE_ID = ? AND PROJECTION.DATE BETWEEN ? AND ?
GROUP BY PROJECTION.ID
ORDER BY PROJECTION.DATE;
'''
IS_USER_IN_USERS = '''
SELECT *
FROM USERS
WHERE USERNAME = ? and PASSWORD LIKE ?;
'''
SELECT_AVAILABLE_SEATS = '''
SELECT ROW, COL
FROM RESERVATIONS
WHERE PROJECTION_ID = ? AND ROW IS NOT NULL AND COL IS NOT NULL;
'''
SELECT_PROJECTION_BY_ID = '''
SELECT *
FROM PROJECTION as p
WHERE p.ID = ?
'''
SELECT_MOVIE_BY_ID = '''
SELECT *
FROM MOVIE as m
WHERE m.ID = ?
'''
|
SELECT_MOVIES_ORDERED_BY_RATING = '''
SELECT *
FROM MOVIE
ORDER BY RATING;
'''
SELECT_PROJECTION_FOR_MOVIE = '''
SELECT p.*, COUNT(r.ROW * r.COL)
FROM PROJECTION as p
LEFT JOIN RESERVATION as r
ON r.PROJECTION_ID = p.ID
WHERE p.MOVIE_ID = ?
GROUP BY p.ID
ORDER BY p.DATE;
'''
SELECT_PROJECTION_ORDERED_BY_DATE = '''
SELECT PROJECTION.*, COUNT(RESERVATION.ROW * RESERVATION.COL)
FROM PROJECTION
LEFT JOIN RESERVATION
ON RESERVATION.PROJECTION_ID = PROJECTION.ID
WHERE PROJECTION.MOVIE_ID = ? AND PROJECTION.DATE BETWEEN ? AND ?
GROUP BY PROJECTION.ID
ORDER BY PROJECTION.DATE;
'''
IS_USER_IN_USERS = '''
SELECT *
FROM USER
WHERE USERNAME = ? and PASSWORD LIKE ?;
'''
SELECT_AVAILABLE_SEATS = '''
SELECT ROW, COL
FROM RESERVATION
WHERE PROJECTION_ID = ? AND ROW IS NOT NULL AND COL IS NOT NULL;
'''
SELECT_PROJECTION_BY_ID = '''
SELECT *
FROM PROJECTION as p
WHERE p.ID = ?
'''
SELECT_MOVIE_BY_ID = '''
SELECT *
FROM MOVIE as m
WHERE m.ID = ?
'''
|
Select Queries Table names changed
|
Select Queries Table names changed
|
Python
|
mit
|
BrickText/JHROM
|
---
+++
@@ -27,13 +27,13 @@
IS_USER_IN_USERS = '''
SELECT *
- FROM USERS
+ FROM USER
WHERE USERNAME = ? and PASSWORD LIKE ?;
'''
SELECT_AVAILABLE_SEATS = '''
SELECT ROW, COL
- FROM RESERVATIONS
+ FROM RESERVATION
WHERE PROJECTION_ID = ? AND ROW IS NOT NULL AND COL IS NOT NULL;
'''
|
01cd080395533b9e8d53f4c203ef6be185d97ebc
|
dbaas/integrations/iaas/manager.py
|
dbaas/integrations/iaas/manager.py
|
from dbaas_cloudstack.provider import CloudStackProvider
from pre_provisioned.pre_provisioned_provider import PreProvisionedProvider
import logging
LOG = logging.getLogger(__name__)
class IaaSManager():
@classmethod
def destroy_instance(cls, database, *args, **kwargs):
plan = database.plan
provider = plan.provider
if provider == plan.PREPROVISIONED:
LOG.info("Destroying pre provisioned database...")
PreProvisionedProvider().destroy_instance(database, *args, **kwargs)
elif provider == plan.CLOUDSTACK:
LOG.info("Destroying cloud stack instance...")
CloudStackProvider().destroy_instance(database, *args, **kwargs)
@classmethod
def create_instance(cls, plan, environment, name):
if plan.provider == plan.PREPROVISIONED:
LOG.info("Creating pre provisioned instance...")
return PreProvisionedProvider().create_instance(plan, environment)
elif plan.provider == plan.CLOUDSTACK:
LOG.info("Creating cloud stack instance...")
return CloudStackProvider().create_instance(plan, environment, name)
|
from dbaas_cloudstack.provider import CloudStackProvider
from pre_provisioned.pre_provisioned_provider import PreProvisionedProvider
from integrations.monitoring.manager import MonitoringManager
import logging
LOG = logging.getLogger(__name__)
class IaaSManager():
@classmethod
def destroy_instance(cls, database, *args, **kwargs):
plan = database.plan
provider = plan.provider
if provider == plan.PREPROVISIONED:
LOG.info("Destroying pre provisioned database...")
PreProvisionedProvider().destroy_instance(database, *args, **kwargs)
elif provider == plan.CLOUDSTACK:
LOG.info("Destroying cloud stack instance...")
MonitoringManager.remove_monitoring(database.databaseinfra)
CloudStackProvider().destroy_instance(database, *args, **kwargs)
@classmethod
def create_instance(cls, plan, environment, name):
if plan.provider == plan.PREPROVISIONED:
LOG.info("Creating pre provisioned instance...")
return PreProvisionedProvider().create_instance(plan, environment)
elif plan.provider == plan.CLOUDSTACK:
LOG.info("Creating cloud stack instance...")
databaseinfra = CloudStackProvider().create_instance(plan, environment, name)
if databaseinfra is not None:
MonitoringManager.create_monitoring(databaseinfra)
return databaseinfra
|
Add call to monitoring app
|
Add call to monitoring app
|
Python
|
bsd-3-clause
|
globocom/database-as-a-service,globocom/database-as-a-service,globocom/database-as-a-service,globocom/database-as-a-service
|
---
+++
@@ -1,5 +1,6 @@
from dbaas_cloudstack.provider import CloudStackProvider
from pre_provisioned.pre_provisioned_provider import PreProvisionedProvider
+from integrations.monitoring.manager import MonitoringManager
import logging
LOG = logging.getLogger(__name__)
@@ -15,6 +16,7 @@
PreProvisionedProvider().destroy_instance(database, *args, **kwargs)
elif provider == plan.CLOUDSTACK:
LOG.info("Destroying cloud stack instance...")
+ MonitoringManager.remove_monitoring(database.databaseinfra)
CloudStackProvider().destroy_instance(database, *args, **kwargs)
@classmethod
@@ -24,4 +26,7 @@
return PreProvisionedProvider().create_instance(plan, environment)
elif plan.provider == plan.CLOUDSTACK:
LOG.info("Creating cloud stack instance...")
- return CloudStackProvider().create_instance(plan, environment, name)
+ databaseinfra = CloudStackProvider().create_instance(plan, environment, name)
+ if databaseinfra is not None:
+ MonitoringManager.create_monitoring(databaseinfra)
+ return databaseinfra
|
2313424f811c59563090e77966d906dd3eb7f127
|
tools/buildbot.py
|
tools/buildbot.py
|
import os
import sys
def usage():
print '%s all -- build all bsp' % os.path.basename(sys.argv[0])
print '%s clean -- clean all bsp' % os.path.basename(sys.argv[0])
print '%s project -- update all prject files' % os.path.basename(sys.argv[0])
BSP_ROOT = '../bsp'
if len(sys.argv) != 2:
usage()
sys.exit(0)
# get command options
command = ''
if sys.argv[1] == 'all':
command = ' '
elif sys.argv[1] == 'clean':
command = ' -c'
elif sys.argv[1] == 'project':
command = ' --target=mdk -s'
else:
usage()
sys.exit(0)
projects = os.listdir(BSP_ROOT)
for item in projects:
project_dir = os.path.join(BSP_ROOT, item)
if os.path.isfile(os.path.join(project_dir, 'SConstruct')):
if os.system('scons --directory=' + project_dir + command) != 0:
print 'build failed!!'
break
|
import os
import sys
def usage():
print '%s all -- build all bsp' % os.path.basename(sys.argv[0])
print '%s clean -- clean all bsp' % os.path.basename(sys.argv[0])
print '%s project -- update all prject files' % os.path.basename(sys.argv[0])
BSP_ROOT = '../bsp'
if len(sys.argv) != 2:
usage()
sys.exit(0)
# get command options
command = ''
if sys.argv[1] == 'all':
command = ' '
elif sys.argv[1] == 'clean':
command = ' -c'
elif sys.argv[1] == 'project':
command = ' --target=mdk -s'
projects = os.listdir(BSP_ROOT)
for item in projects:
project_dir = os.path.join(BSP_ROOT, item)
if os.path.isfile(os.path.join(project_dir, 'template.uvproj')):
print ('prepare MDK project file on ' + project_dir)
os.system('scons --directory=' + project_dir + command)
sys.exit(0)
else:
usage()
sys.exit(0)
projects = os.listdir(BSP_ROOT)
for item in projects:
project_dir = os.path.join(BSP_ROOT, item)
if os.path.isfile(os.path.join(project_dir, 'SConstruct')):
if os.system('scons --directory=' + project_dir + command) != 0:
print 'build failed!!'
break
|
Add better way to generate MDK project file.
|
[tools] Add better way to generate MDK project file.
|
Python
|
apache-2.0
|
RT-Thread/rt-thread,FlyLu/rt-thread,weiyuliang/rt-thread,ArdaFu/rt-thread,weiyuliang/rt-thread,geniusgogo/rt-thread,weety/rt-thread,igou/rt-thread,weiyuliang/rt-thread,hezlog/rt-thread,armink/rt-thread,wolfgangz2013/rt-thread,armink/rt-thread,ArdaFu/rt-thread,hezlog/rt-thread,AubrCool/rt-thread,geniusgogo/rt-thread,yongli3/rt-thread,hezlog/rt-thread,nongxiaoming/rt-thread,weety/rt-thread,igou/rt-thread,zhaojuntao/rt-thread,zhaojuntao/rt-thread,yongli3/rt-thread,gbcwbz/rt-thread,zhaojuntao/rt-thread,RT-Thread/rt-thread,igou/rt-thread,nongxiaoming/rt-thread,weety/rt-thread,zhaojuntao/rt-thread,FlyLu/rt-thread,armink/rt-thread,wolfgangz2013/rt-thread,gbcwbz/rt-thread,igou/rt-thread,nongxiaoming/rt-thread,geniusgogo/rt-thread,armink/rt-thread,weiyuliang/rt-thread,zhaojuntao/rt-thread,AubrCool/rt-thread,zhaojuntao/rt-thread,wolfgangz2013/rt-thread,hezlog/rt-thread,weety/rt-thread,hezlog/rt-thread,geniusgogo/rt-thread,geniusgogo/rt-thread,FlyLu/rt-thread,wolfgangz2013/rt-thread,ArdaFu/rt-thread,hezlog/rt-thread,igou/rt-thread,RT-Thread/rt-thread,FlyLu/rt-thread,yongli3/rt-thread,weiyuliang/rt-thread,nongxiaoming/rt-thread,gbcwbz/rt-thread,yongli3/rt-thread,armink/rt-thread,zhaojuntao/rt-thread,RT-Thread/rt-thread,nongxiaoming/rt-thread,geniusgogo/rt-thread,gbcwbz/rt-thread,yongli3/rt-thread,igou/rt-thread,RT-Thread/rt-thread,wolfgangz2013/rt-thread,yongli3/rt-thread,wolfgangz2013/rt-thread,weety/rt-thread,igou/rt-thread,ArdaFu/rt-thread,gbcwbz/rt-thread,RT-Thread/rt-thread,weiyuliang/rt-thread,FlyLu/rt-thread,FlyLu/rt-thread,yongli3/rt-thread,AubrCool/rt-thread,RT-Thread/rt-thread,geniusgogo/rt-thread,FlyLu/rt-thread,weety/rt-thread,armink/rt-thread,gbcwbz/rt-thread,weety/rt-thread,ArdaFu/rt-thread,nongxiaoming/rt-thread,gbcwbz/rt-thread,AubrCool/rt-thread,armink/rt-thread,nongxiaoming/rt-thread,ArdaFu/rt-thread,wolfgangz2013/rt-thread,AubrCool/rt-thread,AubrCool/rt-thread,ArdaFu/rt-thread,weiyuliang/rt-thread,hezlog/rt-thread,AubrCool/rt-thread
|
---
+++
@@ -19,6 +19,17 @@
command = ' -c'
elif sys.argv[1] == 'project':
command = ' --target=mdk -s'
+
+ projects = os.listdir(BSP_ROOT)
+ for item in projects:
+ project_dir = os.path.join(BSP_ROOT, item)
+
+ if os.path.isfile(os.path.join(project_dir, 'template.uvproj')):
+ print ('prepare MDK project file on ' + project_dir)
+
+ os.system('scons --directory=' + project_dir + command)
+
+ sys.exit(0)
else:
usage()
sys.exit(0)
|
155822548be11161aefdb0d93d5ec86095ab3624
|
rt.py
|
rt.py
|
import queue
import threading
def loop(queue, actor):
while True:
message = queue.get()
actor.behavior(message)
class Actor(object):
def __init__(self):
pass
def _start_loop(self):
self.queue = queue.Queue()
self.dispatcher = threading.Thread(
target=loop,
args=(self.queue, self))
self.dispatcher.start()
def __call__(self, message):
self.queue.put(message)
@classmethod
def create(cls, *args):
actor = cls(*args)
actor._start_loop()
return actor
|
import queue
import threading
def indiviual_loop(queue, actor):
while True:
message = queue.get()
actor.behavior(message)
def global_loop(queue):
while True:
actor, message = queue.get()
actor.behavior(message)
class EventLoop(object):
loop = None
def __init__(self):
self.queue = queue.Queue()
self.thread = threading.Thread(
target=global_loop,
args=(self.queue,),
name='global-loop')
self.thread.start()
def schedule(self, message, target):
self.queue.put((target, message))
@classmethod
def get_loop(cls):
if cls.loop is None:
cls.loop = cls()
return cls.loop
class AbstractActor(object):
def __call__(self, message):
self._put(message)
def _put(self, message):
raise NotImplementedError()
def _ensure_loop(self):
pass
@classmethod
def create(cls, *args):
actor = cls(*args)
actor._ensure_loop()
return actor
class ActorOwnLoop(AbstractActor):
def _put(self, message):
self.queue.put(message)
def _ensure_loop(self):
self.queue = queue.Queue()
self.dispatcher = threading.Thread(
target=indiviual_loop,
args=(self.queue, self),
name=self._thread_name())
self.dispatcher.start()
def _thread_name(self):
return '{}-{}'.format(
self.__class__.__name__,
hex(id(self)))
class ActorGlobalLoop(AbstractActor):
def _put(self, message):
self.loop.schedule(message, self)
def _ensure_loop(self):
self.loop = EventLoop.get_loop()
Actor = ActorGlobalLoop
|
Refactor to allow different event loops
|
Refactor to allow different event loops
|
Python
|
mit
|
waltermoreira/tartpy
|
---
+++
@@ -2,29 +2,84 @@
import threading
-def loop(queue, actor):
+def indiviual_loop(queue, actor):
while True:
message = queue.get()
actor.behavior(message)
+
+def global_loop(queue):
+ while True:
+ actor, message = queue.get()
+ actor.behavior(message)
+
+
+class EventLoop(object):
+
+ loop = None
+
+ def __init__(self):
+ self.queue = queue.Queue()
+ self.thread = threading.Thread(
+ target=global_loop,
+ args=(self.queue,),
+ name='global-loop')
+ self.thread.start()
+
+ def schedule(self, message, target):
+ self.queue.put((target, message))
-class Actor(object):
+ @classmethod
+ def get_loop(cls):
+ if cls.loop is None:
+ cls.loop = cls()
+ return cls.loop
+
- def __init__(self):
+class AbstractActor(object):
+
+ def __call__(self, message):
+ self._put(message)
+
+ def _put(self, message):
+ raise NotImplementedError()
+
+ def _ensure_loop(self):
pass
- def _start_loop(self):
- self.queue = queue.Queue()
- self.dispatcher = threading.Thread(
- target=loop,
- args=(self.queue, self))
- self.dispatcher.start()
-
- def __call__(self, message):
- self.queue.put(message)
-
@classmethod
def create(cls, *args):
actor = cls(*args)
- actor._start_loop()
+ actor._ensure_loop()
return actor
+
+
+class ActorOwnLoop(AbstractActor):
+
+ def _put(self, message):
+ self.queue.put(message)
+
+ def _ensure_loop(self):
+ self.queue = queue.Queue()
+ self.dispatcher = threading.Thread(
+ target=indiviual_loop,
+ args=(self.queue, self),
+ name=self._thread_name())
+ self.dispatcher.start()
+
+ def _thread_name(self):
+ return '{}-{}'.format(
+ self.__class__.__name__,
+ hex(id(self)))
+
+
+class ActorGlobalLoop(AbstractActor):
+
+ def _put(self, message):
+ self.loop.schedule(message, self)
+
+ def _ensure_loop(self):
+ self.loop = EventLoop.get_loop()
+
+
+Actor = ActorGlobalLoop
|
ef73cc38e08db709a33321cde47ceb6dfce3504d
|
chrome/test/functional/chromeos_private_view.py
|
chrome/test/functional/chromeos_private_view.py
|
#!/usr/bin/env python
# Copyright (c) 2012 The Chromium Authors. All rights reserved.
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
import os
import pyauto_functional # must be imported before pyauto
import pyauto
class ChromeosPrivateViewTest(pyauto.PyUITest):
"""Basic tests for ChromeOS Private View.
Requires ChromeOS to be logged in.
"""
def _GetExtensionInfoById(self, extensions, id):
for x in extensions:
if x['id'] == id:
return x
return None
def testInstallPrivateViewExtension(self):
"""Basic installation test for Private View on ChromeOS."""
crx_file_path = os.path.abspath(
os.path.join(self.DataDir(), 'pyauto_private', 'apps',
'privateview-chrome-1.0.800_RELEASE.crx'))
ext_id = self.InstallExtension(crx_file_path)
self.assertTrue(ext_id, 'Failed to install extension.')
extension = self._GetExtensionInfoById(self.GetExtensionsInfo(), ext_id)
self.assertTrue(extension['is_enabled'],
msg='Extension was not enabled on installation')
self.assertFalse(extension['allowed_in_incognito'],
msg='Extension was allowed in incognito on installation.')
if __name__ == '__main__':
pyauto_functional.Main()
|
#!/usr/bin/env python
# Copyright (c) 2012 The Chromium Authors. All rights reserved.
# Use of this source code is governed by a BSD-style license that can be
# found in the LICENSE file.
import os
import pyauto_functional # must be imported before pyauto
import pyauto
class ChromeosPrivateViewTest(pyauto.PyUITest):
"""Basic tests for ChromeOS Private View.
Requires ChromeOS to be logged in.
"""
def _GetExtensionInfoById(self, extensions, id):
for x in extensions:
if x['id'] == id:
return x
return None
def testInstallPrivateViewExtension(self):
"""Basic installation test for Private View on ChromeOS."""
crx_file_path = os.path.abspath(
os.path.join(self.DataDir(), 'pyauto_private', 'apps',
'privateview-chrome-2.2.28_RELEASE.crx'))
ext_id = self.InstallExtension(crx_file_path)
self.assertTrue(ext_id, 'Failed to install extension.')
extension = self._GetExtensionInfoById(self.GetExtensionsInfo(), ext_id)
self.assertTrue(extension['is_enabled'],
msg='Extension was not enabled on installation')
self.assertFalse(extension['allowed_in_incognito'],
msg='Extension was allowed in incognito on installation.')
if __name__ == '__main__':
pyauto_functional.Main()
|
Modify pyauto test ChromeosPrivateViewTest to use 2.2.28 data file.
|
Modify pyauto test ChromeosPrivateViewTest to use 2.2.28 data file.
BUG=none
TEST=This is a test.
Review URL: https://chromiumcodereview.appspot.com/10389084
git-svn-id: de016e52bd170d2d4f2344f9bf92d50478b649e0@136454 0039d316-1c4b-4281-b951-d872f2087c98
|
Python
|
bsd-3-clause
|
TheTypoMaster/chromium-crosswalk,dednal/chromium.src,zcbenz/cefode-chromium,chuan9/chromium-crosswalk,Fireblend/chromium-crosswalk,ChromiumWebApps/chromium,timopulkkinen/BubbleFish,Jonekee/chromium.src,zcbenz/cefode-chromium,krieger-od/nwjs_chromium.src,Chilledheart/chromium,TheTypoMaster/chromium-crosswalk,hujiajie/pa-chromium,jaruba/chromium.src,Just-D/chromium-1,hgl888/chromium-crosswalk-efl,Just-D/chromium-1,markYoungH/chromium.src,hujiajie/pa-chromium,axinging/chromium-crosswalk,PeterWangIntel/chromium-crosswalk,axinging/chromium-crosswalk,dednal/chromium.src,Fireblend/chromium-crosswalk,chuan9/chromium-crosswalk,axinging/chromium-crosswalk,M4sse/chromium.src,PeterWangIntel/chromium-crosswalk,junmin-zhu/chromium-rivertrail,junmin-zhu/chromium-rivertrail,krieger-od/nwjs_chromium.src,keishi/chromium,timopulkkinen/BubbleFish,crosswalk-project/chromium-crosswalk-efl,M4sse/chromium.src,mogoweb/chromium-crosswalk,ChromiumWebApps/chromium,Jonekee/chromium.src,mohamed--abdel-maksoud/chromium.src,Just-D/chromium-1,ChromiumWebApps/chromium,hgl888/chromium-crosswalk-efl,crosswalk-project/chromium-crosswalk-efl,pozdnyakov/chromium-crosswalk,ltilve/chromium,patrickm/chromium.src,mohamed--abdel-maksoud/chromium.src,krieger-od/nwjs_chromium.src,PeterWangIntel/chromium-crosswalk,PeterWangIntel/chromium-crosswalk,dushu1203/chromium.src,jaruba/chromium.src,pozdnyakov/chromium-crosswalk,ChromiumWebApps/chromium,hgl888/chromium-crosswalk,robclark/chromium,Pluto-tv/chromium-crosswalk,dednal/chromium.src,bright-sparks/chromium-spacewalk,patrickm/chromium.src,krieger-od/nwjs_chromium.src,ltilve/chromium,ChromiumWebApps/chromium,krieger-od/nwjs_chromium.src,M4sse/chromium.src,nacl-webkit/chrome_deps,crosswalk-project/chromium-crosswalk-efl,chuan9/chromium-crosswalk,ltilve/chromium,Chilledheart/chromium,markYoungH/chromium.src,TheTypoMaster/chromium-crosswalk,hgl888/chromium-crosswalk,bright-sparks/chromium-spacewalk,robclark/chromium,ondra-novak/chromium.src,ltilve/chromium,Jonekee/chromium.src,hgl888/chromium-crosswalk,jaruba/chromium.src,ChromiumWebApps/chromium,pozdnyakov/chromium-crosswalk,PeterWangIntel/chromium-crosswalk,Jonekee/chromium.src,Jonekee/chromium.src,M4sse/chromium.src,hgl888/chromium-crosswalk-efl,crosswalk-project/chromium-crosswalk-efl,fujunwei/chromium-crosswalk,jaruba/chromium.src,dushu1203/chromium.src,keishi/chromium,hgl888/chromium-crosswalk-efl,axinging/chromium-crosswalk,dushu1203/chromium.src,hgl888/chromium-crosswalk-efl,Just-D/chromium-1,dushu1203/chromium.src,dushu1203/chromium.src,robclark/chromium,hujiajie/pa-chromium,bright-sparks/chromium-spacewalk,ltilve/chromium,Just-D/chromium-1,mohamed--abdel-maksoud/chromium.src,Chilledheart/chromium,Pluto-tv/chromium-crosswalk,hujiajie/pa-chromium,markYoungH/chromium.src,chuan9/chromium-crosswalk,fujunwei/chromium-crosswalk,mogoweb/chromium-crosswalk,Fireblend/chromium-crosswalk,keishi/chromium,ChromiumWebApps/chromium,junmin-zhu/chromium-rivertrail,nacl-webkit/chrome_deps,Jonekee/chromium.src,littlstar/chromium.src,junmin-zhu/chromium-rivertrail,dushu1203/chromium.src,junmin-zhu/chromium-rivertrail,mogoweb/chromium-crosswalk,Chilledheart/chromium,keishi/chromium,krieger-od/nwjs_chromium.src,TheTypoMaster/chromium-crosswalk,ondra-novak/chromium.src,markYoungH/chromium.src,pozdnyakov/chromium-crosswalk,timopulkkinen/BubbleFish,zcbenz/cefode-chromium,timopulkkinen/BubbleFish,junmin-zhu/chromium-rivertrail,nacl-webkit/chrome_deps,krieger-od/nwjs_chromium.src,hgl888/chromium-crosswalk,robclark/chromium,mohamed--abdel-maksoud/chromium.src,TheTypoMaster/chromium-crosswalk,bright-sparks/chromium-spacewalk,fujunwei/chromium-crosswalk,anirudhSK/chromium,keishi/chromium,axinging/chromium-crosswalk,M4sse/chromium.src,fujunwei/chromium-crosswalk,TheTypoMaster/chromium-crosswalk,Pluto-tv/chromium-crosswalk,bright-sparks/chromium-spacewalk,dednal/chromium.src,keishi/chromium,zcbenz/cefode-chromium,markYoungH/chromium.src,timopulkkinen/BubbleFish,anirudhSK/chromium,mogoweb/chromium-crosswalk,hgl888/chromium-crosswalk-efl,dushu1203/chromium.src,PeterWangIntel/chromium-crosswalk,Jonekee/chromium.src,hgl888/chromium-crosswalk,littlstar/chromium.src,jaruba/chromium.src,ChromiumWebApps/chromium,keishi/chromium,TheTypoMaster/chromium-crosswalk,mogoweb/chromium-crosswalk,timopulkkinen/BubbleFish,axinging/chromium-crosswalk,markYoungH/chromium.src,Jonekee/chromium.src,Fireblend/chromium-crosswalk,mohamed--abdel-maksoud/chromium.src,patrickm/chromium.src,M4sse/chromium.src,M4sse/chromium.src,Chilledheart/chromium,markYoungH/chromium.src,Chilledheart/chromium,mogoweb/chromium-crosswalk,hujiajie/pa-chromium,dednal/chromium.src,mohamed--abdel-maksoud/chromium.src,junmin-zhu/chromium-rivertrail,Chilledheart/chromium,zcbenz/cefode-chromium,axinging/chromium-crosswalk,patrickm/chromium.src,mohamed--abdel-maksoud/chromium.src,ltilve/chromium,anirudhSK/chromium,mohamed--abdel-maksoud/chromium.src,keishi/chromium,hujiajie/pa-chromium,littlstar/chromium.src,nacl-webkit/chrome_deps,chuan9/chromium-crosswalk,bright-sparks/chromium-spacewalk,markYoungH/chromium.src,zcbenz/cefode-chromium,dednal/chromium.src,zcbenz/cefode-chromium,pozdnyakov/chromium-crosswalk,anirudhSK/chromium,hujiajie/pa-chromium,fujunwei/chromium-crosswalk,Just-D/chromium-1,timopulkkinen/BubbleFish,littlstar/chromium.src,Jonekee/chromium.src,chuan9/chromium-crosswalk,ondra-novak/chromium.src,anirudhSK/chromium,Just-D/chromium-1,krieger-od/nwjs_chromium.src,robclark/chromium,dushu1203/chromium.src,axinging/chromium-crosswalk,robclark/chromium,timopulkkinen/BubbleFish,Fireblend/chromium-crosswalk,nacl-webkit/chrome_deps,jaruba/chromium.src,fujunwei/chromium-crosswalk,mohamed--abdel-maksoud/chromium.src,ltilve/chromium,patrickm/chromium.src,ondra-novak/chromium.src,nacl-webkit/chrome_deps,hgl888/chromium-crosswalk,TheTypoMaster/chromium-crosswalk,jaruba/chromium.src,PeterWangIntel/chromium-crosswalk,timopulkkinen/BubbleFish,robclark/chromium,chuan9/chromium-crosswalk,Jonekee/chromium.src,junmin-zhu/chromium-rivertrail,ChromiumWebApps/chromium,patrickm/chromium.src,pozdnyakov/chromium-crosswalk,hgl888/chromium-crosswalk-efl,axinging/chromium-crosswalk,Pluto-tv/chromium-crosswalk,patrickm/chromium.src,robclark/chromium,hgl888/chromium-crosswalk,junmin-zhu/chromium-rivertrail,dednal/chromium.src,Just-D/chromium-1,mogoweb/chromium-crosswalk,Pluto-tv/chromium-crosswalk,anirudhSK/chromium,zcbenz/cefode-chromium,markYoungH/chromium.src,Chilledheart/chromium,PeterWangIntel/chromium-crosswalk,anirudhSK/chromium,keishi/chromium,dednal/chromium.src,Jonekee/chromium.src,Fireblend/chromium-crosswalk,nacl-webkit/chrome_deps,jaruba/chromium.src,chuan9/chromium-crosswalk,pozdnyakov/chromium-crosswalk,hgl888/chromium-crosswalk-efl,mohamed--abdel-maksoud/chromium.src,nacl-webkit/chrome_deps,M4sse/chromium.src,jaruba/chromium.src,ltilve/chromium,littlstar/chromium.src,ondra-novak/chromium.src,ondra-novak/chromium.src,littlstar/chromium.src,hgl888/chromium-crosswalk-efl,Pluto-tv/chromium-crosswalk,ChromiumWebApps/chromium,zcbenz/cefode-chromium,keishi/chromium,junmin-zhu/chromium-rivertrail,zcbenz/cefode-chromium,littlstar/chromium.src,ltilve/chromium,hujiajie/pa-chromium,dednal/chromium.src,axinging/chromium-crosswalk,krieger-od/nwjs_chromium.src,jaruba/chromium.src,ChromiumWebApps/chromium,hujiajie/pa-chromium,pozdnyakov/chromium-crosswalk,Pluto-tv/chromium-crosswalk,nacl-webkit/chrome_deps,keishi/chromium,anirudhSK/chromium,anirudhSK/chromium,M4sse/chromium.src,fujunwei/chromium-crosswalk,timopulkkinen/BubbleFish,PeterWangIntel/chromium-crosswalk,anirudhSK/chromium,littlstar/chromium.src,pozdnyakov/chromium-crosswalk,M4sse/chromium.src,chuan9/chromium-crosswalk,nacl-webkit/chrome_deps,dushu1203/chromium.src,crosswalk-project/chromium-crosswalk-efl,hgl888/chromium-crosswalk,mogoweb/chromium-crosswalk,robclark/chromium,krieger-od/nwjs_chromium.src,crosswalk-project/chromium-crosswalk-efl,markYoungH/chromium.src,crosswalk-project/chromium-crosswalk-efl,jaruba/chromium.src,dednal/chromium.src,mohamed--abdel-maksoud/chromium.src,hujiajie/pa-chromium,anirudhSK/chromium,bright-sparks/chromium-spacewalk,ondra-novak/chromium.src,bright-sparks/chromium-spacewalk,crosswalk-project/chromium-crosswalk-efl,junmin-zhu/chromium-rivertrail,ondra-novak/chromium.src,axinging/chromium-crosswalk,Fireblend/chromium-crosswalk,patrickm/chromium.src,markYoungH/chromium.src,dushu1203/chromium.src,krieger-od/nwjs_chromium.src,mogoweb/chromium-crosswalk,zcbenz/cefode-chromium,Chilledheart/chromium,ChromiumWebApps/chromium,ondra-novak/chromium.src,M4sse/chromium.src,TheTypoMaster/chromium-crosswalk,Pluto-tv/chromium-crosswalk,Pluto-tv/chromium-crosswalk,crosswalk-project/chromium-crosswalk-efl,Fireblend/chromium-crosswalk,pozdnyakov/chromium-crosswalk,hujiajie/pa-chromium,robclark/chromium,hgl888/chromium-crosswalk-efl,nacl-webkit/chrome_deps,timopulkkinen/BubbleFish,pozdnyakov/chromium-crosswalk,Fireblend/chromium-crosswalk,anirudhSK/chromium,hgl888/chromium-crosswalk,fujunwei/chromium-crosswalk,dushu1203/chromium.src,mogoweb/chromium-crosswalk,patrickm/chromium.src,fujunwei/chromium-crosswalk,dednal/chromium.src,Just-D/chromium-1,bright-sparks/chromium-spacewalk
|
---
+++
@@ -25,7 +25,7 @@
"""Basic installation test for Private View on ChromeOS."""
crx_file_path = os.path.abspath(
os.path.join(self.DataDir(), 'pyauto_private', 'apps',
- 'privateview-chrome-1.0.800_RELEASE.crx'))
+ 'privateview-chrome-2.2.28_RELEASE.crx'))
ext_id = self.InstallExtension(crx_file_path)
self.assertTrue(ext_id, 'Failed to install extension.')
extension = self._GetExtensionInfoById(self.GetExtensionsInfo(), ext_id)
|
db43b3b3079842fb2baf6d181ef39374acf0053c
|
.gitlab/linters/check-makefiles.py
|
.gitlab/linters/check-makefiles.py
|
#!/usr/bin/env python3
"""
Warn for use of `--interactive` inside Makefiles (#11468).
Encourage the use of `$(TEST_HC_OPTS_INTERACTIVE)` instead of
`$(TEST_HC_OPTS) --interactive -ignore-dot-ghci -v0`. It's too easy to
forget one of those flags when adding a new test.
"""
from linter import run_linters, RegexpLinter
linters = [
RegexpLinter(r'--interactive',
message = "Warning: Use `$(TEST_HC_OPTS_INTERACTIVE)` instead of `--interactive -ignore-dot-ghci -v0`."
).add_path_filter(lambda path: path.name == 'Makefile')
]
if __name__ == '__main__':
run_linters(linters,
subdir='testsuite')
|
#!/usr/bin/env python3
"""
Linters for testsuite makefiles
"""
from linter import run_linters, RegexpLinter
"""
Warn for use of `--interactive` inside Makefiles (#11468).
Encourage the use of `$(TEST_HC_OPTS_INTERACTIVE)` instead of
`$(TEST_HC_OPTS) --interactive -ignore-dot-ghci -v0`. It's too easy to
forget one of those flags when adding a new test.
"""
interactive_linter = \
RegexpLinter(r'--interactive',
message = "Warning: Use `$(TEST_HC_OPTS_INTERACTIVE)` instead of `--interactive -ignore-dot-ghci -v0`."
).add_path_filter(lambda path: path.name == 'Makefile')
test_hc_quotes_linter = \
RegexpLinter('\t\\$\\(TEST_HC\\)',
message = "Warning: $(TEST_HC) should be quoted in Makefiles.",
).add_path_filter(lambda path: path.name == 'Makefile')
linters = [
interactive_linter,
test_hc_quotes_linter,
]
if __name__ == '__main__':
run_linters(linters,
subdir='testsuite')
|
Add linter to catch unquoted use of $(TEST_HC)
|
linters: Add linter to catch unquoted use of $(TEST_HC)
This is a common bug that creeps into Makefiles (e.g. see T12674).
|
Python
|
bsd-3-clause
|
sdiehl/ghc,sdiehl/ghc,sdiehl/ghc,sdiehl/ghc,sdiehl/ghc,sdiehl/ghc,sdiehl/ghc
|
---
+++
@@ -1,4 +1,10 @@
#!/usr/bin/env python3
+
+"""
+Linters for testsuite makefiles
+"""
+
+from linter import run_linters, RegexpLinter
"""
Warn for use of `--interactive` inside Makefiles (#11468).
@@ -7,13 +13,19 @@
`$(TEST_HC_OPTS) --interactive -ignore-dot-ghci -v0`. It's too easy to
forget one of those flags when adding a new test.
"""
-
-from linter import run_linters, RegexpLinter
-
-linters = [
+interactive_linter = \
RegexpLinter(r'--interactive',
message = "Warning: Use `$(TEST_HC_OPTS_INTERACTIVE)` instead of `--interactive -ignore-dot-ghci -v0`."
).add_path_filter(lambda path: path.name == 'Makefile')
+
+test_hc_quotes_linter = \
+ RegexpLinter('\t\\$\\(TEST_HC\\)',
+ message = "Warning: $(TEST_HC) should be quoted in Makefiles.",
+ ).add_path_filter(lambda path: path.name == 'Makefile')
+
+linters = [
+ interactive_linter,
+ test_hc_quotes_linter,
]
if __name__ == '__main__':
|
dc1a7bc4d674fd6e7235222612f1d147112d77db
|
src/nodeconductor_assembly_waldur/packages/migrations/0002_openstack_packages.py
|
src/nodeconductor_assembly_waldur/packages/migrations/0002_openstack_packages.py
|
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import migrations, models
import nodeconductor.core.fields
class Migration(migrations.Migration):
dependencies = [
('openstack', '0022_volume_device'),
('structure', '0037_remove_customer_billing_backend_id'),
('packages', '0001_initial'),
]
operations = [
migrations.CreateModel(
name='OpenStackPackage',
fields=[
('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),
('uuid', nodeconductor.core.fields.UUIDField()),
('service_settings', models.ForeignKey(related_name='+', to='structure.ServiceSettings')),
],
options={
'abstract': False,
},
),
migrations.RemoveField(
model_name='packagetemplate',
name='type',
),
migrations.AddField(
model_name='packagetemplate',
name='service_settings',
field=models.ForeignKey(related_name='+', default=1, to='structure.ServiceSettings'),
preserve_default=False,
),
migrations.AddField(
model_name='openstackpackage',
name='template',
field=models.ForeignKey(related_name='openstack_packages', to='packages.PackageTemplate', help_text='Tenant will be created based on this template.'),
),
migrations.AddField(
model_name='openstackpackage',
name='tenant',
field=models.ForeignKey(related_name='+', to='openstack.Tenant'),
),
]
|
# -*- coding: utf-8 -*-
from __future__ import unicode_literals
from django.db import migrations, models
import nodeconductor.core.fields
class Migration(migrations.Migration):
dependencies = [
('openstack', '0022_volume_device'),
('structure', '0037_remove_customer_billing_backend_id'),
('packages', '0001_initial'),
]
operations = [
migrations.CreateModel(
name='OpenStackPackage',
fields=[
('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)),
('uuid', nodeconductor.core.fields.UUIDField()),
('service_settings', models.ForeignKey(related_name='+', to='structure.ServiceSettings')),
],
options={
'abstract': False,
},
),
migrations.RemoveField(
model_name='packagetemplate',
name='type',
),
migrations.AddField(
model_name='packagetemplate',
name='service_settings',
field=models.ForeignKey(related_name='+', to='structure.ServiceSettings'),
preserve_default=False,
),
migrations.AddField(
model_name='openstackpackage',
name='template',
field=models.ForeignKey(related_name='openstack_packages', to='packages.PackageTemplate', help_text='Tenant will be created based on this template.'),
),
migrations.AddField(
model_name='openstackpackage',
name='tenant',
field=models.ForeignKey(related_name='+', to='openstack.Tenant'),
),
]
|
Remove useless default from migration
|
Remove useless default from migration
- wal-26
|
Python
|
mit
|
opennode/waldur-mastermind,opennode/nodeconductor-assembly-waldur,opennode/nodeconductor-assembly-waldur,opennode/nodeconductor-assembly-waldur,opennode/waldur-mastermind,opennode/waldur-mastermind,opennode/waldur-mastermind
|
---
+++
@@ -32,7 +32,7 @@
migrations.AddField(
model_name='packagetemplate',
name='service_settings',
- field=models.ForeignKey(related_name='+', default=1, to='structure.ServiceSettings'),
+ field=models.ForeignKey(related_name='+', to='structure.ServiceSettings'),
preserve_default=False,
),
migrations.AddField(
|
5e504ae82c37b47ad2c1781cc74d13cdccc72495
|
dthm4kaiako/config/__init__.py
|
dthm4kaiako/config/__init__.py
|
"""Configuration for Django system."""
__version__ = "0.16.2"
__version_info__ = tuple(
[
int(num) if num.isdigit() else num
for num in __version__.replace("-", ".", 1).split(".")
]
)
|
"""Configuration for Django system."""
__version__ = "0.16.3"
__version_info__ = tuple(
[
int(num) if num.isdigit() else num
for num in __version__.replace("-", ".", 1).split(".")
]
)
|
Increment version number to 0.16.3
|
Increment version number to 0.16.3
|
Python
|
mit
|
uccser/cs4teachers,uccser/cs4teachers,uccser/cs4teachers,uccser/cs4teachers
|
---
+++
@@ -1,6 +1,6 @@
"""Configuration for Django system."""
-__version__ = "0.16.2"
+__version__ = "0.16.3"
__version_info__ = tuple(
[
int(num) if num.isdigit() else num
|
3dbc981e62c2d153913557b62083f60888fa7e83
|
ynr/apps/ynr_refactoring/management/commands/ynr_refactoring_remove_legacy_IDs.py
|
ynr/apps/ynr_refactoring/management/commands/ynr_refactoring_remove_legacy_IDs.py
|
import json
from django.core.management.base import BaseCommand
from django.db import transaction
from people.models import Person
from candidates.views.version_data import get_change_metadata
from popolo.models import Identifier
class Command(BaseCommand):
def handle(self, *args, **options):
schemes = ("yournextmp-candidate", "popit-person")
# We can't use the GFK any more because we just deleted it, but the
# content is still there
identifiers = Identifier.objects.filter(scheme__in=schemes).values_list(
"object_id", flat=True
)
for person in Person.objects.filter(pk__in=identifiers).filter(pk=502):
with transaction.atomic():
meta_data = get_change_metadata(
None, "Removing legacy identifiers"
)
meta_data["username"] = "CandidateBot"
person.record_version(meta_data)
person.save()
|
import json
from django.core.management.base import BaseCommand
from django.db import transaction
from people.models import Person
from candidates.views.version_data import get_change_metadata
from popolo.models import Identifier
class Command(BaseCommand):
def handle(self, *args, **options):
schemes = ("yournextmp-candidate", "popit-person")
# We can't use the GFK any more because we just deleted it, but the
# content is still there
identifiers = Identifier.objects.filter(scheme__in=schemes).values_list(
"object_id", flat=True
)
for person in Person.objects.filter(pk__in=identifiers):
with transaction.atomic():
meta_data = get_change_metadata(
None, "Removing legacy identifiers"
)
meta_data["username"] = "CandidateBot"
person.record_version(meta_data)
person.save()
|
Remove IDs for all candidates, not just Zac
|
Remove IDs for all candidates, not just Zac
|
Python
|
agpl-3.0
|
DemocracyClub/yournextrepresentative,DemocracyClub/yournextrepresentative,DemocracyClub/yournextrepresentative
|
---
+++
@@ -17,7 +17,7 @@
identifiers = Identifier.objects.filter(scheme__in=schemes).values_list(
"object_id", flat=True
)
- for person in Person.objects.filter(pk__in=identifiers).filter(pk=502):
+ for person in Person.objects.filter(pk__in=identifiers):
with transaction.atomic():
meta_data = get_change_metadata(
None, "Removing legacy identifiers"
|
e5f4dc01e94694bf9bfcae3ecd6eca34a33a24eb
|
openquake/__init__.py
|
openquake/__init__.py
|
# -*- coding: utf-8 -*-
# vim: tabstop=4 shiftwidth=4 softtabstop=4
#
# Copyright (C) 2010-2016 GEM Foundation
#
# OpenQuake is free software: you can redistribute it and/or modify it
# under the terms of the GNU Affero General Public License as published
# by the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# OpenQuake is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with OpenQuake. If not, see <http://www.gnu.org/licenses/>.
__import__('pkg_resources').declare_namespace(__name__)
|
# -*- coding: utf-8 -*-
# vim: tabstop=4 shiftwidth=4 softtabstop=4
#
# Copyright (C) 2010-2016 GEM Foundation
#
# OpenQuake is free software: you can redistribute it and/or modify it
# under the terms of the GNU Affero General Public License as published
# by the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
#
# OpenQuake is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
#
# You should have received a copy of the GNU Affero General Public License
# along with OpenQuake. If not, see <http://www.gnu.org/licenses/>.
# Make the namespace compatible with old setuptools, like the one
# provided by QGIS 2.1x on Windows
try:
__import__('pkg_resources').declare_namespace(__name__)
except ImportError:
__path__ = __import__('pkgutil').extend_path(__path__, __name__)
|
Make the openquake namespace compatible with old setuptools
|
Make the openquake namespace compatible with old setuptools
|
Python
|
agpl-3.0
|
gem/oq-engine,gem/oq-engine,gem/oq-engine,gem/oq-engine,gem/oq-engine
|
---
+++
@@ -16,4 +16,9 @@
# You should have received a copy of the GNU Affero General Public License
# along with OpenQuake. If not, see <http://www.gnu.org/licenses/>.
-__import__('pkg_resources').declare_namespace(__name__)
+# Make the namespace compatible with old setuptools, like the one
+# provided by QGIS 2.1x on Windows
+try:
+ __import__('pkg_resources').declare_namespace(__name__)
+except ImportError:
+ __path__ = __import__('pkgutil').extend_path(__path__, __name__)
|
5a69162e82c2c6031587448b975f5867c94873ed
|
pyramid_es/dotdict.py
|
pyramid_es/dotdict.py
|
class DotDict(dict):
__getattr__ = dict.__getitem__
__setattr__ = dict.__setitem__
__delattr__ = dict.__delitem__
def __init__(self, d={}):
for key, value in d.items():
if hasattr(value, 'keys'):
value = DotDict(value)
if isinstance(value, list):
value = [DotDict(el) if hasattr(el, 'keys') else el
for el in value]
self[key] = value
def __repr__(self):
return '<DotDict(%s)>' % dict.__repr__(self)
|
class DotDict(dict):
__getattr__ = dict.__getitem__
__setattr__ = dict.__setitem__
__delattr__ = dict.__delitem__
def __init__(self, d={}):
for key, value in d.items():
if hasattr(value, 'keys'):
value = DotDict(value)
if isinstance(value, list):
value = [DotDict(el) if hasattr(el, 'keys') else el
for el in value]
self[key] = value
def __repr__(self):
return '<%s(%s)>' % (self.__class__.__name__, dict.__repr__(self))
|
Make DotDict repr() use class name so that it doesn't print misleading results if subclassed
|
Make DotDict repr() use class name so that it doesn't print misleading results if subclassed
|
Python
|
mit
|
storborg/pyramid_es
|
---
+++
@@ -14,4 +14,4 @@
self[key] = value
def __repr__(self):
- return '<DotDict(%s)>' % dict.__repr__(self)
+ return '<%s(%s)>' % (self.__class__.__name__, dict.__repr__(self))
|
9a240f0efab9be036fe39f9b2b63cc399e5f8134
|
registration/admin.py
|
registration/admin.py
|
from django.contrib import admin
from registration.models import RegistrationProfile
class RegistrationAdmin(admin.ModelAdmin):
list_display = ('__unicode__', 'activation_key_expired')
search_fields = ('user__username', 'user__first_name')
admin.site.register(RegistrationProfile, RegistrationAdmin)
|
from django.contrib import admin
from registration.models import RegistrationProfile
class RegistrationAdmin(admin.ModelAdmin):
list_display = ('__unicode__', 'activation_key_expired')
raw_id_fields = ['user']
search_fields = ('user__username', 'user__first_name')
admin.site.register(RegistrationProfile, RegistrationAdmin)
|
Use raw_id_fields for the relation from RegistrationProfile to User, for sites which have huge numbers of users.
|
Use raw_id_fields for the relation from RegistrationProfile to User, for sites which have huge numbers of users.
|
Python
|
bsd-3-clause
|
dinie/django-registration,Avenza/django-registration,FundedByMe/django-registration,dinie/django-registration,FundedByMe/django-registration
|
---
+++
@@ -5,6 +5,7 @@
class RegistrationAdmin(admin.ModelAdmin):
list_display = ('__unicode__', 'activation_key_expired')
+ raw_id_fields = ['user']
search_fields = ('user__username', 'user__first_name')
|
552d216f2cbcb780c08cc351a7456fb97ac3c3dd
|
app.py
|
app.py
|
from flask import Flask, request, render_template
import os
app = Flask(__name__)
app.debug = True
# Secret Key setting based on debug setting
if app.debug:
app.secret_key = "T3st_s3cret_k3y!~$@"
else:
app.secret_key = os.urandom(30)
@app.route("/domain", methods=["GET", "POST"])
def domain():
if request.method == "GET":
# search domain
pass
elif request.method == "POST":
# register domain
pass
@app.route("/")
def index():
return render_template("index.html")
if __name__ == "__main__":
app.run(host="0.0.0.0", threaded=True)
|
from flask import Flask, request, render_template
from route import Route
from route.db import init_db, db_session
import os
app = Flask(__name__)
app.debug = True
# Secret Key setting based on debug setting
if app.debug:
app.secret_key = "T3st_s3cret_k3y!~$@"
else:
init_db()
app.secret_key = os.urandom(30)
@app.teardown_request
def remove_session(exception=None):
db_session.remove()
@app.route("/domain", methods=["GET", "POST"])
def domain():
ip = request.args.get("ip")
domain = request.args.get("domain")
if (ip and domain) is None:
return "parameter error", 400
r = Route(ip, domain)
if request.method == "GET":
# search domain
r.search()
elif request.method == "POST":
# register domain
user = register.args.get("user")
if user is not None:
r.register(user)
else:
return "user not found", 400
del r
return "{} - {}".format(ip, domain)
@app.route("/")
def index():
return render_template("index.html")
if __name__ == "__main__":
app.run(host="0.0.0.0", threaded=True)
|
Add domain route and case handling
|
Add domain route and case handling
|
Python
|
apache-2.0
|
bunseokbot/proxy_register,bunseokbot/proxy_register
|
---
+++
@@ -1,4 +1,8 @@
from flask import Flask, request, render_template
+
+from route import Route
+
+from route.db import init_db, db_session
import os
@@ -10,18 +14,41 @@
if app.debug:
app.secret_key = "T3st_s3cret_k3y!~$@"
else:
+ init_db()
app.secret_key = os.urandom(30)
+
+
+@app.teardown_request
+def remove_session(exception=None):
+ db_session.remove()
@app.route("/domain", methods=["GET", "POST"])
def domain():
+ ip = request.args.get("ip")
+ domain = request.args.get("domain")
+
+ if (ip and domain) is None:
+ return "parameter error", 400
+
+ r = Route(ip, domain)
+
if request.method == "GET":
# search domain
- pass
+ r.search()
elif request.method == "POST":
# register domain
- pass
+ user = register.args.get("user")
+
+ if user is not None:
+ r.register(user)
+ else:
+ return "user not found", 400
+
+ del r
+
+ return "{} - {}".format(ip, domain)
@app.route("/")
|
f4500e6422f1c6af8e9ce7d2d79d81e7479f0b7f
|
Instanssi/admin_programme/forms.py
|
Instanssi/admin_programme/forms.py
|
# -*- coding: utf-8 -*-
from django import forms
from uni_form.helper import FormHelper
from uni_form.layout import Submit, Layout, Fieldset, ButtonHolder
from Instanssi.ext_programme.models import ProgrammeEvent
class ProgrammeEventForm(forms.ModelForm):
def __init__(self, *args, **kwargs):
super(ProgrammeEventForm, self).__init__(*args, **kwargs)
self.helper = FormHelper()
self.helper.layout = Layout(
Fieldset(
u'',
'title',
'description',
'presenters',
'presenters_titles',
'icon_original',
'email',
'home_url',
'twitter_url',
'github_url',
'facebook_url',
'linkedin_url',
'wiki_url',
ButtonHolder (
Submit('submit', u'Tallenna')
)
)
)
class Meta:
model = ProgrammeEvent
exclude = ('event',)
|
# -*- coding: utf-8 -*-
from django import forms
from uni_form.helper import FormHelper
from uni_form.layout import Submit, Layout, Fieldset, ButtonHolder
from Instanssi.ext_programme.models import ProgrammeEvent
class ProgrammeEventForm(forms.ModelForm):
def __init__(self, *args, **kwargs):
super(ProgrammeEventForm, self).__init__(*args, **kwargs)
self.helper = FormHelper()
self.helper.layout = Layout(
Fieldset(
u'',
'title',
'description',
'start',
'end',
'presenters',
'presenters_titles',
'icon_original',
'email',
'home_url',
'twitter_url',
'github_url',
'facebook_url',
'linkedin_url',
'wiki_url',
ButtonHolder (
Submit('submit', u'Tallenna')
)
)
)
class Meta:
model = ProgrammeEvent
exclude = ('event','icon_small',)
|
Fix form to reflect model change
|
admin_programme: Fix form to reflect model change
|
Python
|
mit
|
Instanssi/Instanssi.org,Instanssi/Instanssi.org,Instanssi/Instanssi.org,Instanssi/Instanssi.org
|
---
+++
@@ -14,6 +14,8 @@
u'',
'title',
'description',
+ 'start',
+ 'end',
'presenters',
'presenters_titles',
'icon_original',
@@ -32,4 +34,4 @@
class Meta:
model = ProgrammeEvent
- exclude = ('event',)
+ exclude = ('event','icon_small',)
|
71cffcb8a8ec7e36dc389a5aa6dc2cc9769a9e97
|
distutils/tests/test_ccompiler.py
|
distutils/tests/test_ccompiler.py
|
import os
import sys
import platform
import textwrap
import sysconfig
import pytest
from distutils import ccompiler
def _make_strs(paths):
"""
Convert paths to strings for legacy compatibility.
"""
if sys.version_info > (3, 8) and platform.system() != "Windows":
return paths
return list(map(os.fspath, paths))
@pytest.fixture
def c_file(tmp_path):
c_file = tmp_path / 'foo.c'
gen_headers = ('Python.h',)
is_windows = platform.system() == "Windows"
plat_headers = ('windows.h',) * is_windows
all_headers = gen_headers + plat_headers
headers = '\n'.join(f'#include <{header}>\n' for header in all_headers)
payload = (
textwrap.dedent(
"""
#headers
void PyInit_foo(void) {}
"""
)
.lstrip()
.replace('#headers', headers)
)
c_file.write_text(payload)
return c_file
def test_set_include_dirs(c_file):
"""
Extensions should build even if set_include_dirs is invoked.
In particular, compiler-specific paths should not be overridden.
"""
compiler = ccompiler.new_compiler()
python = sysconfig.get_paths()['include']
compiler.set_include_dirs([python])
compiler.compile(_make_strs([c_file]))
|
import os
import sys
import platform
import textwrap
import sysconfig
import pytest
from distutils import ccompiler
def _make_strs(paths):
"""
Convert paths to strings for legacy compatibility.
"""
if sys.version_info > (3, 8) and platform.system() != "Windows":
return paths
return list(map(os.fspath, paths))
@pytest.fixture
def c_file(tmp_path):
c_file = tmp_path / 'foo.c'
gen_headers = ('Python.h',)
is_windows = platform.system() == "Windows"
plat_headers = ('windows.h',) * is_windows
all_headers = gen_headers + plat_headers
headers = '\n'.join(f'#include <{header}>\n' for header in all_headers)
payload = (
textwrap.dedent(
"""
#headers
void PyInit_foo(void) {}
"""
)
.lstrip()
.replace('#headers', headers)
)
c_file.write_text(payload)
return c_file
def test_set_include_dirs(c_file):
"""
Extensions should build even if set_include_dirs is invoked.
In particular, compiler-specific paths should not be overridden.
"""
compiler = ccompiler.new_compiler()
python = sysconfig.get_paths()['include']
compiler.set_include_dirs([python])
compiler.compile(_make_strs([c_file]))
# do it again, setting include dirs after any initialization
compiler.set_include_dirs([python])
compiler.compile(_make_strs([c_file]))
|
Extend the test to compile a second time after setting include dirs again.
|
Extend the test to compile a second time after setting include dirs again.
|
Python
|
mit
|
pypa/setuptools,pypa/setuptools,pypa/setuptools
|
---
+++
@@ -49,3 +49,7 @@
python = sysconfig.get_paths()['include']
compiler.set_include_dirs([python])
compiler.compile(_make_strs([c_file]))
+
+ # do it again, setting include dirs after any initialization
+ compiler.set_include_dirs([python])
+ compiler.compile(_make_strs([c_file]))
|
792b9d42167208b39a5e37f2ddf608547238af4f
|
django_graph_api/graphql/utils.py
|
django_graph_api/graphql/utils.py
|
from traceback import format_exc
from django.conf import settings
class GraphQLError(Exception):
def __init__(self, message):
super(GraphQLError, self).__init__(message)
self.message = message
self.location = {'line': 0, 'column': 0}
if settings.DEBUG:
print(format_exc().split('/n'))
def format(self):
return {'message': self.message}
def __eq__(self, other):
return self.__class__ == other.__class__ and self.message == other.message
def __hash__(self):
return super(GraphQLError, self).__hash__() + self.message.__hash__()
|
from traceback import format_exc
from django.conf import settings
class GraphQLError(Exception):
def __init__(self, message):
super(GraphQLError, self).__init__(message)
self.message = message
if settings.DEBUG:
self.traceback = format_exc().split('\n')
def format(self):
return {'message': self.message}
def __eq__(self, other):
return self.__class__ == other.__class__ and self.message == other.message
def __hash__(self):
return super(GraphQLError, self).__hash__() + self.message.__hash__()
|
Undo error changes for now
|
Undo error changes for now
|
Python
|
mit
|
melinath/django-graph-api,melinath/django-graph-api
|
---
+++
@@ -7,9 +7,8 @@
def __init__(self, message):
super(GraphQLError, self).__init__(message)
self.message = message
- self.location = {'line': 0, 'column': 0}
if settings.DEBUG:
- print(format_exc().split('/n'))
+ self.traceback = format_exc().split('\n')
def format(self):
return {'message': self.message}
|
08489ea2c1596a067b482878ff4450db43c08612
|
conf.py
|
conf.py
|
# -*- coding: utf-8 -*-
#
# on_rtd is whether we are on readthedocs.org
import os
on_rtd = os.environ.get('READTHEDOCS', None) == 'True'
if not on_rtd: # only import and set the theme if we're building docs locally
import sphinx_rtd_theme
html_theme = 'sphinx_rtd_theme'
html_theme_path = [sphinx_rtd_theme.get_html_theme_path()]
# otherwise, readthedocs.org uses their theme by default, so no need to specify it
project = 'FIWARE-Stream-Oriented-GE'
|
# -*- coding: utf-8 -*-
#
# on_rtd is whether we are on readthedocs.org
import os
on_rtd = os.environ.get('READTHEDOCS', None) == 'True'
if not on_rtd: # only import and set the theme if we're building docs locally
import sphinx_rtd_theme
html_theme = 'sphinx_rtd_theme'
html_theme_path = [sphinx_rtd_theme.get_html_theme_path()]
# otherwise, readthedocs.org uses their theme by default, so no need to specify it
project = 'FIWARE-Stream-Oriented-GE'
html_theme_options = {
'cssfiles': ['https://fiware.org/style/fiware_readthedocs.css']
}
|
Add custom CSS style to FIWARE doc
|
Add custom CSS style to FIWARE doc
Change-Id: I74293d488e0cd762ad023b94879ee618a4016110
|
Python
|
apache-2.0
|
Kurento/doc-kurento,SanMi86/doc-kurento,SanMi86/doc-kurento,SanMi86/doc-kurento,Kurento/doc-kurento,Kurento/doc-kurento,SanMi86/doc-kurento
|
---
+++
@@ -12,3 +12,6 @@
# otherwise, readthedocs.org uses their theme by default, so no need to specify it
project = 'FIWARE-Stream-Oriented-GE'
+html_theme_options = {
+ 'cssfiles': ['https://fiware.org/style/fiware_readthedocs.css']
+}
|
0fbc02b40f4414d96686d879aa9f7611e8fbb85d
|
singlet/config.py
|
singlet/config.py
|
# vim: fdm=indent
# author: Fabio Zanini
# date: 02/08/17
# content: Support module for filenames related to the Google Sheet APIs.
# Modules
import os
import yaml
# Globals
config_filename = os.getenv(
'SINGLET_CONFIG_FILENAME',
os.getenv('HOME') + '/.singlet/config.yml')
with open(config_filename) as stream:
config = yaml.load(stream)
# Warnings that should be seen only once
config['_once_warnings'] = []
config.reset_once_warings = lambda: config['_once_warnings'] = []
|
# vim: fdm=indent
# author: Fabio Zanini
# date: 02/08/17
# content: Support module for filenames related to the Google Sheet APIs.
# Modules
import os
import yaml
# Globals
config_filename = os.getenv(
'SINGLET_CONFIG_FILENAME',
os.getenv('HOME') + '/.singlet/config.yml')
with open(config_filename) as stream:
config = yaml.load(stream)
# Warnings that should be seen only once
config['_once_warnings'] = []
|
Remove function to reset _once_warnings (messy)
|
Remove function to reset _once_warnings (messy)
|
Python
|
mit
|
iosonofabio/singlet,iosonofabio/singlet
|
---
+++
@@ -16,4 +16,3 @@
# Warnings that should be seen only once
config['_once_warnings'] = []
-config.reset_once_warings = lambda: config['_once_warnings'] = []
|
ed7d0c5f8b64185f9fc612b44e4182b12a0fa62e
|
yunity/users/factories.py
|
yunity/users/factories.py
|
from django.contrib.auth import get_user_model
from factory import DjangoModelFactory, CREATE_STRATEGY, LazyAttribute, PostGeneration, SubFactory
from yunity.walls.factories import Wall
from yunity.utils.tests.fake import faker
class User(DjangoModelFactory):
class Meta:
model = get_user_model()
strategy = CREATE_STRATEGY
is_active = True
is_staff = False
display_name = LazyAttribute(lambda _: faker.name())
first_name = LazyAttribute(lambda _: faker.name())
last_name = LazyAttribute(lambda _: faker.name())
email = LazyAttribute(lambda _: faker.email())
password = PostGeneration(lambda obj, *args, **kwargs: obj.set_password(obj.display_name))
wall = SubFactory(Wall)
|
from django.contrib.auth import get_user_model
from factory import DjangoModelFactory, CREATE_STRATEGY, LazyAttribute, PostGeneration, SubFactory
from yunity.walls.factories import Wall
from yunity.utils.tests.fake import faker
class User(DjangoModelFactory):
class Meta:
model = get_user_model()
strategy = CREATE_STRATEGY
is_active = True
is_staff = False
display_name = LazyAttribute(lambda _: faker.name())
first_name = LazyAttribute(lambda _: faker.name())
last_name = LazyAttribute(lambda _: faker.name())
email = LazyAttribute(lambda _: faker.email())
# Use display_name as password, as it is readable
password = PostGeneration(lambda obj, *args, **kwargs: obj.set_password(obj.display_name))
wall = SubFactory(Wall)
|
Comment about display_name == password
|
Comment about display_name == password
|
Python
|
agpl-3.0
|
yunity/foodsaving-backend,yunity/yunity-core,yunity/foodsaving-backend,yunity/yunity-core,yunity/foodsaving-backend
|
---
+++
@@ -16,6 +16,8 @@
first_name = LazyAttribute(lambda _: faker.name())
last_name = LazyAttribute(lambda _: faker.name())
email = LazyAttribute(lambda _: faker.email())
+
+ # Use display_name as password, as it is readable
password = PostGeneration(lambda obj, *args, **kwargs: obj.set_password(obj.display_name))
wall = SubFactory(Wall)
|
aed9b3066f9d796e5c89e38d833c87e130a421c3
|
auth0/v2/blacklists.py
|
auth0/v2/blacklists.py
|
from .rest import RestClient
class Blacklists(object):
def __init__(self, domain, jwt_token):
url = 'https://%s/api/v2/blacklists/tokens' % domain
self.client = RestClient(endpoint=url, jwt=jwt_token)
def get(self, aud=None):
params = {
'aud': aud
}
return self.client.get(params=params)
def create(self, jti, aud=''):
return self.client.post(data={'jti': jti, 'aud': aud})
|
from .rest import RestClient
class Blacklists(object):
def __init__(self, domain, jwt_token):
self.url = 'https://%s/api/v2/blacklists/tokens' % domain
self.client = RestClient(jwt=jwt_token)
def get(self, aud=None):
params = {
'aud': aud
}
return self.client.get(self.url, params=params)
def create(self, jti, aud=''):
return self.client.post(self.url, data={'jti': jti, 'aud': aud})
|
Fix Blacklists usage of RestClient
|
Fix Blacklists usage of RestClient
|
Python
|
mit
|
auth0/auth0-python,auth0/auth0-python
|
---
+++
@@ -3,16 +3,15 @@
class Blacklists(object):
def __init__(self, domain, jwt_token):
- url = 'https://%s/api/v2/blacklists/tokens' % domain
-
- self.client = RestClient(endpoint=url, jwt=jwt_token)
+ self.url = 'https://%s/api/v2/blacklists/tokens' % domain
+ self.client = RestClient(jwt=jwt_token)
def get(self, aud=None):
params = {
'aud': aud
}
- return self.client.get(params=params)
+ return self.client.get(self.url, params=params)
def create(self, jti, aud=''):
- return self.client.post(data={'jti': jti, 'aud': aud})
+ return self.client.post(self.url, data={'jti': jti, 'aud': aud})
|
e49163ceecc5da949fe01281a87b56be513784d5
|
abbr/languages/pt_br/dictionary.py
|
abbr/languages/pt_br/dictionary.py
|
# Copyright 2016 Adler Brediks Medrado
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
word_list = {
'Professor': 'Prof.',
'Senhor': 'Sr.',
'Doutor': 'Dr.',
}
|
# Copyright 2016 Adler Brediks Medrado
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
word_list = {
'Professor': 'Prof.',
'Senhor': 'Sr.',
'Doutor': 'Dr.',
'Avenida': 'Av.',
'Travessa': 'Tv.',
'Número': 'Nº',
'Primeiro': '1º',
'Primeira': '1ª',
'Segundo': '2º',
'Segunda': '2ª',
}
|
Add new words to wordlist
|
Add new words to wordlist
|
Python
|
apache-2.0
|
adlermedrado/abbr
|
---
+++
@@ -16,4 +16,11 @@
'Professor': 'Prof.',
'Senhor': 'Sr.',
'Doutor': 'Dr.',
+ 'Avenida': 'Av.',
+ 'Travessa': 'Tv.',
+ 'Número': 'Nº',
+ 'Primeiro': '1º',
+ 'Primeira': '1ª',
+ 'Segundo': '2º',
+ 'Segunda': '2ª',
}
|
92e5ff34737feef0d196e25b97dbc817b502a59d
|
demo.py
|
demo.py
|
from FbFeed import NewsFeed
username = raw_input('Enter your email id registered with facebook : ')
password = raw_input('Enter your Password : ')
print('Creating new session on Firefox..')
fb = NewsFeed(username,password)
print('Logging into your facebook account')
fb.login()
#Add people to group
print('Add people to Feed Group')
count = int(raw_input('How many people would you like to add ?: '))
for i in range(count):
name = raw_input()
fb.add(name)
print('Leave me running on your system , I will notify you whenever these people perform any public activity')
fb.read()
|
from FbFeed import NewsFeed
import getpass
username = raw_input('Enter your email id registered with facebook : ')
password = getpass.getpass(prompt='Enter your Password : ',stream=None)
print('Creating new session on Firefox..')
fb = NewsFeed(username,password)
print('Logging into your facebook account')
fb.login()
#Add people to group
print('Add people to Feed Group')
count = int(raw_input('How many people would you like to add ?: '))
for i in range(count):
name = raw_input()
fb.add(name)
print('Leave me running on your system , I will notify you whenever these people perform any public activity')
fb.read()
|
Hide password in terminal input
|
Hide password in terminal input
|
Python
|
mit
|
ashishpahwa7/Fb-Feedirator
|
---
+++
@@ -1,8 +1,8 @@
from FbFeed import NewsFeed
-
+import getpass
username = raw_input('Enter your email id registered with facebook : ')
-password = raw_input('Enter your Password : ')
+password = getpass.getpass(prompt='Enter your Password : ',stream=None)
print('Creating new session on Firefox..')
fb = NewsFeed(username,password)
print('Logging into your facebook account')
|
deeaed14e40b9deca39c46ec7879f775606898c0
|
Instanssi/dblog/handlers.py
|
Instanssi/dblog/handlers.py
|
# -*- coding: utf-8 -*-
from logging import Handler
from datetime import datetime
class DBLogHandler(Handler, object):
def __init__(self):
super(DBLogHandler, self).__init__()
def emit(self, record):
from models import DBLogEntry as _LogEntry
entry = _LogEntry()
entry.level = record.levelname
entry.message = self.format(record)
if record.event:
entry.event = record.event
if record.user:
entry.user = record.user
entry.save()
|
# -*- coding: utf-8 -*-
from logging import Handler
from datetime import datetime
class DBLogHandler(Handler, object):
def __init__(self):
super(DBLogHandler, self).__init__()
def emit(self, record):
from models import DBLogEntry as _LogEntry
entry = _LogEntry()
entry.level = record.levelname
entry.message = self.format(record)
entry.module = record.name
try:
entry.event = record.event
except:
pass
try:
entry.user = record.user
except:
pass
entry.save()
|
Handle optional field saving with exceptions, save module name.
|
dblog: Handle optional field saving with exceptions, save module name.
|
Python
|
mit
|
Instanssi/Instanssi.org,Instanssi/Instanssi.org,Instanssi/Instanssi.org,Instanssi/Instanssi.org
|
---
+++
@@ -13,9 +13,14 @@
entry = _LogEntry()
entry.level = record.levelname
entry.message = self.format(record)
- if record.event:
+ entry.module = record.name
+ try:
entry.event = record.event
- if record.user:
+ except:
+ pass
+ try:
entry.user = record.user
+ except:
+ pass
entry.save()
|
407a032acb307e5f936437aec4975ef69133d0c5
|
DisplayAdapter/testing/test_display_adapter/test_display_driver/test_display_drivers.py
|
DisplayAdapter/testing/test_display_adapter/test_display_driver/test_display_drivers.py
|
"""
This module contains the testing framework for the display driver functionality, and is responsible for testing whether
the pi can correctly and sufficiently connect to the display.
"""
from mock import patch
from display_adapter.display_driver.display_drivers import DisplayDriver
class TestDisplayDriver(object):
"""
This class tests the functionality of the DisplayDriver class, ensuring that it can correctly connect the raspberry
pi to the display.
"""
@patch("serial.Serial")
def test_init(self, dc_mock):
"""
This method tests initialisation of the display driver, ensuring it has correctly set up the database helper
and the display controller.
"""
dd = DisplayDriver()
assert dd
assert hasattr(dd, "_db_helper")
assert hasattr(dd, "_display_controller")
|
"""
This module contains the testing framework for the display driver functionality, and is responsible for testing whether
the pi can correctly and sufficiently connect to the display.
"""
from mock import patch
from datetime import datetime
from display_adapter.display_driver.display_drivers import minutify, DisplayDriver
class TestDisplayDriver(object):
"""
This class tests the functionality of the DisplayDriver class, ensuring that it can correctly connect the raspberry
pi to the display.
"""
@patch("serial.Serial")
def test_init(self, dc_mock):
"""
This method tests initialisation of the display driver, ensuring it has correctly set up the database helper
and the display controller.
"""
dd = DisplayDriver()
assert dd
assert hasattr(dd, "_db_helper")
assert hasattr(dd, "_display_controller")
def test_minutify():
"""
This function tests the functionality of the minutify function linked to the Display Driver. The expected result
of this test is for a datetime object to be made accurate to the minute.
"""
dt = datetime.now().replace(second=30, microsecond=40000)
accurate_dt = minutify(dt)
# Assert the datetime object has been minutified correctly (seconds and microseconds == 0)
assert accurate_dt.second == 0 and accurate_dt.microsecond == 0
|
Test functionality has been added. Paired by Richard and Michael.
|
Test functionality has been added. Paired by Richard and Michael.
The functionality for the minutify function has now been tested; and the tests work (Support 231)
|
Python
|
mit
|
CO600GOL/Game_of_life,CO600GOL/Game_of_life,CO600GOL/Game_of_life
|
---
+++
@@ -4,7 +4,8 @@
"""
from mock import patch
-from display_adapter.display_driver.display_drivers import DisplayDriver
+from datetime import datetime
+from display_adapter.display_driver.display_drivers import minutify, DisplayDriver
class TestDisplayDriver(object):
"""
@@ -23,3 +24,15 @@
assert dd
assert hasattr(dd, "_db_helper")
assert hasattr(dd, "_display_controller")
+
+
+def test_minutify():
+ """
+ This function tests the functionality of the minutify function linked to the Display Driver. The expected result
+ of this test is for a datetime object to be made accurate to the minute.
+ """
+
+ dt = datetime.now().replace(second=30, microsecond=40000)
+ accurate_dt = minutify(dt)
+ # Assert the datetime object has been minutified correctly (seconds and microseconds == 0)
+ assert accurate_dt.second == 0 and accurate_dt.microsecond == 0
|
3e4094d838df26afadfa7074a16c6a21f06ba6b8
|
bamp/logs.py
|
bamp/logs.py
|
import logging
class ExceptionFilter(logging.Filter):
def __init__(self, debug=None):
self.debug = debug
def filter(self, record):
if self.debug:
return True
# clear exceptions when not in debug
if record.levelname == 'ERROR' and record.exc_info:
record.exc_info = None
return True
LOGGING = {
'version': 1,
'filters': {
'exc_filter': {
'()': ExceptionFilter,
'debug': False
}
},
'handlers': {
'console': {
'class': 'logging.StreamHandler',
'filters': ['exc_filter']
}
},
'loggers': {
'bamp': {
'level': 'DEBUG',
'handlers': ['console']
},
}
}
|
import logging
class DebugFilter(logging.Filter):
def __init__(self, debug=None):
self.debug = debug
def filter(self, record):
if self.debug:
return True
return False
LOGGING = {
'version': 1,
'filters': {
'exc_filter': {
'()': DebugFilter,
'debug': False
}
},
'handlers': {
'console': {
'class': 'logging.StreamHandler',
'filters': ['exc_filter']
}
},
'loggers': {
'bamp': {
'level': 'DEBUG',
'handlers': ['console']
},
}
}
|
Disable printing error messages not in debug
|
Disable printing error messages not in debug
|
Python
|
mit
|
inirudebwoy/bamp
|
---
+++
@@ -1,24 +1,20 @@
import logging
-class ExceptionFilter(logging.Filter):
+class DebugFilter(logging.Filter):
def __init__(self, debug=None):
self.debug = debug
def filter(self, record):
if self.debug:
return True
-
- # clear exceptions when not in debug
- if record.levelname == 'ERROR' and record.exc_info:
- record.exc_info = None
- return True
+ return False
LOGGING = {
'version': 1,
'filters': {
'exc_filter': {
- '()': ExceptionFilter,
+ '()': DebugFilter,
'debug': False
}
},
|
61d7c9e99398874745d11720cd8d985bdc3d7514
|
demoapp/views.py
|
demoapp/views.py
|
from demoapp.forms import DemoLoginForm
from django.shortcuts import render_to_response
from django.shortcuts import redirect
from demoapp import app_settings
from demoapp.utils import get_salt
def login_view(request):
if request.method == 'POST':
form = DemoLoginForm(request.POST)
if form.is_valid():
response = redirect('/')
response.set_signed_cookie(app_settings.COOKIE_NAME, 'demo access granted', salt=get_salt(request))
return response
else:
form = DemoLoginForm()
return render_to_response('demoapp/login.html', {'form': form})
|
from demoapp.forms import DemoLoginForm
from django.shortcuts import render
from django.shortcuts import redirect
from demoapp import app_settings
from demoapp.utils import get_salt
def login_view(request):
if request.method == 'POST':
form = DemoLoginForm(request.POST)
if form.is_valid():
response = redirect('/')
response.set_signed_cookie(app_settings.COOKIE_NAME, 'demo access granted', salt=get_salt(request))
return response
else:
form = DemoLoginForm()
return render(request, 'demoapp/login.html', {'form': form})
|
Add context to login view
|
Add context to login view
|
Python
|
unlicense
|
dboczek/django-demo,dboczek/django-demo
|
---
+++
@@ -1,5 +1,5 @@
from demoapp.forms import DemoLoginForm
-from django.shortcuts import render_to_response
+from django.shortcuts import render
from django.shortcuts import redirect
from demoapp import app_settings
from demoapp.utils import get_salt
@@ -14,4 +14,4 @@
return response
else:
form = DemoLoginForm()
- return render_to_response('demoapp/login.html', {'form': form})
+ return render(request, 'demoapp/login.html', {'form': form})
|
fc3408e0d8336ca2324b272dbb4aa0e69914a27c
|
build_chrome_webapp.py
|
build_chrome_webapp.py
|
import os.path
from shutil import copyfile
try:
from jinja2 import Template
except:
print "Could not import Jinja2, run 'easy_install Jinja2'"
exit()
output_dir = os.path.join('./', 'chrome_webstore')
if not os.path.exists(output_dir):
os.makedirs(output_dir)
def add_background_script():
copyfile('chrome_webstore_background.js', os.path.join(output_dir, 'background.js'))
def add_manifest():
copyfile('chrome_webstore_manifest.json', os.path.join(output_dir, 'manifest.json'))
def render_main_template():
f = open('./html/index.html')
template = Template(f.read().decode('utf-8'))
f.close()
html = template.render(og_tag='', url='', ON_PRODUCTION=True, ON_DEV=False, USE_PRODUCTION_JAVASCRIPT=True)
f = open(os.path.join(output_dir, 'index.html'), 'w')
f.write(html.encode('utf-8'))
f.close()
print "Template rendered"
add_manifest()
add_background_script()
render_main_template()
|
import os.path
from shutil import copyfile
from shutil import copytree
from shutil import rmtree
try:
from jinja2 import Template
except:
print "Could not import Jinja2, run 'easy_install Jinja2'"
exit()
output_dir = os.path.join('./', 'chrome_webstore')
if os.path.exists(output_dir):
rmtree(output_dir)
os.makedirs(output_dir)
def add_background_script():
copyfile('chrome_webstore_background.js', os.path.join(output_dir, 'background.js'))
def copy_static_dirs():
copytree('images', os.path.join(output_dir, 'images'))
copytree('styles', os.path.join(output_dir, 'styles'))
copytree('scripts', os.path.join(output_dir, 'scripts'))
def add_manifest():
copyfile('chrome_webstore_manifest.json', os.path.join(output_dir, 'manifest.json'))
def render_main_template():
f = open('./html/index.html')
template = Template(f.read().decode('utf-8'))
f.close()
html = template.render(og_tag='', url='', ON_PRODUCTION=True, ON_DEV=False, USE_PRODUCTION_JAVASCRIPT=True)
f = open(os.path.join(output_dir, 'index.html'), 'w')
f.write(html.encode('utf-8'))
f.close()
print "Template rendered"
add_manifest()
add_background_script()
render_main_template()
copy_static_dirs()
|
Copy static dirs as well
|
Copy static dirs as well
|
Python
|
mit
|
youtify/youtify,youtify/youtify,youtify/youtify
|
---
+++
@@ -1,5 +1,7 @@
import os.path
from shutil import copyfile
+from shutil import copytree
+from shutil import rmtree
try:
from jinja2 import Template
except:
@@ -8,11 +10,18 @@
output_dir = os.path.join('./', 'chrome_webstore')
-if not os.path.exists(output_dir):
- os.makedirs(output_dir)
+if os.path.exists(output_dir):
+ rmtree(output_dir)
+
+os.makedirs(output_dir)
def add_background_script():
copyfile('chrome_webstore_background.js', os.path.join(output_dir, 'background.js'))
+
+def copy_static_dirs():
+ copytree('images', os.path.join(output_dir, 'images'))
+ copytree('styles', os.path.join(output_dir, 'styles'))
+ copytree('scripts', os.path.join(output_dir, 'scripts'))
def add_manifest():
copyfile('chrome_webstore_manifest.json', os.path.join(output_dir, 'manifest.json'))
@@ -32,3 +41,4 @@
add_manifest()
add_background_script()
render_main_template()
+copy_static_dirs()
|
5c1615d9e3e16c1f48be41733997d8c39355bcb4
|
django_distributed_queue/utils.py
|
django_distributed_queue/utils.py
|
class LazyModel(object):
"""
It's a helper class that is used in case you have model_class and object
primary key. You might need to use only object id.
If you try to access other fields of the model class then we will query the
database to get that object and provide you with any field and method
transparently proxying them.
"""
def __init__(self, model_class, pk):
self._model_class = model_class
self.pk = pk
self._instance = None
def __getattribute__(self, attr_name):
# Hiding traces of decoration.
if attr_name in ('__init__', '__getattribute__', '_model_class', 'pk',
'_instance'):
# Stopping recursion.
return object.__getattribute__(self, attr_name)
# All other attr_names, including auto-defined by system in self, are
# searched in decorated self.instance, e.g.: __module__, __class__, etc.
if self._instance is None:
self._instance = self._model_class.objects.get(pk=self.pk)
# Raises correct AttributeError if name is not found in decorated self.func.
return getattr(self.instance, attr_name)
|
class LazyModel(object):
"""
It's a helper class that is used in case you have model_class and object
primary key. You might need to use only object id.
If you try to access other fields of the model class then we will query the
database to get that object and provide you with any field and method
transparently proxying them.
"""
def __init__(self, model_class, pk):
self._model_class = model_class
self.pk = pk
self._instance = None
def __getattribute__(self, attr_name):
# Hiding traces of decoration.
if attr_name in ('__init__', '__getattribute__', '_model_class', 'pk',
'_instance'):
# Stopping recursion.
return object.__getattribute__(self, attr_name)
# All other attr_names, including auto-defined by system in self, are
# searched in decorated self.instance, e.g.: __module__, __class__, etc.
if self._instance is None:
self._instance = self._model_class.objects.get(pk=self.pk)
# Raises correct AttributeError if name is not found in decorated self.func.
return getattr(self._instance, attr_name)
|
Fix for LazyModel infinite recursion
|
Fix for LazyModel infinite recursion
|
Python
|
mit
|
ProstoKSI/django-distributed-queue
|
---
+++
@@ -23,4 +23,4 @@
if self._instance is None:
self._instance = self._model_class.objects.get(pk=self.pk)
# Raises correct AttributeError if name is not found in decorated self.func.
- return getattr(self.instance, attr_name)
+ return getattr(self._instance, attr_name)
|
df1617a7518f66d87470f948e057e4d7d7d8f026
|
driller/tasks.py
|
driller/tasks.py
|
import redis
from celery import Celery
from .driller import Driller
app = Celery('tasks', broker='amqp://guest@localhost//', backend='redis://localhost')
redis_pool = redis.ConnectionPool(host='localhost', port=6379, db=1)
@app.task
def drill(binary, input, fuzz_bitmap, qemu_dir):
redis_inst = redis.Redis(connection_pool=redis_pool)
driller = Driller(binary, input, fuzz_bitmap, qemu_dir, redis=redis_inst)
return driller.drill()
|
import redis
from celery import Celery
from .driller import Driller
import config
backend_url = "redis://%s:%d" % (config.REDIS_HOST, config.REDIS_PORT)
app = Celery('tasks', broker=config.BROKER_URL, backend=backend_url)
redis_pool = redis.ConnectionPool(host=config.REDIS_HOST, port=config.REDIS_PORT, db=config.REDIS_DB)
@app.task
def drill(binary, input, fuzz_bitmap, qemu_dir):
redis_inst = redis.Redis(connection_pool=redis_pool)
driller = Driller(binary, input, fuzz_bitmap, qemu_dir, redis=redis_inst)
return driller.drill()
|
Connect to Celery using config options
|
Connect to Celery using config options
|
Python
|
bsd-2-clause
|
shellphish/driller
|
---
+++
@@ -2,9 +2,11 @@
from celery import Celery
from .driller import Driller
+import config
-app = Celery('tasks', broker='amqp://guest@localhost//', backend='redis://localhost')
-redis_pool = redis.ConnectionPool(host='localhost', port=6379, db=1)
+backend_url = "redis://%s:%d" % (config.REDIS_HOST, config.REDIS_PORT)
+app = Celery('tasks', broker=config.BROKER_URL, backend=backend_url)
+redis_pool = redis.ConnectionPool(host=config.REDIS_HOST, port=config.REDIS_PORT, db=config.REDIS_DB)
@app.task
def drill(binary, input, fuzz_bitmap, qemu_dir):
|
a85d148eb00f83052a97d66da8ff9dd79b40f172
|
.ycm_extra_conf.py
|
.ycm_extra_conf.py
|
import os
def FlagsForFile(filename, **kwargs):
flags = ['-std=c++14', '-I/usr/local/include']
proj_root = os.path.dirname(os.path.abspath(__file__))
libcanon_include = ''.join(['-I', proj_root, '/deps/libcanon/include'])
proj_include = ''.join(['-I', proj_root, '/drudge'])
flags.extend([libcanon_include, proj_include])
return {'flags': flags}
|
import os
import subprocess
def FlagsForFile(filename, **kwargs):
flags = ['-std=c++14', '-I/usr/local/include']
proj_root = os.path.dirname(os.path.abspath(__file__))
libcanon_include = ''.join(['-I', proj_root, '/deps/libcanon/include'])
python_include = subprocess.run(
["pkg-config", '--cflags', 'python3'], stdout=subprocess.PIPE
).stdout.decode("utf-8")
proj_include = ''.join(['-I', proj_root, '/drudge'])
flags.extend([libcanon_include, proj_include])
return {'flags': flags}
|
Add Python inclusion path to YCM config
|
Add Python inclusion path to YCM config
In the script, the path is read from the result from pkg-config. So it
should work in most places.
|
Python
|
mit
|
tschijnmo/drudge,tschijnmo/drudge,tschijnmo/drudge
|
---
+++
@@ -1,4 +1,5 @@
import os
+import subprocess
def FlagsForFile(filename, **kwargs):
@@ -6,6 +7,9 @@
proj_root = os.path.dirname(os.path.abspath(__file__))
libcanon_include = ''.join(['-I', proj_root, '/deps/libcanon/include'])
+ python_include = subprocess.run(
+ ["pkg-config", '--cflags', 'python3'], stdout=subprocess.PIPE
+ ).stdout.decode("utf-8")
proj_include = ''.join(['-I', proj_root, '/drudge'])
flags.extend([libcanon_include, proj_include])
|
238da6f5cb5409409f54980f4ce018fda897a766
|
API/chat/models.py
|
API/chat/models.py
|
from django.db import models
class Channel(models.Model):
def __str__(self):
return self.name
name = models.CharField(max_length=20, unique=True)
class Message(models.Model):
def __str__(self):
return self.text
def to_dict(self):
serializable_fields = ('text', 'datetime_start', 'datetime_sent', 'username')
return {key: getattr(self, key) for key in serializable_fields}
text = models.TextField(max_length=2000)
datetime_start = models.DateTimeField(default=None)
datetime_sent = models.DateTimeField(default=None, null=True)
typing = models.BooleanField(default=False)
username = models.CharField(max_length=20)
channel = models.ForeignKey(Channel)
|
from django.db import models
class Channel(models.Model):
def __str__(self):
return self.name
name = models.CharField(max_length=20, unique=True)
class Message(models.Model):
def __str__(self):
return self.text
def to_dict(self):
serializable_fields = ('text', 'datetime_start', 'datetime_sent', 'username')
return {key: getattr(self, key) for key in serializable_fields}
TEXT = 'text'
IMAGE = 'image'
MESSAGE_TYPE = (
(TEXT, 'text'),
(IMAGE, 'image'),
)
text = models.TextField(max_length=2000)
datetime_start = models.DateTimeField(default=None)
datetime_sent = models.DateTimeField(default=None, null=True)
typing = models.BooleanField(default=False)
username = models.CharField(max_length=20)
channel = models.ForeignKey(Channel)
message_type = models.CharField(max_length=10,
choices=MESSAGE_TYPE,
default=TEXT)
|
Add message_type field into message model
|
Add message_type field into message model
|
Python
|
mit
|
dionyziz/ting,gtklocker/ting,dionyziz/ting,gtklocker/ting,gtklocker/ting,dionyziz/ting,mbalamat/ting,mbalamat/ting,gtklocker/ting,dionyziz/ting,mbalamat/ting,mbalamat/ting
|
---
+++
@@ -16,9 +16,20 @@
serializable_fields = ('text', 'datetime_start', 'datetime_sent', 'username')
return {key: getattr(self, key) for key in serializable_fields}
+ TEXT = 'text'
+ IMAGE = 'image'
+
+ MESSAGE_TYPE = (
+ (TEXT, 'text'),
+ (IMAGE, 'image'),
+ )
+
text = models.TextField(max_length=2000)
datetime_start = models.DateTimeField(default=None)
datetime_sent = models.DateTimeField(default=None, null=True)
typing = models.BooleanField(default=False)
username = models.CharField(max_length=20)
channel = models.ForeignKey(Channel)
+ message_type = models.CharField(max_length=10,
+ choices=MESSAGE_TYPE,
+ default=TEXT)
|
7d898ec04733d25c1df33c8faf151f2b42a69ec9
|
base/components/people/constants.py
|
base/components/people/constants.py
|
from model_utils import Choices
from ohashi.constants import OTHER
BLOOD_TYPE = Choices('A', 'B', 'O', 'AB')
CLASSIFICATIONS = Choices(
(1, 'major', 'Major Unit'),
(2, 'minor', 'Minor Unit'),
(4, 'temporary', 'Temporary Unit'),
(5, 'subunit', 'Sub-Unit'),
(7, 'supergroup', 'Supergroup'),
('Special Units', [
(3, 'shuffle', 'Shuffle Unit'),
(6, 'revival', 'Revival Unit'),
(8, 'satoyama', 'Satoyama Unit'),
]),
(OTHER, 'other', 'Other')
)
PHOTO_SOURCES = Choices(
(1, 'promotional', 'Promotional Photo'),
(2, 'blog', 'Blog Photo'),
(OTHER, 'other', 'Other')
)
SCOPE = Choices(
(1, 'hp', 'Hello! Project'),
(2, 'ufa', 'Up Front Agency'),
(OTHER, 'other', 'Other')
)
STATUS = Choices(
(1, 'active', 'Active'),
(2, 'former', 'Former'),
(OTHER, 'other', 'Other')
)
|
from model_utils import Choices
from ohashi.constants import OTHER
BLOOD_TYPE = Choices('A', 'B', 'O', 'AB')
CLASSIFICATIONS = Choices(
(1, 'major', 'Major Unit'),
(2, 'minor', 'Minor Unit'),
(4, 'temporary', 'Temporary Unit'),
(5, 'subunit', 'Sub-Unit'),
(7, 'supergroup', 'Supergroup'),
('Special Units', [
(3, 'shuffle', 'Shuffle Unit'),
(6, 'revival', 'Revival Unit'),
(8, 'satoyama', 'Satoyama Unit'),
(9, 'satoumi', 'Satoumi Unit'),
]),
(OTHER, 'other', 'Other')
)
PHOTO_SOURCES = Choices(
(1, 'promotional', 'Promotional Photo'),
(2, 'blog', 'Blog Photo'),
(OTHER, 'other', 'Other')
)
SCOPE = Choices(
(1, 'hp', 'Hello! Project'),
(2, 'ufa', 'Up Front Agency'),
(OTHER, 'other', 'Other')
)
STATUS = Choices(
(1, 'active', 'Active'),
(2, 'former', 'Former'),
(OTHER, 'other', 'Other')
)
|
Add Satoumi as a classification.
|
Add Satoumi as a classification.
|
Python
|
apache-2.0
|
hello-base/web,hello-base/web,hello-base/web,hello-base/web
|
---
+++
@@ -15,6 +15,7 @@
(3, 'shuffle', 'Shuffle Unit'),
(6, 'revival', 'Revival Unit'),
(8, 'satoyama', 'Satoyama Unit'),
+ (9, 'satoumi', 'Satoumi Unit'),
]),
(OTHER, 'other', 'Other')
)
|
f828ac9ee5082a9a0b5e215c4c814e7f35db11b6
|
planetstack/core/models/__init__.py
|
planetstack/core/models/__init__.py
|
from .plcorebase import PlCoreBase
from .planetstack import PlanetStack
from .project import Project
from .singletonmodel import SingletonModel
from .service import Service
from .service import ServiceAttribute
from .tag import Tag
from .role import Role
from .site import Site,Deployment, DeploymentRole, DeploymentPrivilege, SiteDeployments
from .dashboard import DashboardView
from .user import User, UserDashboardView
from .serviceclass import ServiceClass
from .slice import Slice, SliceDeployments
from .site import SitePrivilege, SiteDeployments
from .userdeployments import UserDeployments
from .image import Image, ImageDeployments
from .node import Node
from .serviceresource import ServiceResource
from .slice import SliceRole
from .slice import SlicePrivilege
from .site import SiteRole
from .site import SitePrivilege
from .planetstack import PlanetStackRole
from .planetstack import PlanetStackPrivilege
from .slicetag import SliceTag
from .flavor import Flavor
from .sliver import Sliver
from .reservation import ReservedResource
from .reservation import Reservation
from .network import Network, NetworkParameterType, NetworkParameter, NetworkSliver, NetworkTemplate, Router, NetworkSlice, NetworkDeployments
from .billing import Account, Invoice, Charge, UsableObject, Payment
|
from .plcorebase import PlCoreBase
from .planetstack import PlanetStack
from .project import Project
from .singletonmodel import SingletonModel
from .service import Service
from .service import ServiceAttribute
from .tag import Tag
from .role import Role
from .site import Site,Deployment, DeploymentRole, DeploymentPrivilege, SiteDeployments
from .dashboard import DashboardView
from .user import User, UserDashboardView
from .serviceclass import ServiceClass
from .slice import Slice, SliceDeployments
from .site import SitePrivilege, SiteDeployments
from .userdeployments import UserDeployments
from .image import Image, ImageDeployments
from .node import Node
from .serviceresource import ServiceResource
from .slice import SliceRole
from .slice import SlicePrivilege
from .credential import UserCredential,SiteCredential,SliceCredential
from .site import SiteRole
from .site import SitePrivilege
from .planetstack import PlanetStackRole
from .planetstack import PlanetStackPrivilege
from .slicetag import SliceTag
from .flavor import Flavor
from .sliver import Sliver
from .reservation import ReservedResource
from .reservation import Reservation
from .network import Network, NetworkParameterType, NetworkParameter, NetworkSliver, NetworkTemplate, Router, NetworkSlice, NetworkDeployments
from .billing import Account, Invoice, Charge, UsableObject, Payment
|
Add credentials module to core list
|
Add credentials module to core list
|
Python
|
apache-2.0
|
wathsalav/xos,wathsalav/xos,wathsalav/xos,wathsalav/xos
|
---
+++
@@ -18,6 +18,7 @@
from .serviceresource import ServiceResource
from .slice import SliceRole
from .slice import SlicePrivilege
+from .credential import UserCredential,SiteCredential,SliceCredential
from .site import SiteRole
from .site import SitePrivilege
from .planetstack import PlanetStackRole
|
644660b6c41f029f271a0b8866387f358f8fdf54
|
frappe/patches/v4_0/enable_scheduler_in_system_settings.py
|
frappe/patches/v4_0/enable_scheduler_in_system_settings.py
|
# Copyright (c) 2013, Web Notes Technologies Pvt. Ltd. and Contributors
# MIT License. See license.txt
from __future__ import unicode_literals
import frappe
from frappe.utils.scheduler import disable_scheduler, enable_scheduler
def execute():
frappe.reload_doc("core", "doctype", "system_settings")
if frappe.db.get_global("disable_scheduler"):
disable_scheduler()
else:
enable_scheduler()
|
# Copyright (c) 2013, Web Notes Technologies Pvt. Ltd. and Contributors
# MIT License. See license.txt
from __future__ import unicode_literals
import frappe
from frappe.utils.scheduler import disable_scheduler, enable_scheduler
from frappe.utils import cint
def execute():
frappe.reload_doc("core", "doctype", "system_settings")
if cint(frappe.db.get_global("disable_scheduler")):
disable_scheduler()
else:
enable_scheduler()
|
Fix in enable scheduler patch
|
Fix in enable scheduler patch
|
Python
|
mit
|
BhupeshGupta/frappe,letzerp/framework,saurabh6790/frappe,rmehta/frappe,elba7r/builder,suyashphadtare/sajil-frappe,rohitw1991/frappe,saguas/frappe,indictranstech/tele-frappe,nerevu/frappe,indictranstech/omnitech-frappe,vCentre/vFRP-6233,gangadharkadam/saloon_frappe,aboganas/frappe,indictranstech/phr-frappe,gangadharkadam/letzfrappe,hatwar/buyback-frappe,mbauskar/tele-frappe,erpletzerp/letzerpcore,Amber-Creative/amber-frappe,indictranstech/ebuy-now-frappe,elba7r/frameworking,vjFaLk/frappe,shitolepriya/test-frappe,mbauskar/Das_frappe,frappe/frappe,gangadharkadam/v4_frappe,gangadhar-kadam/helpdesk-frappe,vCentre/vFRP-6233,RicardoJohann/frappe,rohitw1991/frappe,ESS-LLP/frappe,nerevu/frappe,gangadharkadam/stfrappe,sbktechnology/trufil-frappe,rohitwaghchaure/frappe_smart,aboganas/frappe,StrellaGroup/frappe,manassolanki/frappe,drukhil/frappe,sbktechnology/sap_frappe,rmehta/frappe,indictranstech/Das_frappe,letzerp/framework,erpletzerp/letzerpcore,MaxMorais/frappe,drukhil/frappe,rohitwaghchaure/frappe-alec,ashokrajbathu/secondrep,hernad/frappe,jevonearth/frappe,suyashphadtare/propshikhari-frappe,BhupeshGupta/frappe,gangadharkadam/v4_frappe,indictranstech/fbd_frappe,gangadharkadam/tailorfrappe,hernad/frappe,shitolepriya/test-frappe,bcornwellmott/frappe,gangadhar-kadam/lgnlvefrape,sbkolate/sap_frappe_v6,indautgrp/frappe,indictranstech/frappe,mbauskar/omnitech-frappe,indictranstech/frappe,praba230890/frappe,gangadharkadam/v6_frappe,indictranstech/osmosis-frappe,pawaranand/phr-frappe,saurabh6790/frappe,rohitwaghchaure/frappe_smart,gangadharkadam/v4_frappe,indictranstech/osmosis-frappe,gangadharkadam/v6_frappe,gangadhar-kadam/verve_live_frappe,adityahase/frappe,bcornwellmott/frappe,almeidapaulopt/frappe,indictranstech/frappe-digitales,indictranstech/reciphergroup-frappe,mbauskar/omnitech-frappe,gangadharkadam/saloon_frappe,gangadhar-kadam/laganfrappe,gangadharkadam/vervefrappe,neilLasrado/frappe,mbauskar/frappe,gangadhar-kadam/verve_live_frappe,MaxMorais/frappe,saurabh6790/test-frappe,rohitw1991/smarttailorfrappe,StrellaGroup/frappe,gangadharkadam/vervefrappe,rohitwaghchaure/frappe-digitales,elba7r/builder,mbauskar/helpdesk-frappe,deveninfotech/deven-frappe,paurosello/frappe,mbauskar/frappe,adityahase/frappe,gangadhar-kadam/lgnlvefrape,indictranstech/trufil-frappe,indictranstech/omnitech-frappe,saurabh6790/test-frappe,ESS-LLP/frappe,RicardoJohann/frappe,pombredanne/frappe,mbauskar/phr-frappe,tundebabzy/frappe,tmimori/frappe,paurosello/frappe,anandpdoshi/frappe,neilLasrado/frappe,maxtorete/frappe,gangadharkadam/frappecontribution,gangadharkadam/frappecontribution,nerevu/frappe,maxtorete/frappe,Tejal011089/digitales_frappe,pawaranand/phr-frappe,saguas/frappe,mbauskar/omnitech-demo-frappe,sbkolate/sap_frappe_v6,sbktechnology/sap_frappe,indictranstech/trufil-frappe,gangadhar-kadam/smrterpfrappe,suyashphadtare/sajil-final-frappe,indictranstech/internal-frappe,pawaranand/phr-frappe,saguas/frappe,erpletzerp/letzerpcore,praba230890/frappe,Amber-Creative/amber-frappe,gangadhar-kadam/lgnlvefrape,indictranstech/osmosis-frappe,ShashaQin/frappe,indautgrp/frappe,gangadharkadam/saloon_frappe,indictranstech/phr-frappe,indictranstech/omnitech-frappe,suyashphadtare/propshikhari-frappe,frappe/frappe,nerevu/frappe,hatwar/buyback-frappe,sbktechnology/trufil-frappe,mbauskar/tele-frappe,aboganas/frappe,indictranstech/frappe-digitales,gangadhar-kadam/helpdesk-frappe,vjFaLk/frappe,almeidapaulopt/frappe,StrellaGroup/frappe,mbauskar/phr-frappe,mbauskar/helpdesk-frappe,indictranstech/frappe-digitales,tundebabzy/frappe,rohitwaghchaure/New_Theme_frappe,indictranstech/trufil-frappe,sbkolate/sap_frappe_v6,hernad/frappe,suyashphadtare/propshikhari-frappe,gangadharkadam/vlinkfrappe,geo-poland/frappe,indictranstech/omnitech-frappe,gangadhar-kadam/smrterpfrappe,gangadhar-kadam/verve_test_frappe,gangadhar-kadam/verve_test_frappe,rohitwaghchaure/frappe,gangadharkadam/office_frappe,gangadharkadam/letzfrappe,chdecultot/frappe,gangadharkadam/letzfrappe,gangadharkadam/saloon_frappe_install,indautgrp/frappe,MaxMorais/frappe,maxtorete/frappe,mhbu50/frappe,vCentre/vFRP-6233,chdecultot/frappe,almeidapaulopt/frappe,indictranstech/tele-frappe,jevonearth/frappe,hatwar/buyback-frappe,gangadharkadam/letzfrappe,mbauskar/Das_frappe,gangadharkadam/v6_frappe,saurabh6790/phr-frappe,gangadhar-kadam/verve_live_frappe,gangadharkadam/stfrappe,indictranstech/Das_frappe,gangadhar-kadam/helpdesk-frappe,mbauskar/phr-frappe,indictranstech/phr-frappe,aboganas/frappe,gangadharkadam/saloon_frappe_install,indictranstech/frappe,rohitwaghchaure/frappe-alec,paurosello/frappe,saurabh6790/phr-frappe,mbauskar/tele-frappe,mbauskar/omnitech-demo-frappe,bohlian/frappe,gangadhar-kadam/verve_frappe,paurosello/frappe,BhupeshGupta/frappe,ESS-LLP/frappe,mhbu50/frappe,chdecultot/frappe,rohitwaghchaure/frappe-alec,gangadhar-kadam/helpdesk-frappe,indictranstech/fbd_frappe,pranalik/frappe-bb,geo-poland/frappe,Tejal011089/digitales_frappe,rohitwaghchaure/vestasi-frappe,mhbu50/frappe,gangadharkadam/shfr,pombredanne/frappe,bohlian/frappe,pawaranand/phr_frappe,gangadhar-kadam/laganfrappe,bohlian/frappe,manassolanki/frappe,indictranstech/trufil-frappe,maxtorete/frappe,jevonearth/frappe,mbauskar/Das_frappe,rohitwaghchaure/frappe-digitales,tmimori/frappe,vqw/frappe,vjFaLk/frappe,saurabh6790/test-frappe,indictranstech/osmosis-frappe,mbauskar/helpdesk-frappe,pombredanne/frappe,vqw/frappe,bohlian/frappe,tmimori/frappe,letzerp/framework,ashokrajbathu/secondrep,geo-poland/frappe,RicardoJohann/frappe,manassolanki/frappe,bcornwellmott/frappe,gangadhar-kadam/verve_frappe,hatwar/buyback-frappe,pranalik/frappe-bb,gangadhar-kadam/verve_test_frappe,yashodhank/frappe,neilLasrado/frappe,gangadharkadam/smrtfrappe,rohitwaghchaure/vestasi-frappe,shitolepriya/test-frappe,indictranstech/ebuy-now-frappe,elba7r/builder,gangadharkadam/office_frappe,indictranstech/tele-frappe,gangadharkadam/johnfrappe,ShashaQin/frappe,indictranstech/reciphergroup-frappe,gangadharkadam/smrtfrappe,adityahase/frappe,vqw/frappe,reachalpineswift/frappe-bench,gangadharkadam/tailorfrappe,rohitwaghchaure/New_Theme_frappe,mbauskar/omnitech-demo-frappe,BhupeshGupta/frappe,gangadharkadam/v5_frappe,gangadharkadam/saloon_frappe_install,hernad/frappe,pawaranand/phr_frappe,indictranstech/tele-frappe,indictranstech/frappe-digitales,gangadharkadam/johnfrappe,gangadharkadam/frappecontribution,anandpdoshi/frappe,rohitwaghchaure/frappe-digitales,gangadhar-kadam/laganfrappe,shitolepriya/test-frappe,deveninfotech/deven-frappe,erpletzerp/letzerpcore,rmehta/frappe,pranalik/frappe-bb,chdecultot/frappe,indictranstech/reciphergroup-frappe,saurabh6790/phr-frappe,mbauskar/helpdesk-frappe,mbauskar/omnitech-frappe,indictranstech/ebuy-now-frappe,gangadharkadam/vervefrappe,indictranstech/Das_frappe,gangadhar-kadam/verve_frappe,vjFaLk/frappe,RicardoJohann/frappe,anandpdoshi/frappe,rohitwaghchaure/frappe,gangadharkadam/shfr,indictranstech/internal-frappe,indictranstech/internal-frappe,praba230890/frappe,gangadharkadam/v5_frappe,rohitwaghchaure/vestasi-frappe,mbauskar/frappe,suyashphadtare/propshikhari-frappe,reachalpineswift/frappe-bench,yashodhank/frappe,indictranstech/reciphergroup-frappe,rohitwaghchaure/vestasi-frappe,indautgrp/frappe,drukhil/frappe,gangadhar-kadam/verve_test_frappe,adityahase/frappe,saurabh6790/phr-frappe,praba230890/frappe,sbktechnology/sap_frappe,elba7r/builder,rmehta/frappe,manassolanki/frappe,gangadharkadam/vlinkfrappe,ShashaQin/frappe,mbauskar/omnitech-demo-frappe,pombredanne/frappe,pawaranand/phr_frappe,tmimori/frappe,deveninfotech/deven-frappe,rohitw1991/smartfrappe,gangadharkadam/v4_frappe,gangadharkadam/vervefrappe,rohitwaghchaure/New_Theme_frappe,MaxMorais/frappe,gangadharkadam/vlinkfrappe,almeidapaulopt/frappe,mhbu50/frappe,saurabh6790/frappe,indictranstech/frappe,neilLasrado/frappe,rohitwaghchaure/frappe,pranalik/frappe-bb,gangadharkadam/v6_frappe,drukhil/frappe,suyashphadtare/sajil-frappe,yashodhank/frappe,vCentre/vFRP-6233,saurabh6790/frappe,elba7r/frameworking,indictranstech/ebuy-now-frappe,indictranstech/fbd_frappe,ShashaQin/frappe,sbkolate/sap_frappe_v6,rohitwaghchaure/frappe,reachalpineswift/frappe-bench,suyashphadtare/sajil-final-frappe,mbauskar/omnitech-frappe,ashokrajbathu/secondrep,saguas/frappe,bcornwellmott/frappe,Tejal011089/digitales_frappe,letzerp/framework,gangadharkadam/v5_frappe,ESS-LLP/frappe,deveninfotech/deven-frappe,suyashphadtare/sajil-final-frappe,saurabh6790/test-frappe,anandpdoshi/frappe,frappe/frappe,mbauskar/Das_frappe,gangadhar-kadam/verve_frappe,jevonearth/frappe,gangadhar-kadam/verve_live_frappe,reachalpineswift/frappe-bench,mbauskar/tele-frappe,sbktechnology/sap_frappe,rohitw1991/smartfrappe,elba7r/frameworking,mbauskar/phr-frappe,Tejal011089/digitales_frappe,rohitwaghchaure/frappe-digitales,gangadharkadam/v5_frappe,elba7r/frameworking,gangadharkadam/frappecontribution,indictranstech/fbd_frappe,indictranstech/phr-frappe,tundebabzy/frappe,suyashphadtare/sajil-frappe,gangadharkadam/vlinkfrappe,Amber-Creative/amber-frappe,tundebabzy/frappe,indictranstech/internal-frappe,gangadharkadam/office_frappe,mbauskar/frappe,Amber-Creative/amber-frappe,indictranstech/Das_frappe,rohitw1991/smarttailorfrappe,vqw/frappe,sbktechnology/trufil-frappe,sbktechnology/trufil-frappe,yashodhank/frappe,ashokrajbathu/secondrep,pawaranand/phr_frappe,gangadharkadam/saloon_frappe_install,gangadharkadam/saloon_frappe
|
---
+++
@@ -4,10 +4,11 @@
from __future__ import unicode_literals
import frappe
from frappe.utils.scheduler import disable_scheduler, enable_scheduler
+from frappe.utils import cint
def execute():
frappe.reload_doc("core", "doctype", "system_settings")
- if frappe.db.get_global("disable_scheduler"):
+ if cint(frappe.db.get_global("disable_scheduler")):
disable_scheduler()
else:
enable_scheduler()
|
fd79823893b9b83a184c2bcd0fbe32fbb51619c9
|
src/server/convert.py
|
src/server/convert.py
|
# midi-beeper-orchestra - program to create an orchestra from PC speakers
# Copyright (C) 2015 The Underscores
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as published
# by the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
import math
def MIDItoHz(MIDIval):
"""
Converts a MIDI note, MIDIval, value to the equivalent hertz value
"""
#return 69 + 12 * math.log((MIDIval/440), 2)
return (2**((MIDIval-69)/12))*440
def hzToMIDI(hz):
"""
Converts hertz, hz, to MIDI note equivalent
"""
midi = 2**((hz-69)/12) * 440
return int(midi + 0.5)
|
# midi-beeper-orchestra - program to create an orchestra from PC speakers
# Copyright (C) 2015 The Underscores
# This program is free software: you can redistribute it and/or modify
# it under the terms of the GNU Affero General Public License as published
# by the Free Software Foundation, either version 3 of the License, or
# (at your option) any later version.
# This program is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
# GNU Affero General Public License for more details.
# You should have received a copy of the GNU Affero General Public License
# along with this program. If not, see <http://www.gnu.org/licenses/>.
import math
def MIDItoHz(MIDIval):
"""
Converts a MIDI note, MIDIval, value to the equivalent hertz value
"""
return (2**((MIDIval-69)/12))*440
def hzToMIDI(hz):
"""
Converts hertz, hz, to MIDI note equivalent
"""
midi = 2**((hz-69)/12) * 440
return int(midi + 0.5)
|
Remove comment containing incorrect conversion function.
|
Remove comment containing incorrect conversion function.
|
Python
|
agpl-3.0
|
TheUnderscores/midi-beeper-orchestra
|
---
+++
@@ -17,7 +17,6 @@
"""
Converts a MIDI note, MIDIval, value to the equivalent hertz value
"""
- #return 69 + 12 * math.log((MIDIval/440), 2)
return (2**((MIDIval-69)/12))*440
def hzToMIDI(hz):
|
d81a1ba12add244cb246efeae5c292a6d995c9b8
|
deadlinks.py
|
deadlinks.py
|
from operator import itemgetter
from itertools import chain
import os
import yaml
import requests
yaml.load_all
directory = "_companies"
flat = chain.from_iterable
def link_status_company(filename):
(name, _) = filename.rsplit(".", 1);
print("==== {name} ====".format(name=name))
docs = filter(None, yaml.load_all(open(os.path.join(directory, filename))))
positions = flat(map(itemgetter("positions"), filter(lambda doc: "positions" in doc, docs)))
def link_status_position(position):
title = position["title"]
url = position["url"]
print("{title} [ {url} ]".format(title=title, url=url))
response = requests.get(url)
status_code_description = requests.status_codes._codes.get(response.status_code, '-')
print("{} {} {}".format(response.status_code, status_code_description, response.history))
print()
list(map(link_status_position, positions))
list(map(link_status_company, sorted(os.listdir(directory))))
|
from operator import itemgetter
from itertools import chain
import os
import yaml
import requests
yaml.load_all
directory = "_companies"
flat = chain.from_iterable
def link_status_company(filename):
(name, _) = filename.rsplit(".", 1);
print("==== {name} ====".format(name=name))
docs = filter(None, yaml.load_all(open(os.path.join(directory, filename))))
positions = flat(map(itemgetter("positions"), filter(lambda doc: "positions" in doc, docs)))
def link_status_position(position):
title = position["title"]
url = position["url"]
print("{title} [ {url} ]".format(title=title, url=url))
try:
response = requests.get(url, timeout=10)
status_code_description = requests.status_codes._codes.get(response.status_code, '-')
print("{} {} {}".format(response.status_code, status_code_description, response.history))
except Exception as e:
print(e)
print()
list(map(link_status_position, positions))
list(map(link_status_company, sorted(os.listdir(directory))))
|
Add timeout to dead links script
|
Add timeout to dead links script
|
Python
|
apache-2.0
|
Stockholm-AI/stockholm-ai,Stockholm-AI/stockholm-ai,Stockholm-AI/stockholm-ai,Stockholm-AI/stockholm-ai,Stockholm-AI/stockholm-ai
|
---
+++
@@ -23,11 +23,14 @@
url = position["url"]
print("{title} [ {url} ]".format(title=title, url=url))
- response = requests.get(url)
+ try:
+ response = requests.get(url, timeout=10)
- status_code_description = requests.status_codes._codes.get(response.status_code, '-')
+ status_code_description = requests.status_codes._codes.get(response.status_code, '-')
- print("{} {} {}".format(response.status_code, status_code_description, response.history))
+ print("{} {} {}".format(response.status_code, status_code_description, response.history))
+ except Exception as e:
+ print(e)
print()
|
64cbf90dbb8378752398d15d75ee5af9aa5b32ed
|
linter.py
|
linter.py
|
#
# linter.py
# Linter for SublimeLinter3, a code checking framework for Sublime Text 3
#
# Written by Aparajita Fishman
# Copyright (c) 2013 Aparajita Fishman
#
# Project: https://github.com/SublimeLinter/SublimeLinter-contrib-pep257
# License: MIT
#
"""This module exports the PEP257 plugin linter class."""
import os
from SublimeLinter.lint import highlight, PythonLinter
class PEP257(PythonLinter):
"""Provides an interface to the pep257 python module/script."""
language = 'python'
cmd = ('pep257@python', '-')
regex = r'^.+?:(?P<line>\d+):(?P<col>\d+): (?P<message>.+)'
default_type = highlight.WARNING
line_col_base = (1, 0) # pep257 uses one-based line and zero-based column numbers
module = 'pep257'
def check(self, code, filename):
"""Run pep257 on code and return the output."""
return self.module.check_source(code, os.path.basename(filename))
|
#
# linter.py
# Linter for SublimeLinter3, a code checking framework for Sublime Text 3
#
# Written by Aparajita Fishman
# Copyright (c) 2013 Aparajita Fishman
#
# Project: https://github.com/SublimeLinter/SublimeLinter-contrib-pep257
# License: MIT
#
"""This module exports the PEP257 plugin linter class."""
import os
from SublimeLinter.lint import highlight, PythonLinter
class PEP257(PythonLinter):
"""Provides an interface to the pep257 python module/script."""
syntax = 'python'
cmd = ('pep257@python', '-')
regex = r'^.+?:(?P<line>\d+):(?P<col>\d+): (?P<message>.+)'
default_type = highlight.WARNING
line_col_base = (1, 0) # pep257 uses one-based line and zero-based column numbers
module = 'pep257'
def check(self, code, filename):
"""Run pep257 on code and return the output."""
return self.module.check_source(code, os.path.basename(filename))
|
Change 'language' to 'syntax', that is more precise terminology.
|
Change 'language' to 'syntax', that is more precise terminology.
|
Python
|
mit
|
SublimeLinter/SublimeLinter-pep257
|
---
+++
@@ -20,7 +20,7 @@
"""Provides an interface to the pep257 python module/script."""
- language = 'python'
+ syntax = 'python'
cmd = ('pep257@python', '-')
regex = r'^.+?:(?P<line>\d+):(?P<col>\d+): (?P<message>.+)'
default_type = highlight.WARNING
|
96b9c25268e98e9464d8b068aa12de113ad1c66f
|
joby/spiders/data_science_jobs.py
|
joby/spiders/data_science_jobs.py
|
# -*- coding: utf-8 -*-
import scrapy
class DataScienceJobsSpider(scrapy.Spider):
name = "data-science-jobs"
allowed_domains = ["www.data-science-jobs.com"]
start_urls = (
'http://www.data-science-jobs.com/',
)
def parse(self, response):
pass
|
# -*- coding: utf-8 -*-
from logging import getLogger
from scrapy.spiders import Rule, CrawlSpider
from scrapy.linkextractors import LinkExtractor
class DataScienceJobsSpider(CrawlSpider):
log = getLogger(__name__)
name = 'data-science-jobs'
allowed_domains = ['www.data-science-jobs.com', 'fonts.googleapis.com', 'jobs.lever.com']
start_urls = ['http://www.data-science-jobs.com/']
test = Rule(LinkExtractor(allow='family'), callback='parse')
test2 = Rule(LinkExtractor(allow='comtravo'), callback='parse')
job_links = Rule(LinkExtractor(allow='detail\/'), callback='parse')
pagination_links = Rule(LinkExtractor(allow='\?page=\d+'), callback='parse')
rules = [job_links, pagination_links, test, test2]
response = None
def parse(self, response):
self.log.info('Parsing %s', response.url)
|
Add more rules for test purposes.
|
Add more rules for test purposes.
|
Python
|
mit
|
cyberbikepunk/job-spiders
|
---
+++
@@ -1,13 +1,21 @@
# -*- coding: utf-8 -*-
-import scrapy
+
+from logging import getLogger
+from scrapy.spiders import Rule, CrawlSpider
+from scrapy.linkextractors import LinkExtractor
-class DataScienceJobsSpider(scrapy.Spider):
- name = "data-science-jobs"
- allowed_domains = ["www.data-science-jobs.com"]
- start_urls = (
- 'http://www.data-science-jobs.com/',
- )
+class DataScienceJobsSpider(CrawlSpider):
+ log = getLogger(__name__)
+ name = 'data-science-jobs'
+ allowed_domains = ['www.data-science-jobs.com', 'fonts.googleapis.com', 'jobs.lever.com']
+ start_urls = ['http://www.data-science-jobs.com/']
+ test = Rule(LinkExtractor(allow='family'), callback='parse')
+ test2 = Rule(LinkExtractor(allow='comtravo'), callback='parse')
+ job_links = Rule(LinkExtractor(allow='detail\/'), callback='parse')
+ pagination_links = Rule(LinkExtractor(allow='\?page=\d+'), callback='parse')
+ rules = [job_links, pagination_links, test, test2]
+ response = None
def parse(self, response):
- pass
+ self.log.info('Parsing %s', response.url)
|
8bfb43e6af19e6335a95768289036fb73caac559
|
pinax/testimonials/templatetags/pinax_testimonials_tags.py
|
pinax/testimonials/templatetags/pinax_testimonials_tags.py
|
from django import template
from ..models import Testimonial
register = template.Library()
@register.simple_tag
def random_testimonials(number):
return Testimonial.objects.filter(active=True).order_by("?")[:number]
@register.simple_tag
def random_testimonial():
queryset = Testimonial.objects.filter(active=True).order_by("?")
return queryset[0] if queryset else None
@register.simple_tag
def testimonials(number):
return Testimonial.objects.filter(active=True).order_by("-added")[:number]
|
from django import template
from ..models import Testimonial
register = template.Library()
@register.simple_tag
def random_testimonials(number):
return Testimonial.objects.filter(active=True).order_by("?")[:number]
@register.simple_tag
def random_testimonial():
queryset = Testimonial.objects.filter(active=True).order_by("?")
return queryset[0] if queryset else None
@register.simple_tag
def testimonials(number=None):
return Testimonial.objects.filter(active=True).order_by("-added")[:number]
|
Make number parameter optional to testimonials tag
|
Make number parameter optional to testimonials tag
|
Python
|
mit
|
pinax/pinax-testimonials
|
---
+++
@@ -17,5 +17,5 @@
@register.simple_tag
-def testimonials(number):
+def testimonials(number=None):
return Testimonial.objects.filter(active=True).order_by("-added")[:number]
|
0692cc324d3759703ee52e117ac19e75d82df6a6
|
tests/config/tests.py
|
tests/config/tests.py
|
from raven.conf import load
from unittest2 import TestCase
class LoadTest(TestCase):
def test_basic(self):
dsn = 'https://foo:bar@sentry.local/1'
res = {}
load(dsn, res)
self.assertEquals(res, {
'SENTRY_PROJECT': '1',
'SENTRY_SERVERS': ['https://sentry.local/api/store/'],
'SENTRY_PUBLIC_KEY': 'foo',
'SENTRY_SECRET_KEY': 'bar',
})
def test_path(self):
dsn = 'https://foo:bar@sentry.local/app/1'
res = {}
load(dsn, res)
self.assertEquals(res, {
'SENTRY_PROJECT': '1',
'SENTRY_SERVERS': ['https://sentry.local/app/api/store/'],
'SENTRY_PUBLIC_KEY': 'foo',
'SENTRY_SECRET_KEY': 'bar',
})
|
import logging
import mock
from raven.conf import load, setup_logging
from unittest2 import TestCase
class LoadTest(TestCase):
def test_basic(self):
dsn = 'https://foo:bar@sentry.local/1'
res = {}
load(dsn, res)
self.assertEquals(res, {
'SENTRY_PROJECT': '1',
'SENTRY_SERVERS': ['https://sentry.local/api/store/'],
'SENTRY_PUBLIC_KEY': 'foo',
'SENTRY_SECRET_KEY': 'bar',
})
def test_path(self):
dsn = 'https://foo:bar@sentry.local/app/1'
res = {}
load(dsn, res)
self.assertEquals(res, {
'SENTRY_PROJECT': '1',
'SENTRY_SERVERS': ['https://sentry.local/app/api/store/'],
'SENTRY_PUBLIC_KEY': 'foo',
'SENTRY_SECRET_KEY': 'bar',
})
class SetupLoggingTest(TestCase):
def test_basic_not_configured(self):
with mock.patch('logging.getLogger', spec=logging.getLogger) as getLogger:
logger = getLogger()
logger.handlers = []
handler = mock.Mock()
result = setup_logging(handler)
self.assertTrue(result)
def test_basic_already_configured(self):
with mock.patch('logging.getLogger', spec=logging.getLogger) as getLogger:
handler = mock.Mock()
logger = getLogger()
logger.handlers = [handler]
result = setup_logging(handler)
self.assertFalse(result)
|
Add basic coverage for the setup_logging method
|
Add basic coverage for the setup_logging method
|
Python
|
bsd-3-clause
|
inspirehep/raven-python,jmagnusson/raven-python,Goldmund-Wyldebeast-Wunderliebe/raven-python,akalipetis/raven-python,inspirehep/raven-python,nikolas/raven-python,hzy/raven-python,smarkets/raven-python,beniwohli/apm-agent-python,icereval/raven-python,jmagnusson/raven-python,percipient/raven-python,dirtycoder/opbeat_python,someonehan/raven-python,johansteffner/raven-python,Photonomie/raven-python,recht/raven-python,lopter/raven-python-old,icereval/raven-python,arthurlogilab/raven-python,patrys/opbeat_python,getsentry/raven-python,alex/raven,jbarbuto/raven-python,lepture/raven-python,inspirehep/raven-python,jmagnusson/raven-python,Goldmund-Wyldebeast-Wunderliebe/raven-python,tarkatronic/opbeat_python,jmp0xf/raven-python,recht/raven-python,Goldmund-Wyldebeast-Wunderliebe/raven-python,smarkets/raven-python,jbarbuto/raven-python,lepture/raven-python,jbarbuto/raven-python,recht/raven-python,ronaldevers/raven-python,icereval/raven-python,daikeren/opbeat_python,ronaldevers/raven-python,nikolas/raven-python,beniwohli/apm-agent-python,someonehan/raven-python,getsentry/raven-python,beniwohli/apm-agent-python,percipient/raven-python,akalipetis/raven-python,smarkets/raven-python,patrys/opbeat_python,collective/mr.poe,inspirehep/raven-python,ticosax/opbeat_python,jmp0xf/raven-python,akalipetis/raven-python,patrys/opbeat_python,daikeren/opbeat_python,hzy/raven-python,danriti/raven-python,danriti/raven-python,jbarbuto/raven-python,ewdurbin/raven-python,ewdurbin/raven-python,daikeren/opbeat_python,tarkatronic/opbeat_python,nikolas/raven-python,Photonomie/raven-python,percipient/raven-python,openlabs/raven,lepture/raven-python,ticosax/opbeat_python,nikolas/raven-python,jmp0xf/raven-python,danriti/raven-python,akheron/raven-python,ewdurbin/raven-python,beniwohli/apm-agent-python,hzy/raven-python,smarkets/raven-python,someonehan/raven-python,icereval/raven-python,dirtycoder/opbeat_python,getsentry/raven-python,Goldmund-Wyldebeast-Wunderliebe/raven-python,akheron/raven-python,dbravender/raven-python,tarkatronic/opbeat_python,Photonomie/raven-python,arthurlogilab/raven-python,ronaldevers/raven-python,dbravender/raven-python,johansteffner/raven-python,johansteffner/raven-python,ticosax/opbeat_python,arthurlogilab/raven-python,dbravender/raven-python,akheron/raven-python,patrys/opbeat_python,arthurlogilab/raven-python,dirtycoder/opbeat_python
|
---
+++
@@ -1,4 +1,6 @@
-from raven.conf import load
+import logging
+import mock
+from raven.conf import load, setup_logging
from unittest2 import TestCase
@@ -24,3 +26,21 @@
'SENTRY_PUBLIC_KEY': 'foo',
'SENTRY_SECRET_KEY': 'bar',
})
+
+
+class SetupLoggingTest(TestCase):
+ def test_basic_not_configured(self):
+ with mock.patch('logging.getLogger', spec=logging.getLogger) as getLogger:
+ logger = getLogger()
+ logger.handlers = []
+ handler = mock.Mock()
+ result = setup_logging(handler)
+ self.assertTrue(result)
+
+ def test_basic_already_configured(self):
+ with mock.patch('logging.getLogger', spec=logging.getLogger) as getLogger:
+ handler = mock.Mock()
+ logger = getLogger()
+ logger.handlers = [handler]
+ result = setup_logging(handler)
+ self.assertFalse(result)
|
1ec2779f5e4470c6ed19b56d16185c6174ab520c
|
tests/test_readers.py
|
tests/test_readers.py
|
# coding: utf-8
try:
import unittest2
except ImportError, e:
import unittest as unittest2
import datetime
import os
from pelican import readers
CUR_DIR = os.path.dirname(__file__)
CONTENT_PATH = os.path.join(CUR_DIR, 'content')
def _filename(*args):
return os.path.join(CONTENT_PATH, *args)
class RstReaderTest(unittest2.TestCase):
def test_article_with_metadata(self):
reader = readers.RstReader({})
content, metadata = reader.read(_filename('article_with_metadata.rst'))
expected = {
'category': 'yeah',
'author': u'Alexis Métaireau',
'title': 'This is a super article !',
'summary': 'Multi-line metadata should be supported\nas well as <strong>inline markup</strong>.',
'date': datetime.datetime(2010, 12, 2, 10, 14),
'tags': ['foo', 'bar', 'foobar'],
}
self.assertDictEqual(metadata, expected)
|
# coding: utf-8
try:
import unittest2 as unittest
except ImportError, e:
import unittest
import datetime
import os
from pelican import readers
CUR_DIR = os.path.dirname(__file__)
CONTENT_PATH = os.path.join(CUR_DIR, 'content')
def _filename(*args):
return os.path.join(CONTENT_PATH, *args)
class RstReaderTest(unittest.TestCase):
def test_article_with_metadata(self):
reader = readers.RstReader({})
content, metadata = reader.read(_filename('article_with_metadata.rst'))
expected = {
'category': 'yeah',
'author': u'Alexis Métaireau',
'title': 'This is a super article !',
'summary': 'Multi-line metadata should be supported\nas well as <strong>inline markup</strong>.',
'date': datetime.datetime(2010, 12, 2, 10, 14),
'tags': ['foo', 'bar', 'foobar'],
}
for key, value in expected.items():
self.assertEquals(value, metadata[key], key)
|
Make the readers tests a bit more verbose.
|
Make the readers tests a bit more verbose.
|
Python
|
agpl-3.0
|
11craft/pelican,51itclub/pelican,simonjj/pelican,sunzhongwei/pelican,goerz/pelican,51itclub/pelican,Rogdham/pelican,lucasplus/pelican,gymglish/pelican,sunzhongwei/pelican,sunzhongwei/pelican,koobs/pelican,UdeskDeveloper/pelican,ingwinlu/pelican,kennethlyn/pelican,karlcow/pelican,zackw/pelican,alexras/pelican,HyperGroups/pelican,rbarraud/pelican,crmackay/pelican,koobs/pelican,iurisilvio/pelican,jimperio/pelican,zackw/pelican,btnpushnmunky/pelican,ehashman/pelican,douglaskastle/pelican,HyperGroups/pelican,deanishe/pelican,iKevinY/pelican,ls2uper/pelican,karlcow/pelican,Rogdham/pelican,Rogdham/pelican,catdog2/pelican,kernc/pelican,number5/pelican,btnpushnmunky/pelican,crmackay/pelican,iurisilvio/pelican,Natim/pelican,GiovanniMoretti/pelican,avaris/pelican,HyperGroups/pelican,51itclub/pelican,deved69/pelican-1,liyonghelpme/myBlog,abrahamvarricatt/pelican,catdog2/pelican,kernc/pelican,jvehent/pelican,11craft/pelican,douglaskastle/pelican,jvehent/pelican,kennethlyn/pelican,garbas/pelican,gymglish/pelican,levanhien8/pelican,iurisilvio/pelican,lazycoder-ru/pelican,UdeskDeveloper/pelican,douglaskastle/pelican,koobs/pelican,treyhunner/pelican,florianjacob/pelican,Summonee/pelican,JeremyMorgan/pelican,joetboole/pelican,alexras/pelican,kennethlyn/pelican,ehashman/pelican,jo-tham/pelican,Polyconseil/pelican,JeremyMorgan/pelican,deved69/pelican-1,treyhunner/pelican,Scheirle/pelican,jimperio/pelican,deved69/pelican-1,avaris/pelican,Scheirle/pelican,btnpushnmunky/pelican,jo-tham/pelican,joetboole/pelican,fbs/pelican,ls2uper/pelican,eevee/pelican,ingwinlu/pelican,jvehent/pelican,TC01/pelican,talha131/pelican,iKevinY/pelican,kernc/pelican,florianjacob/pelican,getpelican/pelican,karlcow/pelican,goerz/pelican,deanishe/pelican,lazycoder-ru/pelican,levanhien8/pelican,JeremyMorgan/pelican,number5/pelican,liyonghelpme/myBlog,Summonee/pelican,levanhien8/pelican,ionelmc/pelican,janaurka/git-debug-presentiation,joetboole/pelican,ls2uper/pelican,farseerfc/pelican,simonjj/pelican,11craft/pelican,TC01/pelican,alexras/pelican,crmackay/pelican,goerz/pelican,liyonghelpme/myBlog,GiovanniMoretti/pelican,janaurka/git-debug-presentiation,lucasplus/pelican,number5/pelican,simonjj/pelican,abrahamvarricatt/pelican,treyhunner/pelican,eevee/pelican,lazycoder-ru/pelican,rbarraud/pelican,garbas/pelican,janaurka/git-debug-presentiation,arty-name/pelican,liyonghelpme/myBlog,rbarraud/pelican,TC01/pelican,catdog2/pelican,ehashman/pelican,Summonee/pelican,talha131/pelican,Scheirle/pelican,farseerfc/pelican,0xMF/pelican,justinmayer/pelican,florianjacob/pelican,deanishe/pelican,garbas/pelican,gymglish/pelican,abrahamvarricatt/pelican,jimperio/pelican,zackw/pelican,sunzhongwei/pelican,liyonghelpme/myBlog,UdeskDeveloper/pelican,getpelican/pelican,Polyconseil/pelican,lucasplus/pelican,eevee/pelican,GiovanniMoretti/pelican
|
---
+++
@@ -1,8 +1,8 @@
# coding: utf-8
try:
- import unittest2
+ import unittest2 as unittest
except ImportError, e:
- import unittest as unittest2
+ import unittest
import datetime
import os
@@ -12,11 +12,12 @@
CUR_DIR = os.path.dirname(__file__)
CONTENT_PATH = os.path.join(CUR_DIR, 'content')
+
def _filename(*args):
return os.path.join(CONTENT_PATH, *args)
-class RstReaderTest(unittest2.TestCase):
+class RstReaderTest(unittest.TestCase):
def test_article_with_metadata(self):
reader = readers.RstReader({})
@@ -29,4 +30,6 @@
'date': datetime.datetime(2010, 12, 2, 10, 14),
'tags': ['foo', 'bar', 'foobar'],
}
- self.assertDictEqual(metadata, expected)
+
+ for key, value in expected.items():
+ self.assertEquals(value, metadata[key], key)
|
f3ea9820a96536e74e6f74f13387140c97ea9f2e
|
backgroundworker.py
|
backgroundworker.py
|
import sys
import os
sys.path.insert(0, "../financialScraper")
import pandas as pd
from financialScraper import getqf
from sqlalchemy import create_engine
running = True
# engine = create_engine(os.environ.get('DATABASE_URL'))
engine = create_engine('postgres://fkwfcpvbchmxps:VCmxue5WFWCOOHt56aqOm4FD_Z@ec2-54-83-205-46.compute-1.amazonaws.com:5432/d376d3nru8envq')
dfdict = getqf.scraper()
df = dfdict['nsdqct.csv']
df.to_sql(name='entries', con = engine, if_exists = 'replace')
|
import sys
import os
sys.path.insert(0, "../financialScraper")
import pandas as pd
from financialScraper import getqf
from sqlalchemy import create_engine
running = True
# engine = create_engine(os.environ.get('DATABASE_URL'))
engine = create_engine('postgres://fkwfcpvbchmxps:VCmxue5WFWCOOHt56aqOm4FD_Z@ec2-54-83-205-46.compute-1.amazonaws.com:5432/d376d3nru8envq')
connection = engine.connect()
dfdict = getqf.scraper()
df = dfdict['nsdqct.csv']
df.to_sql(name='entries', con = connection, if_exists = 'replace')
connection.close()
|
Add engine connection, and close engine connection to worker dyno
|
Add engine connection, and close engine connection to worker dyno
|
Python
|
mit
|
caseymacphee/green_quote,caseymacphee/green_quote
|
---
+++
@@ -9,7 +9,11 @@
# engine = create_engine(os.environ.get('DATABASE_URL'))
+
engine = create_engine('postgres://fkwfcpvbchmxps:VCmxue5WFWCOOHt56aqOm4FD_Z@ec2-54-83-205-46.compute-1.amazonaws.com:5432/d376d3nru8envq')
+connection = engine.connect()
dfdict = getqf.scraper()
df = dfdict['nsdqct.csv']
-df.to_sql(name='entries', con = engine, if_exists = 'replace')
+df.to_sql(name='entries', con = connection, if_exists = 'replace')
+
+connection.close()
|
8868cb556851d3caf227281873d619ec3ddc726a
|
matador/commands/deploy_ticket.py
|
matador/commands/deploy_ticket.py
|
#!/usr/bin/env python
from .command import Command
from matador import utils
class DeployTicket(Command):
def _add_arguments(self, parser):
parser.add_argument(
'-e', '--environment',
type=str,
required=True,
help='Agresso environment name')
def _execute(self):
project_folder = utils.project_folder()
self._logger.info(project_folder)
working_folder = utils.working_folder('uog01', self.args.environment)
self._logger.info(working_folder)
project = utils.project()
self._logger.info(project)
self._logger.info(utils.is_git_repository())
|
#!/usr/bin/env python
from .command import Command
from matador import utils
class DeployTicket(Command):
def _add_arguments(self, parser):
parser.prog = 'matador deploy-ticket'
parser.add_argument(
'-e', '--environment',
type=str,
required=True,
help='Agresso environment name')
def _execute(self):
project_folder = utils.project_folder()
self._logger.info(project_folder)
working_folder = utils.working_folder('uog01', self.args.environment)
self._logger.info(working_folder)
project = utils.project()
self._logger.info(project)
self._logger.info(utils.is_git_repository())
|
Add program name to parser
|
Add program name to parser
|
Python
|
mit
|
Empiria/matador
|
---
+++
@@ -6,6 +6,7 @@
class DeployTicket(Command):
def _add_arguments(self, parser):
+ parser.prog = 'matador deploy-ticket'
parser.add_argument(
'-e', '--environment',
type=str,
|
8f82336aed62a18b2c6f824fcf0e6b1a1d00b8d3
|
tests/test_astroid.py
|
tests/test_astroid.py
|
# -*- coding: utf-8 -*-
from __future__ import unicode_literals, print_function
import re
import astroid
from . import tools, test_mark_tokens
class TestAstroid(test_mark_tokens.TestMarkTokens):
is_astroid_test = True
module = astroid
nodes_classes = astroid.ALL_NODE_CLASSES
context_classes = [
(astroid.Name, astroid.DelName, astroid.AssignName),
(astroid.Attribute, astroid.DelAttr, astroid.AssignAttr),
]
@staticmethod
def iter_fields(node):
"""
Yield a tuple of ``(fieldname, value)`` for each field
that is present on *node*.
Similar to ast.iter_fields, but for astroid and ignores context
"""
for field in node._astroid_fields + node._other_fields:
if field == 'ctx':
continue
yield field, getattr(node, field)
@classmethod
def create_mark_checker(cls, source):
builder = astroid.builder.AstroidBuilder()
tree = builder.string_build(source)
return tools.MarkChecker(source, tree=tree)
|
# -*- coding: utf-8 -*-
from __future__ import unicode_literals, print_function
import astroid
from astroid.node_classes import NodeNG
from . import tools, test_mark_tokens
class TestAstroid(test_mark_tokens.TestMarkTokens):
is_astroid_test = True
module = astroid
nodes_classes = NodeNG
context_classes = [
(astroid.Name, astroid.DelName, astroid.AssignName),
(astroid.Attribute, astroid.DelAttr, astroid.AssignAttr),
]
@staticmethod
def iter_fields(node):
"""
Yield a tuple of ``(fieldname, value)`` for each field
that is present on *node*.
Similar to ast.iter_fields, but for astroid and ignores context
"""
for field in node._astroid_fields + node._other_fields:
if field == 'ctx':
continue
yield field, getattr(node, field)
@classmethod
def create_mark_checker(cls, source):
builder = astroid.builder.AstroidBuilder()
tree = builder.string_build(source)
return tools.MarkChecker(source, tree=tree)
|
Use NodeNG instead of astroid.ALL_NODE_CLASSES
|
Use NodeNG instead of astroid.ALL_NODE_CLASSES
|
Python
|
apache-2.0
|
gristlabs/asttokens
|
---
+++
@@ -1,9 +1,8 @@
# -*- coding: utf-8 -*-
from __future__ import unicode_literals, print_function
-import re
-
import astroid
+from astroid.node_classes import NodeNG
from . import tools, test_mark_tokens
@@ -13,7 +12,7 @@
is_astroid_test = True
module = astroid
- nodes_classes = astroid.ALL_NODE_CLASSES
+ nodes_classes = NodeNG
context_classes = [
(astroid.Name, astroid.DelName, astroid.AssignName),
(astroid.Attribute, astroid.DelAttr, astroid.AssignAttr),
|
59a345f29d77026c9bd8b2ec27df71253dc3d1e6
|
tests/test_helpers.py
|
tests/test_helpers.py
|
from scrapi.base import helpers
class TestHelpers(object):
def test_format_one_tag(self):
single_tag = ' A single tag '
single_output = helpers.format_tags(single_tag)
assert single_output == ['a single tag']
assert isinstance(single_output, list)
def test_format_many_tags(self):
many_tags = [' A', 'Bunch', ' oftags ']
many_output = helpers.format_tags(many_tags)
assert set(many_output) == set(['a', 'bunch', 'oftags'])
def test_format_sep_tags(self):
sep_tags = ['These, we know', 'should be many']
sep_output = helpers.format_tags(sep_tags, sep=',')
assert set(sep_output) == set(['these', 'we know', 'should be many'])
def test_extract_doi(self):
identifiers = 'doi: THIS_IS_A_DOI!'
valid_doi = helpers.oai_extract_doi(identifiers)
assert valid_doi == 'THIS_IS_A_DOI!'
def test_oai_extract_url(self):
identifiers = 'I might be a url but rly I am naaaahhttt'
extraction_attempt = helpers.oai_extract_url(identifiers)
extraction_attempt
def test_process_contributors(self):
args = ['Stardust Rhodes', 'Golddust Rhodes', 'Dusty Rhodes']
response = helpers.oai_process_contributors(args)
assert isinstance(response, list)
|
from scrapi.base import helpers
class TestHelpers(object):
def test_format_one_tag(self):
single_tag = ' A single tag '
single_output = helpers.format_tags(single_tag)
assert single_output == ['a single tag']
assert isinstance(single_output, list)
def test_format_many_tags(self):
many_tags = [' A', 'Bunch', ' oftags ']
many_output = helpers.format_tags(many_tags)
assert set(many_output) == set(['a', 'bunch', 'oftags'])
def test_format_sep_tags(self):
sep_tags = ['These, we know', 'should be many']
sep_output = helpers.format_tags(sep_tags, sep=',')
assert set(sep_output) == set(['these', 'we know', 'should be many'])
def test_extract_dois(self):
identifiers = 'doi: THIS_IS_A_DOI!'
valid_doi = helpers.oai_extract_dois(identifiers)
assert valid_doi == 'THIS_IS_A_DOI!'
def test_oai_extract_url(self):
identifiers = 'I might be a url but rly I am naaaahhttt'
extraction_attempt = helpers.oai_extract_url(identifiers)
extraction_attempt
def test_process_contributors(self):
args = ['Stardust Rhodes', 'Golddust Rhodes', 'Dusty Rhodes']
response = helpers.oai_process_contributors(args)
assert isinstance(response, list)
|
Update test to use new name of helper for dois
|
Update test to use new name of helper for dois
|
Python
|
apache-2.0
|
fabianvf/scrapi,CenterForOpenScience/scrapi,CenterForOpenScience/scrapi,mehanig/scrapi,icereval/scrapi,jeffreyliu3230/scrapi,ostwald/scrapi,felliott/scrapi,erinspace/scrapi,fabianvf/scrapi,felliott/scrapi,mehanig/scrapi,alexgarciac/scrapi,erinspace/scrapi
|
---
+++
@@ -19,9 +19,9 @@
sep_output = helpers.format_tags(sep_tags, sep=',')
assert set(sep_output) == set(['these', 'we know', 'should be many'])
- def test_extract_doi(self):
+ def test_extract_dois(self):
identifiers = 'doi: THIS_IS_A_DOI!'
- valid_doi = helpers.oai_extract_doi(identifiers)
+ valid_doi = helpers.oai_extract_dois(identifiers)
assert valid_doi == 'THIS_IS_A_DOI!'
def test_oai_extract_url(self):
|
71f007559ca2adf69d09f3c6f720383bfd784897
|
students/forms.py
|
students/forms.py
|
from django.contrib.auth import get_user_model
from django.core.exceptions import ObjectDoesNotExist, ValidationError
from django.utils.translation import ugettext_lazy as _
from registration.forms import RegistrationForm
from .models import WhitelistedUsername
User = get_user_model()
class ExclusiveRegistrationForm(RegistrationForm):
def __init__(self, *args, **kwargs):
super(ExclusiveRegistrationForm, self).__init__(*args, **kwargs)
self.fields['username'].label = 'Student number'
def clean(self):
# TODO: try catch KeyError here to avoid empty form error
form_username = self.cleaned_data['username']
try:
# If this runs without raising an exception, then the username is in
# our database of whitelisted usernames.
WhitelistedUsername.objects.get(username=form_username.lower())
except ObjectDoesNotExist:
err = ValidationError(_('Unrecognised student number. Are you a CS1 student at UCT?s'), code='invalid')
self.add_error(User.USERNAME_FIELD, err)
super(ExclusiveRegistrationForm, self).clean()
|
from django.contrib.auth import get_user_model
from django.contrib.auth.forms import AuthenticationForm
from django.core.exceptions import ObjectDoesNotExist, ValidationError
from django.utils.translation import ugettext_lazy as _
from registration.forms import RegistrationForm
from .models import WhitelistedUsername
User = get_user_model()
class CustomAuthenticationForm(AuthenticationForm):
def __init__(self, *args, **kwargs):
super(CustomAuthenticationForm, self).__init__(*args, **kwargs)
self.fields['username'].label = 'Student number'
class ExclusiveRegistrationForm(RegistrationForm):
def __init__(self, *args, **kwargs):
super(ExclusiveRegistrationForm, self).__init__(*args, **kwargs)
self.fields['username'].label = 'Student number'
def clean(self):
# TODO: try catch KeyError here to avoid empty form error
form_username = self.cleaned_data['username']
try:
# If this runs without raising an exception, then the username is in
# our database of whitelisted usernames.
WhitelistedUsername.objects.get(username=form_username.lower())
except ObjectDoesNotExist:
err = ValidationError(_('Unrecognised student number. Are you a CS1 student at UCT?s'), code='invalid')
self.add_error(User.USERNAME_FIELD, err)
super(ExclusiveRegistrationForm, self).clean()
|
Add custom authentication form to set the label 'username' label to 'student number' for the login form.
|
Add custom authentication form to set the label 'username' label to 'student number' for the login form.
|
Python
|
mit
|
muhummadPatel/raspied,muhummadPatel/raspied,muhummadPatel/raspied
|
---
+++
@@ -1,4 +1,5 @@
from django.contrib.auth import get_user_model
+from django.contrib.auth.forms import AuthenticationForm
from django.core.exceptions import ObjectDoesNotExist, ValidationError
from django.utils.translation import ugettext_lazy as _
@@ -8,6 +9,12 @@
User = get_user_model()
+
+
+class CustomAuthenticationForm(AuthenticationForm):
+ def __init__(self, *args, **kwargs):
+ super(CustomAuthenticationForm, self).__init__(*args, **kwargs)
+ self.fields['username'].label = 'Student number'
class ExclusiveRegistrationForm(RegistrationForm):
|
48075a16190bbcc3d260dfa242a5553b129de8a8
|
tests/test_see.py
|
tests/test_see.py
|
#!/usr/bin/env python
# encoding: utf-8
"""
Unit tests for see.py
"""
from __future__ import print_function, unicode_literals
try:
import unittest2 as unittest
except ImportError:
import unittest
import see
class TestSee(unittest.TestCase):
def test_line_width(self):
# Arrange
default_width = 1
max_width = 1
# Act
width = see.line_width(default_width, max_width)
# Assert
self.assertIsInstance(width, int)
self.assertEqual(width, 1)
def test_regex_filter(self):
# Arrange
names = ["george", "helen"]
pat = "or*"
# Act
out = see.regex_filter(names, pat)
# Assert
self.assertIsInstance(out, tuple)
self.assertEqual(out, ("george",))
def test_fn_filter(self):
# Arrange
names = ["george", "helen"]
pat = "*or*"
# Act
out = see.fn_filter(names, pat)
# Assert
self.assertIsInstance(out, tuple)
self.assertEqual(out, ("george",))
def test_see_with_no_args(self):
# Act
out = see.see()
# Assert
self.assertIsInstance(out, see._SeeOutput)
if __name__ == '__main__':
unittest.main()
# End of file
|
#!/usr/bin/env python
# encoding: utf-8
"""
Unit tests for see.py
"""
from __future__ import print_function, unicode_literals
try:
import unittest2 as unittest
except ImportError:
import unittest
import os
import sys
sys.path.insert(0, os.path.dirname(__file__))
import see
class TestSee(unittest.TestCase):
def test_line_width(self):
# Arrange
default_width = 1
max_width = 1
# Act
width = see.line_width(default_width, max_width)
# Assert
self.assertIsInstance(width, int)
self.assertEqual(width, 1)
def test_regex_filter(self):
# Arrange
names = ["george", "helen"]
pat = "or*"
# Act
out = see.regex_filter(names, pat)
# Assert
self.assertIsInstance(out, tuple)
self.assertEqual(out, ("george",))
def test_fn_filter(self):
# Arrange
names = ["george", "helen"]
pat = "*or*"
# Act
out = see.fn_filter(names, pat)
# Assert
self.assertIsInstance(out, tuple)
self.assertEqual(out, ("george",))
def test_see_with_no_args(self):
# Act
out = see.see()
# Assert
self.assertIsInstance(out, see._SeeOutput)
if __name__ == '__main__':
unittest.main()
# End of file
|
Update tests to import see
|
Update tests to import see
|
Python
|
bsd-3-clause
|
araile/see
|
---
+++
@@ -8,6 +8,11 @@
import unittest2 as unittest
except ImportError:
import unittest
+
+import os
+import sys
+
+sys.path.insert(0, os.path.dirname(__file__))
import see
|
053a8f9deb8bfc0bb93cddcd48c8a7817bfe8c48
|
loom/analysis.py
|
loom/analysis.py
|
import six
from datetime import datetime, date
from collections import Mapping, Iterable
from jsonmapping.transforms import transliterate
IGNORE_FIELDS = ['$schema', '$sources', '$latin', '$text', '$attrcount',
'$linkcount', 'id']
def latinize(text):
""" Transliterate text to latin. """
if text is None or not len(text):
return text
return transliterate(text).lower()
def extract_text(data, sep=' : '):
""" Get all the instances of text from a given object, recursively. """
if isinstance(data, Mapping):
values = []
for k, v in data.items():
if k in IGNORE_FIELDS:
continue
values.append(v)
data = values
if isinstance(data, (date, datetime)):
data = data.isoformat()
elif isinstance(data, (int, float)):
data = six.text_type(data)
if isinstance(data, six.string_types):
return data
if isinstance(data, Iterable):
text = [extract_text(d, sep=sep) for d in data]
return sep.join([t for t in text if t is not None])
|
import six
from datetime import datetime, date
from collections import Mapping, Iterable
from jsonmapping.transforms import transliterate
IGNORE_FIELDS = ['$schema', '$sources', '$latin', '$text', '$attrcount',
'$linkcount', 'id']
def latinize(text):
""" Transliterate text to latin. """
if text is None or not len(text):
return text
return transliterate(text).lower()
def extract_text(data):
""" Get all the instances of text from a given object, recursively. """
if isinstance(data, Mapping):
values = []
for k, v in data.items():
if k in IGNORE_FIELDS:
continue
values.append(v)
data = values
if isinstance(data, (date, datetime)):
data = data.isoformat()
elif isinstance(data, (int, float)):
data = six.text_type(data)
if isinstance(data, six.string_types):
return [data]
if isinstance(data, Iterable):
values = []
for d in data:
values.extend(extract_text(d))
return values
|
Make text a list in the index.
|
Make text a list in the index.
|
Python
|
agpl-3.0
|
occrp/loom,occrp/datamapper
|
---
+++
@@ -15,7 +15,7 @@
return transliterate(text).lower()
-def extract_text(data, sep=' : '):
+def extract_text(data):
""" Get all the instances of text from a given object, recursively. """
if isinstance(data, Mapping):
values = []
@@ -29,7 +29,9 @@
elif isinstance(data, (int, float)):
data = six.text_type(data)
if isinstance(data, six.string_types):
- return data
+ return [data]
if isinstance(data, Iterable):
- text = [extract_text(d, sep=sep) for d in data]
- return sep.join([t for t in text if t is not None])
+ values = []
+ for d in data:
+ values.extend(extract_text(d))
+ return values
|
af79bd6dce28e8147994a8fe2afb4df742dcd3eb
|
client/test_server_proxy.py
|
client/test_server_proxy.py
|
"""Code snippet to test the Java ServerProxy interface to the Pings server."""
import ServerProxy, ClientInfo
sp = ServerProxy('localhost', 6543)
if False:
# Need to change permissions on ServerProxy Java class for this to work.
print 'Calling doJsonRequest directly...'
r = sp.doJsonRequest('/get_pings', None)
print r
print
info = ClientInfo()
pings = sp.getPings(info)
print pings.token
print pings.addresses[0]
print pings.geoip_info[0]
# Fill in results.
for i in range(len(pings.addresses)):
pings.results[i] = 'FOO %d' % i
print
print 'Submitting results'
sp.submitResults(info, pings)
|
"""Code snippet to test the Java ServerProxy interface to the Pings server."""
import ServerProxy, ClientInfo
sp = ServerProxy('localhost', 6543)
if False:
# Need to change permissions on ServerProxy Java class for this to work.
print 'Calling doJsonRequest directly...'
r = sp.doJsonRequest('/get_pings', None)
print r
print
info = ClientInfo()
pings = sp.getPings(info)
print 'Token', pings.token
print 'First address', pings.addresses[0]
print 'Geoip for first address', pings.geoip_info[0]
print 'Client Geoip', pings.client_geoip
# Fill in results.
for i in range(len(pings.addresses)):
pings.results[i] = 'FOO %d' % i
print
print 'Submitting results'
sp.submitResults(info, pings)
|
Add display of client geoip info. And print description of what we are printing.
|
Add display of client geoip info. And print description of what we are printing.
|
Python
|
bsd-3-clause
|
lisa-lab/pings,lisa-lab/pings,lisa-lab/pings,lisa-lab/pings
|
---
+++
@@ -13,9 +13,10 @@
print
info = ClientInfo()
pings = sp.getPings(info)
-print pings.token
-print pings.addresses[0]
-print pings.geoip_info[0]
+print 'Token', pings.token
+print 'First address', pings.addresses[0]
+print 'Geoip for first address', pings.geoip_info[0]
+print 'Client Geoip', pings.client_geoip
# Fill in results.
for i in range(len(pings.addresses)):
|
fe5edfe737a774aa86cce578321fbb7fb4c8795e
|
tagcache/utils.py
|
tagcache/utils.py
|
# -*- encoding: utf-8 -*-
import os
import errno
def ensure_intermediate_dir(path):
"""
Basiclly equivalent to command `mkdir -p`
"""
try:
os.makedirs(os.path.dirname(path))
except OSError, e:
if e.errno != errno.EEXIST:
raise e
def open_file(filename, flag, mode=0777):
"""
Wrapper of `os.open` which ensure intermediate dirs are created as well.
"""
try:
return os.open(filename, flag, mode=mode)
except OSError, e:
if e.errno != errno.ENOENT or not (flag & os.O_CREAT):
raise e
# a directory component not exists
ensure_intermediate_dir(filename)
# second try
return os.open(filename, flag, mode=mode)
def link_file(src, dst):
"""
Wrapper of `os.link` which ensure intermediate dirs are created as well.
"""
try:
return os.link(src, dst)
except OSError, e:
if e.errno != errno.ENOENT:
raise e
ensure_intermediate_dir(dst)
return os.link(src, dst)
def rename_file(old, new):
"""
Wrapper of `os.rename` which ensure intermediate dirs are created as well.
"""
try:
return os.rename(old, new)
except OSError, e:
if e.errno != errno.ENOENT:
raise e
ensure_intermediate_dir(new)
return os.rename(old, new)
|
# -*- encoding: utf-8 -*-
import os
import errno
def ensure_intermediate_dir(path):
"""
Basiclly equivalent to command `mkdir -p`
"""
try:
os.makedirs(os.path.dirname(path))
except OSError, e:
if e.errno != errno.EEXIST:
raise e
def open_file(filename, flag, mode=0777):
"""
Wrapper of `os.open` which ensure intermediate dirs are created as well.
"""
try:
return os.open(filename, flag, mode)
except OSError, e:
if e.errno != errno.ENOENT or not (flag & os.O_CREAT):
raise e
# a directory component not exists
ensure_intermediate_dir(filename)
# second try
return os.open(filename, flag, mode)
def link_file(src, dst):
"""
Wrapper of `os.link` which ensure intermediate dirs are created as well.
"""
try:
return os.link(src, dst)
except OSError, e:
if e.errno != errno.ENOENT:
raise e
ensure_intermediate_dir(dst)
return os.link(src, dst)
def rename_file(old, new):
"""
Wrapper of `os.rename` which ensure intermediate dirs are created as well.
"""
try:
return os.rename(old, new)
except OSError, e:
if e.errno != errno.ENOENT:
raise e
ensure_intermediate_dir(new)
return os.rename(old, new)
|
Fix a bug in file_open (os.open does not take keyword argument).
|
Fix a bug in file_open (os.open does not take keyword argument).
|
Python
|
mit
|
huangjunwen/tagcache
|
---
+++
@@ -28,7 +28,7 @@
"""
try:
- return os.open(filename, flag, mode=mode)
+ return os.open(filename, flag, mode)
except OSError, e:
@@ -40,7 +40,7 @@
ensure_intermediate_dir(filename)
# second try
- return os.open(filename, flag, mode=mode)
+ return os.open(filename, flag, mode)
def link_file(src, dst):
|
4ef8681f9dcd0f92be524925d3cacdae68c45616
|
tests/conftest.py
|
tests/conftest.py
|
# -*- coding: utf-8 -*-
from pytest import fixture
from iamport import Iamport
DEFAULT_TEST_IMP_KEY = 'imp_apikey'
DEFAULT_TEST_IMP_SECRET = ('ekKoeW8RyKuT0zgaZsUtXXTLQ4AhPFW3ZGseDA6bkA5lamv9O'
'qDMnxyeB9wqOsuO9W3Mx9YSJ4dTqJ3f')
def pytest_addoption(parser):
parser.addoption('--imp-key', default=DEFAULT_TEST_IMP_KEY,
help='iamport client key for testing '
'[default: %default]')
parser.addoption('--imp-secret', default=DEFAULT_TEST_IMP_SECRET,
help='iamport secret key for testing '
'[default: %default]')
@fixture
def iamport(request):
imp_key = request.config.getoption('--imp-key')
imp_secret = request.config.getoption('--imp-secret')
return Iamport(imp_key=imp_key, imp_secret=imp_secret)
|
# -*- coding: utf-8 -*-
from pytest import fixture
from iamport import Iamport
DEFAULT_TEST_IMP_KEY = 'imp_apikey'
DEFAULT_TEST_IMP_SECRET = (
'ekKoeW8RyKuT0zgaZsUtXXTLQ4AhPFW3ZGseDA6b'
'kA5lamv9OqDMnxyeB9wqOsuO9W3Mx9YSJ4dTqJ3f'
)
def pytest_addoption(parser):
parser.addoption(
'--imp-key',
default=DEFAULT_TEST_IMP_KEY,
help='iamport client key for testing '
'[default: %(default)s]'
)
parser.addoption(
'--imp-secret',
default=DEFAULT_TEST_IMP_SECRET,
help='iamport secret key for testing '
'[default: %(default)s]'
)
@fixture
def iamport(request):
imp_key = request.config.getoption('--imp-key')
imp_secret = request.config.getoption('--imp-secret')
return Iamport(imp_key=imp_key, imp_secret=imp_secret)
|
Change %default to %(default)s for removing warning
|
Change %default to %(default)s for removing warning
|
Python
|
mit
|
iamport/iamport-rest-client-python
|
---
+++
@@ -3,18 +3,27 @@
from iamport import Iamport
+
DEFAULT_TEST_IMP_KEY = 'imp_apikey'
-DEFAULT_TEST_IMP_SECRET = ('ekKoeW8RyKuT0zgaZsUtXXTLQ4AhPFW3ZGseDA6bkA5lamv9O'
- 'qDMnxyeB9wqOsuO9W3Mx9YSJ4dTqJ3f')
+DEFAULT_TEST_IMP_SECRET = (
+ 'ekKoeW8RyKuT0zgaZsUtXXTLQ4AhPFW3ZGseDA6b'
+ 'kA5lamv9OqDMnxyeB9wqOsuO9W3Mx9YSJ4dTqJ3f'
+)
def pytest_addoption(parser):
- parser.addoption('--imp-key', default=DEFAULT_TEST_IMP_KEY,
- help='iamport client key for testing '
- '[default: %default]')
- parser.addoption('--imp-secret', default=DEFAULT_TEST_IMP_SECRET,
- help='iamport secret key for testing '
- '[default: %default]')
+ parser.addoption(
+ '--imp-key',
+ default=DEFAULT_TEST_IMP_KEY,
+ help='iamport client key for testing '
+ '[default: %(default)s]'
+ )
+ parser.addoption(
+ '--imp-secret',
+ default=DEFAULT_TEST_IMP_SECRET,
+ help='iamport secret key for testing '
+ '[default: %(default)s]'
+ )
@fixture
|
3a1615238d4500f0fa7b9eea9ee2bfe460bc21f9
|
cax/tasks/purity.py
|
cax/tasks/purity.py
|
"""Add electron lifetime
"""
from sympy.parsing.sympy_parser import parse_expr
from pax import units
from cax import config
from cax.task import Task
class AddElectronLifetime(Task):
"Add electron lifetime to dataset"
def __init__(self):
self.collection_purity = config.mongo_collection('purity')
Task.__init__(self)
def each_run(self):
if 'processor' in self.run_doc:
return
# Fetch the latest electron lifetime fit
doc = self.collection_purity.find_one(sort=(('calculation_time',
-1),))
function = parse_expr(doc['electron_lifetime_function'])
# Compute value from this function on this dataset
lifetime = function.evalf(subs={"t" : self.run_doc['start'].timestamp()})
run_number = self.run_doc['number']
self.log.info("Run %d: calculated lifetime of %d us" % (run_number,
lifetime))
if not config.DATABASE_LOG:
return
# Update run database
key = 'processor.DEFAULT.electron_lifetime_liquid'
self.collection.find_and_modify({'_id': self.run_doc['_id']},
{'$set': {key: lifetime * units.us}})
|
"""Add electron lifetime
"""
from sympy.parsing.sympy_parser import parse_expr
from pax import units
from cax import config
from cax.task import Task
class AddElectronLifetime(Task):
"Add electron lifetime to dataset"
def __init__(self):
self.collection_purity = config.mongo_collection('purity')
Task.__init__(self)
def each_run(self):
if 'processor' in self.run_doc:
return
# Fetch the latest electron lifetime fit
doc = self.collection_purity.find_one(sort=(('calculation_time',
-1),))
function = parse_expr(doc['electron_lifetime_function'])
# Compute value from this function on this dataset
lifetime = function.evalf(subs={"t" : self.run_doc['start'].timestamp()})
lifetime = float(lifetime) # Convert away from Sympy type.
run_number = self.run_doc['number']
self.log.info("Run %d: calculated lifetime of %d us" % (run_number,
lifetime))
if not config.DATABASE_LOG:
return
# Update run database
key = 'processor.DEFAULT.electron_lifetime_liquid'
self.collection.find_and_modify({'_id': self.run_doc['_id']},
{'$set': {key: lifetime * units.us}})
|
Convert lifetime to float instead of sympy type.
|
Convert lifetime to float instead of sympy type.
|
Python
|
isc
|
XENON1T/cax,XENON1T/cax
|
---
+++
@@ -28,6 +28,7 @@
# Compute value from this function on this dataset
lifetime = function.evalf(subs={"t" : self.run_doc['start'].timestamp()})
+ lifetime = float(lifetime) # Convert away from Sympy type.
run_number = self.run_doc['number']
self.log.info("Run %d: calculated lifetime of %d us" % (run_number,
|
46511322dc8d738cc43561025bca3298946da2e6
|
server.py
|
server.py
|
from swiftdav.swiftdav import SwiftProvider, WsgiDAVDomainController
from waitress import serve
from wsgidav.wsgidav_app import DEFAULT_CONFIG, WsgiDAVApp
proxy = 'http://127.0.0.1:8080/auth/v1.0'
insecure = False # Set to True to disable SSL certificate validation
config = DEFAULT_CONFIG.copy()
config.update({
"provider_mapping": {"": SwiftProvider()},
"verbose": 1,
"propsmanager": True,
"locksmanager": True,
"acceptbasic": True,
"acceptdigest": False,
"defaultdigest": False,
"domaincontroller": WsgiDAVDomainController(proxy, insecure)
})
app = WsgiDAVApp(config)
serve(app, host="0.0.0.0", port=8000)
|
from swiftdav.swiftdav import SwiftProvider, WsgiDAVDomainController
from waitress import serve
from wsgidav.wsgidav_app import DEFAULT_CONFIG, WsgiDAVApp
proxy = 'http://127.0.0.1:8080/auth/v1.0'
insecure = False # Set to True to disable SSL certificate validation
config = DEFAULT_CONFIG.copy()
config.update({
"provider_mapping": {"": SwiftProvider()},
"verbose": 1,
"propsmanager": True,
"locksmanager": True,
"acceptbasic": True,
"acceptdigest": False,
"defaultdigest": False,
"domaincontroller": WsgiDAVDomainController(proxy, insecure)
})
app = WsgiDAVApp(config)
serve(app, host="0.0.0.0", port=8000, max_request_body_size=5*1024*1024*1024)
|
Increase waitress setting max_request_body_size to 5GiB
|
Increase waitress setting max_request_body_size to 5GiB
Python waitress limits the body size to 1GiB by default, thus
uploading of larger objects will fail if this value is not
increased.
Please note that this value should be increased if your Swift
cluster supports uploading of objects larger than 5GiB.
|
Python
|
apache-2.0
|
cschwede/swiftdav,cschwede/swiftdav
|
---
+++
@@ -18,4 +18,4 @@
})
app = WsgiDAVApp(config)
-serve(app, host="0.0.0.0", port=8000)
+serve(app, host="0.0.0.0", port=8000, max_request_body_size=5*1024*1024*1024)
|
5ef05f6c75e3dd0febffd527aebfdf8c2ed6bb51
|
server.py
|
server.py
|
import argparse
import json
from flask import Flask, request
parser = argparse.ArgumentParser(description="Start a Blindstore server.")
parser.add_argument('-d', '--debug', action='store_true',
help="enable Flask debug mode. DO NOT use in production.")
args = parser.parse_args()
NUM_RECORDS = 5
RECORD_SIZE = 64
app = Flask(__name__)
@app.route('/db_size')
def get_db_size():
return json.dumps({'num_records': NUM_RECORDS, 'record_size': RECORD_SIZE})
@app.route('/retrieve', methods=['POST'])
def get():
public_key = request.form['PUBLIC_KEY']
enc_index = request.form['ENC_INDEX']
return "/retrieve index '{index}' with key '{key}'".format(index=enc_index, key=public_key)
@app.route('/set', methods=['POST'])
def put():
enc_index = request.form['ENC_INDEX']
enc_data = request.form['ENC_DATA']
return "/set '{index}' to '{data}'".format(data=enc_data, index=enc_index)
if __name__ == '__main__':
app.run(debug=args.debug)
|
import argparse
import json
from flask import Flask, request
parser = argparse.ArgumentParser(description="Start a Blindstore server.")
parser.add_argument('-d', '--debug', action='store_true',
help="enable Flask debug mode. DO NOT use in production.")
args = parser.parse_args()
NUM_RECORDS = 5
RECORD_SIZE = 64
app = Flask(__name__)
@app.route('/db_size')
def get_db_size():
return json.dumps({'num_records': NUM_RECORDS, 'record_size': RECORD_SIZE}), \
200, {'Content-Type': 'text/json'}
@app.route('/retrieve', methods=['POST'])
def get():
public_key = request.form['PUBLIC_KEY']
enc_index = request.form['ENC_INDEX']
return "/retrieve index '{index}' with key '{key}'".format(index=enc_index, key=public_key)
@app.route('/set', methods=['POST'])
def put():
enc_index = request.form['ENC_INDEX']
enc_data = request.form['ENC_DATA']
return "/set '{index}' to '{data}'".format(data=enc_data, index=enc_index)
if __name__ == '__main__':
app.run(debug=args.debug)
|
Return Content-Type header with JSON size info
|
Return Content-Type header with JSON size info
|
Python
|
mit
|
blindstore/blindstore-old-scarab
|
---
+++
@@ -15,7 +15,8 @@
@app.route('/db_size')
def get_db_size():
- return json.dumps({'num_records': NUM_RECORDS, 'record_size': RECORD_SIZE})
+ return json.dumps({'num_records': NUM_RECORDS, 'record_size': RECORD_SIZE}), \
+ 200, {'Content-Type': 'text/json'}
@app.route('/retrieve', methods=['POST'])
def get():
|
0f7ebf148ab3f88fc983e60f689a9c740ae64e47
|
outgoing_mail.py
|
outgoing_mail.py
|
#!/usr/bin/env python
#
# Copyright 2010 Eric Entzel <eric@ubermac.net>
#
from google.appengine.api import mail
from google.appengine.ext.webapp import template
import os
from_address = 'admin@' + os.environ['APPLICATION_ID'] + '.appspotmail.com'
def send(to, template_name, values):
path = os.path.join(os.path.dirname(__file__), 'email_templates', template_name)
message = mail.EmailMessage(sender=from_address, to=to)
message.subject = template.render(path + '.subject', values)
message.body = template.render(path + '.body', values)
message.send()
|
#!/usr/bin/env python
#
# Copyright 2010 Eric Entzel <eric@ubermac.net>
#
from google.appengine.api import mail
from google.appengine.ext.webapp import template
import os
from_address = 'EventBot <admin@' + os.environ['APPLICATION_ID'] + '.appspotmail.com>'
def send(to, template_name, values):
path = os.path.join(os.path.dirname(__file__), 'email_templates', template_name)
message = mail.EmailMessage(sender=from_address, to=to)
message.subject = template.render(path + '.subject', values)
message.body = template.render(path + '.body', values)
message.send()
|
Add display name for from address
|
Add display name for from address
|
Python
|
mit
|
eentzel/myeventbot,eentzel/myeventbot,eentzel/myeventbot,eentzel/myeventbot,eentzel/myeventbot
|
---
+++
@@ -8,7 +8,7 @@
import os
-from_address = 'admin@' + os.environ['APPLICATION_ID'] + '.appspotmail.com'
+from_address = 'EventBot <admin@' + os.environ['APPLICATION_ID'] + '.appspotmail.com>'
def send(to, template_name, values):
|
3822b5b142d54f83aadf7e366f2b5b925f557e1a
|
test/testUtils/__init__.py
|
test/testUtils/__init__.py
|
import ibmiotf.application
import os
class AbstractTest(object):
WIOTP_API_KEY=os.getenv("WIOTP_API_KEY")
WIOTP_API_TOKEN=os.getenv("WIOTP_API_TOKEN")
ORG_ID = os.getenv("WIOTP_ORG_ID")
appOptions = {'auth-key': WIOTP_API_KEY, 'auth-token': WIOTP_API_TOKEN}
setupAppClient = ibmiotf.application.Client(appOptions)
|
import ibmiotf.application
import os
class AbstractTest(object):
WIOTP_API_KEY=os.getenv("WIOTP_API_KEY")
WIOTP_API_TOKEN=os.getenv("WIOTP_API_TOKEN")
ORG_ID = os.getenv("WIOTP_ORG_ID")
if WIOTP_API_KEY is None:
raise Exception("WIOTP_API_KEY environment variable is not set")
if WIOTP_API_TOKEN is None:
raise Exception("WIOTP_API_TOKEN environment variable is not set")
if ORG_ID is None:
raise Exception("WIOTP_ORG_ID environment variable is not set")
appOptions = {'auth-key': WIOTP_API_KEY, 'auth-token': WIOTP_API_TOKEN}
setupAppClient = ibmiotf.application.Client(appOptions)
|
Make tests throw better error if env vars are missing
|
Make tests throw better error if env vars are missing
|
Python
|
epl-1.0
|
ibm-watson-iot/iot-python,ibm-watson-iot/iot-python,ibm-messaging/iot-python
|
---
+++
@@ -7,5 +7,12 @@
WIOTP_API_TOKEN=os.getenv("WIOTP_API_TOKEN")
ORG_ID = os.getenv("WIOTP_ORG_ID")
+ if WIOTP_API_KEY is None:
+ raise Exception("WIOTP_API_KEY environment variable is not set")
+ if WIOTP_API_TOKEN is None:
+ raise Exception("WIOTP_API_TOKEN environment variable is not set")
+ if ORG_ID is None:
+ raise Exception("WIOTP_ORG_ID environment variable is not set")
+
appOptions = {'auth-key': WIOTP_API_KEY, 'auth-token': WIOTP_API_TOKEN}
setupAppClient = ibmiotf.application.Client(appOptions)
|
b853abc579f5dfaab896cf57c39268a36c109a83
|
tests/test_address_book.py
|
tests/test_address_book.py
|
from unittest import TestCase
class AddressBookTestCase(TestCase):
def test_add_person(self):
pass
def test_add_group(self):
pass
def test_find_person_by_first_name(self):
pass
def test_find_person_by_last_name(self):
pass
def test_find_person_by_email(self):
passjjj
|
from unittest import TestCase
class AddressBookTestCase(TestCase):
def test_add_person(self):
person = Person(
'John',
'Doe',
['Russian Federation, Kemerovo region, Kemerovo, Kirova street 23, apt. 42'],
['+79834772053']
)
self.address_book.add_person(person)
self.assertIn(person, self.address_book)
def test_add_group(self):
pass
def test_find_person_by_first_name(self):
pass
def test_find_person_by_last_name(self):
pass
def test_find_person_by_email(self):
passjjj
|
Test person adding to addressbook
|
Test person adding to addressbook
|
Python
|
mit
|
dizpers/python-address-book-assignment
|
---
+++
@@ -4,7 +4,14 @@
class AddressBookTestCase(TestCase):
def test_add_person(self):
- pass
+ person = Person(
+ 'John',
+ 'Doe',
+ ['Russian Federation, Kemerovo region, Kemerovo, Kirova street 23, apt. 42'],
+ ['+79834772053']
+ )
+ self.address_book.add_person(person)
+ self.assertIn(person, self.address_book)
def test_add_group(self):
pass
|
097eae49564a8eefd66d903d8e8cd900054ef147
|
characters/views.py
|
characters/views.py
|
from django.shortcuts import get_object_or_404, redirect, render
from django.views import generic
from characters.forms import CharacterForm
from characters.models import Character, Class, Race
class CharacterIndexView(generic.ListView):
template_name = 'characters/index.html'
context_object_name = 'all_characters' # better than 'object_list'
def get_queryset(self):
return Character.objects.all()
class CharacterDetailView(generic.DetailView):
model = Character
template_name = 'characters/view_character.html'
def create_character(request):
form = CharacterForm(request.POST or None)
if request.method == 'POST' and form.is_valid():
character = Character(
name=request.POST['name'],
background=request.POST['background'],
race_id=1,
cclass_id=1
)
character.save()
return redirect('characters:view', character_id=character.id)
context = {'form': form}
return render(request, 'characters/create_character.html', context)
|
from django.shortcuts import get_object_or_404, redirect, render
from django.views import generic
from characters.forms import CharacterForm
from characters.models import Character, Class, Race
class CharacterIndexView(generic.ListView):
template_name = 'characters/index.html'
context_object_name = 'all_characters' # better than 'object_list'
def get_queryset(self):
return Character.objects.all().order_by('name')
class CharacterDetailView(generic.DetailView):
model = Character
template_name = 'characters/view_character.html'
def create_character(request):
form = CharacterForm(request.POST or None)
if request.method == 'POST' and form.is_valid():
character = Character(
name=request.POST['name'],
background=request.POST['background'],
race_id=1,
cclass_id=1
)
character.save()
return redirect('characters:view', character_id=character.id)
context = {'form': form}
return render(request, 'characters/create_character.html', context)
|
Order character listing by name
|
Order character listing by name
|
Python
|
mit
|
mpirnat/django-tutorial-v2
|
---
+++
@@ -11,7 +11,7 @@
context_object_name = 'all_characters' # better than 'object_list'
def get_queryset(self):
- return Character.objects.all()
+ return Character.objects.all().order_by('name')
class CharacterDetailView(generic.DetailView):
|
fefde8aef88cbfb13cb1f0bfcd3ac476ad7a903c
|
spacy/download.py
|
spacy/download.py
|
from __future__ import print_function
import sys
import sputnik
from sputnik.package_list import (PackageNotFoundException,
CompatiblePackageNotFoundException)
from . import about
def download(lang, force=False, fail_on_exist=True):
if force:
sputnik.purge(about.__title__, about.__version__)
try:
sputnik.package(about.__title__, about.__version__,
about.__models__.get(lang, lang))
if fail_on_exist:
print("Model already installed. Please run 'python -m "
"spacy.%s.download --force' to reinstall." % lang, file=sys.stderr)
sys.exit(0)
except (PackageNotFoundException, CompatiblePackageNotFoundException):
pass
package = sputnik.install(about.__title__, about.__version__,
about.__models__.get(lang, lang))
try:
sputnik.package(about.__title__, about.__version__,
about.__models__.get(lang, lang))
except (PackageNotFoundException, CompatiblePackageNotFoundException):
print("Model failed to install. Please run 'python -m "
"spacy.%s.download --force'." % lang, file=sys.stderr)
sys.exit(1)
print("Model successfully installed.", file=sys.stderr)
|
from __future__ import print_function
import sys
import sputnik
from sputnik.package_list import (PackageNotFoundException,
CompatiblePackageNotFoundException)
from . import about
from . import util
def download(lang, force=False, fail_on_exist=True):
if force:
sputnik.purge(about.__title__, about.__version__)
try:
sputnik.package(about.__title__, about.__version__,
about.__models__.get(lang, lang))
if fail_on_exist:
print("Model already installed. Please run 'python -m "
"spacy.%s.download --force' to reinstall." % lang, file=sys.stderr)
sys.exit(0)
except (PackageNotFoundException, CompatiblePackageNotFoundException):
pass
package = sputnik.install(about.__title__, about.__version__,
about.__models__.get(lang, lang))
try:
sputnik.package(about.__title__, about.__version__,
about.__models__.get(lang, lang))
except (PackageNotFoundException, CompatiblePackageNotFoundException):
print("Model failed to install. Please run 'python -m "
"spacy.%s.download --force'." % lang, file=sys.stderr)
sys.exit(1)
data_path = util.get_data_path()
print("Model successfully installed to %s" % data_path, file=sys.stderr)
|
Make installation print data path.
|
Make installation print data path.
|
Python
|
mit
|
explosion/spaCy,explosion/spaCy,aikramer2/spaCy,recognai/spaCy,aikramer2/spaCy,recognai/spaCy,oroszgy/spaCy.hu,Gregory-Howard/spaCy,recognai/spaCy,Gregory-Howard/spaCy,recognai/spaCy,Gregory-Howard/spaCy,banglakit/spaCy,oroszgy/spaCy.hu,oroszgy/spaCy.hu,spacy-io/spaCy,spacy-io/spaCy,raphael0202/spaCy,banglakit/spaCy,explosion/spaCy,honnibal/spaCy,recognai/spaCy,oroszgy/spaCy.hu,raphael0202/spaCy,Gregory-Howard/spaCy,aikramer2/spaCy,aikramer2/spaCy,aikramer2/spaCy,honnibal/spaCy,spacy-io/spaCy,explosion/spaCy,banglakit/spaCy,Gregory-Howard/spaCy,oroszgy/spaCy.hu,honnibal/spaCy,banglakit/spaCy,raphael0202/spaCy,raphael0202/spaCy,Gregory-Howard/spaCy,spacy-io/spaCy,recognai/spaCy,aikramer2/spaCy,explosion/spaCy,raphael0202/spaCy,banglakit/spaCy,explosion/spaCy,oroszgy/spaCy.hu,spacy-io/spaCy,banglakit/spaCy,spacy-io/spaCy,honnibal/spaCy,raphael0202/spaCy
|
---
+++
@@ -7,6 +7,7 @@
CompatiblePackageNotFoundException)
from . import about
+from . import util
def download(lang, force=False, fail_on_exist=True):
@@ -34,4 +35,5 @@
"spacy.%s.download --force'." % lang, file=sys.stderr)
sys.exit(1)
- print("Model successfully installed.", file=sys.stderr)
+ data_path = util.get_data_path()
+ print("Model successfully installed to %s" % data_path, file=sys.stderr)
|
267a768bd1ccc87c3c1f54c4ac520a0e12e5fd5f
|
moksha/tests/test_clientsockets.py
|
moksha/tests/test_clientsockets.py
|
import webtest
import moksha.tests.utils as testutils
from moksha.api.widgets.live import get_moksha_socket
from moksha.middleware import make_moksha_middleware
from tw2.core import make_middleware as make_tw2_middleware
class TestClientSocketDumb:
def _setUp(self):
def kernel(config):
def app(environ, start_response):
start_response('200 OK', [('Content-Type', 'text/html')])
socket = get_moksha_socket(config)
return map(str, [socket.display()])
app = make_moksha_middleware(app, config)
app = make_tw2_middleware(app, config)
app = webtest.TestApp(app)
self.app = app
for _setup, name in testutils.make_setup_functions(kernel):
yield _setup, name
def _tearDown(self):
pass
@testutils.crosstest
def test_middleware_wrap(self):
targets = ['moksha_websocket', 'TCPSocket']
response = self.app.get('/')
assert(any([target in response for target in targets]))
|
import webtest
import moksha.tests.utils as testutils
from moksha.api.widgets.live import get_moksha_socket
from moksha.middleware import make_moksha_middleware
from tw2.core import make_middleware as make_tw2_middleware
class TestClientSocketDumb:
def _setUp(self):
def kernel(config):
def app(environ, start_response):
start_response('200 OK', [('Content-Type', 'text/html')])
socket = get_moksha_socket(config)
return map(str, [socket.display()])
app = make_moksha_middleware(app, config)
app = make_tw2_middleware(app, config)
app = webtest.TestApp(app)
self.app = app
for _setup, name in testutils.make_setup_functions(kernel):
yield _setup, name
def _tearDown(self):
pass
@testutils.crosstest
def test_has_socket_str(self):
targets = ['moksha_websocket', 'TCPSocket']
response = self.app.get('/')
assert(any([target in response for target in targets]))
|
Rename test. Fix copy/pasta forgetfulness.
|
Rename test. Fix copy/pasta forgetfulness.
|
Python
|
apache-2.0
|
pombredanne/moksha,mokshaproject/moksha,mokshaproject/moksha,mokshaproject/moksha,pombredanne/moksha,pombredanne/moksha,mokshaproject/moksha,pombredanne/moksha
|
---
+++
@@ -27,7 +27,7 @@
pass
@testutils.crosstest
- def test_middleware_wrap(self):
+ def test_has_socket_str(self):
targets = ['moksha_websocket', 'TCPSocket']
response = self.app.get('/')
assert(any([target in response for target in targets]))
|
68cec560ad108e0e6a081ed92aab2f06a7545821
|
tests/conftest.py
|
tests/conftest.py
|
import pytest
@pytest.fixture(autouse=True)
def tagschecker(request):
tags = set(request.config.getini('TAGS'))
tags_marker = request.node.get_marker('tags')
xfailtags_marker = request.node.get_marker('xfailtags')
skiptags_marker = request.node.get_marker('skiptags')
if xfailtags_marker and not tags.isdisjoint(set(xfailtags_marker.args)):
request.node.add_marker(pytest.mark.xfail())
elif (
tags_marker and tags.isdisjoint(set(tags_marker.args)) or
skiptags_marker and not tags.isdisjoint(set(skiptags_marker.args))
):
pytest.skip('skipped for this tags: {}'.format(tags))
|
import pytest
from docker import Client
@pytest.fixture(scope="session")
def docker_client():
client = Client(base_url='unix://var/run/docker.sock', timeout=180)
return client
@pytest.fixture(autouse=True)
def tagschecker(request):
tags = set(request.config.getini('TAGS'))
tags_marker = request.node.get_marker('tags')
xfailtags_marker = request.node.get_marker('xfailtags')
skiptags_marker = request.node.get_marker('skiptags')
if xfailtags_marker and not tags.isdisjoint(set(xfailtags_marker.args)):
request.node.add_marker(pytest.mark.xfail())
elif (
tags_marker and tags.isdisjoint(set(tags_marker.args)) or
skiptags_marker and not tags.isdisjoint(set(skiptags_marker.args))
):
pytest.skip('skipped for this tags: {}'.format(tags))
|
Increase docker-py timeout to 180
|
Increase docker-py timeout to 180
|
Python
|
mit
|
dincamihai/salt-toaster,dincamihai/salt-toaster
|
---
+++
@@ -1,4 +1,11 @@
import pytest
+from docker import Client
+
+
+@pytest.fixture(scope="session")
+def docker_client():
+ client = Client(base_url='unix://var/run/docker.sock', timeout=180)
+ return client
@pytest.fixture(autouse=True)
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.