commit
stringlengths
40
40
old_file
stringlengths
4
150
new_file
stringlengths
4
150
old_contents
stringlengths
0
3.26k
new_contents
stringlengths
1
4.43k
subject
stringlengths
15
501
message
stringlengths
15
4.06k
lang
stringclasses
4 values
license
stringclasses
13 values
repos
stringlengths
5
91.5k
diff
stringlengths
0
4.35k
5d98c4109322d3346609fdb1bf44cfd4a16dc94c
project_management/tests/test_main.py
project_management/tests/test_main.py
""" Test main pm functionality """ from cement.utils import test from pmtools import PmController, PmApp from test_default import config_defaults, PmTestApp class PmMainTest(test.CementTestCase): app_class = PmTestApp def test_1_help(self): app = self.make_app(argv=['--help']) app.setup() print dir(app) print app.argv print dir(app.config) print dir(app._meta) print app._meta.label print app._meta.base_controller.Meta.label app.run() app.close()
Add tests for main app
Add tests for main app
Python
mit
SciLifeLab/scilifelab,kate-v-stepanova/scilifelab,senthil10/scilifelab,senthil10/scilifelab,SciLifeLab/scilifelab,jun-wan/scilifelab,SciLifeLab/scilifelab,jun-wan/scilifelab,senthil10/scilifelab,jun-wan/scilifelab,kate-v-stepanova/scilifelab,SciLifeLab/scilifelab,kate-v-stepanova/scilifelab,kate-v-stepanova/scilifelab,jun-wan/scilifelab,senthil10/scilifelab
--- +++ @@ -0,0 +1,23 @@ +""" +Test main pm functionality +""" + +from cement.utils import test +from pmtools import PmController, PmApp +from test_default import config_defaults, PmTestApp + +class PmMainTest(test.CementTestCase): + app_class = PmTestApp + + def test_1_help(self): + app = self.make_app(argv=['--help']) + app.setup() + print dir(app) + print app.argv + print dir(app.config) + print dir(app._meta) + print app._meta.label + print app._meta.base_controller.Meta.label + app.run() + app.close() +
68b5e3438b771afe4a70019a12fdd6d60a587c7e
py/1-bit-and-2-bit-characters.py
py/1-bit-and-2-bit-characters.py
class Solution(object): def isOneBitCharacter(self, bits): """ :type bits: List[int] :rtype: bool """ s0, s1 = True, False for i, b in enumerate(bits[:-1]): if b == 1: s0, s1 = s1 and bits[i - 1] == 1, s0 else: s0, s1 = True, False return s0
Add py solution for 717. 1-bit and 2-bit Characters
Add py solution for 717. 1-bit and 2-bit Characters 717. 1-bit and 2-bit Characters: https://leetcode.com/problems/1-bit-and-2-bit-characters/
Python
apache-2.0
ckclark/leetcode,ckclark/leetcode,ckclark/leetcode,ckclark/leetcode,ckclark/leetcode,ckclark/leetcode
--- +++ @@ -0,0 +1,14 @@ +class Solution(object): + def isOneBitCharacter(self, bits): + """ + :type bits: List[int] + :rtype: bool + """ + s0, s1 = True, False + for i, b in enumerate(bits[:-1]): + if b == 1: + s0, s1 = s1 and bits[i - 1] == 1, s0 + else: + s0, s1 = True, False + return s0 +
0738a6a4739db2a34659e469b4d262a971bc4938
heppyplotlib/errorcalc.py
heppyplotlib/errorcalc.py
"""Functions for calculating errors by combining different datasets.""" import math def asymmetric_hessian_error(value_lists): """Calculate the asymmetric hessian error from a list of datasets, where the first dataset stems from a PDF CV run.""" lows = [] highs = [] transposed_value_lists = zip(*value_lists) for values in transposed_value_lists: central_value = values[0] evs = values[1:] error = [0.0, 0.0] for i in range(0, len(evs), 2): (ev_p, ev_m) = evs[i:i+2] error[0] += max(ev_p - central_value, ev_m - central_value, 0)**2 error[1] += max(central_value - ev_p, central_value - ev_m, 0)**2 high = central_value + math.sqrt(error[0]) low = central_value - math.sqrt(error[1]) highs.append(high) lows.append(low) return (lows, highs)
Add a calc function for asymmetric hessian errors
Add a calc function for asymmetric hessian errors
Python
mit
ebothmann/heppyplotlib
--- +++ @@ -0,0 +1,23 @@ +"""Functions for calculating errors by combining different datasets.""" + +import math + +def asymmetric_hessian_error(value_lists): + """Calculate the asymmetric hessian error from a list of datasets, + where the first dataset stems from a PDF CV run.""" + lows = [] + highs = [] + transposed_value_lists = zip(*value_lists) + for values in transposed_value_lists: + central_value = values[0] + evs = values[1:] + error = [0.0, 0.0] + for i in range(0, len(evs), 2): + (ev_p, ev_m) = evs[i:i+2] + error[0] += max(ev_p - central_value, ev_m - central_value, 0)**2 + error[1] += max(central_value - ev_p, central_value - ev_m, 0)**2 + high = central_value + math.sqrt(error[0]) + low = central_value - math.sqrt(error[1]) + highs.append(high) + lows.append(low) + return (lows, highs)
aeb657dae9decb0ea25361e530fdbe0399a1650e
tests/test_validators.py
tests/test_validators.py
""" test_validators ~~~~~~~~~~~~~~ Unittests for bundled validators. :copyright: 2007-2008 by James Crasta, Thomas Johansson. :license: MIT, see LICENSE.txt for details. """ from py.test import raises from wtforms.validators import ValidationError, length, url, not_empty, email, ip_address class DummyForm(object): pass class DummyField(object): def __init__(self, data): self.data = data form = DummyForm() def test_email(): assert email(form, DummyField('foo@bar.dk')) == None assert email(form, DummyField('123@bar.dk')) == None assert email(form, DummyField('foo@456.dk')) == None assert email(form, DummyField('foo@bar456.info')) == None raises(ValidationError, email, form, DummyField('foo')) == None raises(ValidationError, email, form, DummyField('bar.dk')) == None raises(ValidationError, email, form, DummyField('foo@')) == None raises(ValidationError, email, form, DummyField('@bar.dk')) == None raises(ValidationError, email, form, DummyField('foo@bar')) == None raises(ValidationError, email, form, DummyField('foo@bar.ab12')) == None raises(ValidationError, email, form, DummyField('foo@bar.abcde')) == None def test_length(): field = DummyField('foobar') assert length(min=2, max=6)(form, field) == None raises(ValidationError, length(min=7), form, field) raises(ValidationError, length(max=5), form, field) def test_url(): assert url()(form, DummyField('http://foobar.dk')) == None assert url()(form, DummyField('http://foobar.dk/')) == None assert url()(form, DummyField('http://foobar.dk/foobar')) == None raises(ValidationError, url(), form, DummyField('http://foobar')) raises(ValidationError, url(), form, DummyField('foobar.dk')) raises(ValidationError, url(), form, DummyField('http://foobar.12')) def test_not_empty(): assert not_empty()(form, DummyField('foobar')) == None raises(ValidationError, not_empty(), form, DummyField('')) raises(ValidationError, not_empty(), form, DummyField(' ')) def test_ip_address(): assert ip_address(form, DummyField('127.0.0.1')) == None raises(ValidationError, ip_address, form, DummyField('abc.0.0.1')) raises(ValidationError, ip_address, form, DummyField('1278.0.0.1')) raises(ValidationError, ip_address, form, DummyField('127.0.0.abc'))
Add first basic unittests using py.test
Add first basic unittests using py.test
Python
bsd-3-clause
mfa/wtforms-clone,maxcountryman/wtforms,mfa/wtforms-clone
--- +++ @@ -0,0 +1,60 @@ +""" + test_validators + ~~~~~~~~~~~~~~ + + Unittests for bundled validators. + + :copyright: 2007-2008 by James Crasta, Thomas Johansson. + :license: MIT, see LICENSE.txt for details. +""" + +from py.test import raises +from wtforms.validators import ValidationError, length, url, not_empty, email, ip_address + +class DummyForm(object): + pass + +class DummyField(object): + def __init__(self, data): + self.data = data + +form = DummyForm() + +def test_email(): + assert email(form, DummyField('foo@bar.dk')) == None + assert email(form, DummyField('123@bar.dk')) == None + assert email(form, DummyField('foo@456.dk')) == None + assert email(form, DummyField('foo@bar456.info')) == None + raises(ValidationError, email, form, DummyField('foo')) == None + raises(ValidationError, email, form, DummyField('bar.dk')) == None + raises(ValidationError, email, form, DummyField('foo@')) == None + raises(ValidationError, email, form, DummyField('@bar.dk')) == None + raises(ValidationError, email, form, DummyField('foo@bar')) == None + raises(ValidationError, email, form, DummyField('foo@bar.ab12')) == None + raises(ValidationError, email, form, DummyField('foo@bar.abcde')) == None + +def test_length(): + field = DummyField('foobar') + assert length(min=2, max=6)(form, field) == None + raises(ValidationError, length(min=7), form, field) + raises(ValidationError, length(max=5), form, field) + +def test_url(): + assert url()(form, DummyField('http://foobar.dk')) == None + assert url()(form, DummyField('http://foobar.dk/')) == None + assert url()(form, DummyField('http://foobar.dk/foobar')) == None + raises(ValidationError, url(), form, DummyField('http://foobar')) + raises(ValidationError, url(), form, DummyField('foobar.dk')) + raises(ValidationError, url(), form, DummyField('http://foobar.12')) + +def test_not_empty(): + assert not_empty()(form, DummyField('foobar')) == None + raises(ValidationError, not_empty(), form, DummyField('')) + raises(ValidationError, not_empty(), form, DummyField(' ')) + +def test_ip_address(): + assert ip_address(form, DummyField('127.0.0.1')) == None + raises(ValidationError, ip_address, form, DummyField('abc.0.0.1')) + raises(ValidationError, ip_address, form, DummyField('1278.0.0.1')) + raises(ValidationError, ip_address, form, DummyField('127.0.0.abc')) +
a205a7032aa72111e1cb44f4989363b4e64be28a
tools/download-wheels.py
tools/download-wheels.py
#!/usr/bin/env python """ Download NumPy wheels from Anaconda staging area. """ import sys import os import re import shutil import argparse import urllib3 from bs4 import BeautifulSoup __version__ = '0.1' ANACONDA_INDEX = 'https://anaconda.org/multibuild-wheels-staging/numpy/files' ANACONDA_FILES = 'https://anaconda.org/multibuild-wheels-staging/numpy/simple' def get_wheel_names(version): """ Get wheel names from Anaconda HTML directory. This looks in the Anaconda multibuild-wheels-staging page and parses the HTML to get all the wheel names for a release version. Parameters ---------- version : str The release version. For instance, "1.18.3". """ tmpl = re.compile('^.*numpy-' + version + '.*\.whl$') http = urllib3.PoolManager(cert_reqs='CERT_REQUIRED') indx = http.request('GET', ANACONDA_INDEX) soup = BeautifulSoup(indx.data, 'html.parser') return soup.findAll(text=tmpl) def download_wheels(version, wheelhouse): """Download release wheels. The release wheels for the given NumPy version are downloaded into the given directory. Parameters ---------- version : str The release version. For instance, "1.18.3". wheelhouse : str Directory in which to download the wheels. """ wheel_names = get_wheel_names(version[0]) http = urllib3.PoolManager(cert_reqs='CERT_REQUIRED') for wheel_name in wheel_names: wheel_url = os.path.join(ANACONDA_FILES, wheel_name) wheel_path = os.path.join(wheelhouse, wheel_name) with open(wheel_path, 'wb') as f: with http.request('GET', wheel_url, preload_content=False,) as r: print(f"Downloading {wheel_name}") shutil.copyfileobj(r, f) if __name__ == '__main__': parser = argparse.ArgumentParser() parser.add_argument( "version", help="NumPy version to download.") parser.add_argument( "-w", "--wheelhouse", default=os.path.join(os.getcwd(), "release", "installers"), help="Directory in which to store downloaded wheels\n" "[defaults to <cwd>/release/installers]") args = parser.parse_args() wheelhouse = os.path.expanduser(args.wheelhouse) download_wheels(args.version, wheelhouse)
Add tool for downloading release wheels from Anaconda.
ENH: Add tool for downloading release wheels from Anaconda. This is a simplified version of terryfy::wheel-uploader that has two advantages over the original: - It works with Anaconda where our wheels are now stored. - It is simplified to match the NumPY workflow.
Python
bsd-3-clause
mattip/numpy,pdebuyl/numpy,endolith/numpy,abalkin/numpy,mhvk/numpy,mhvk/numpy,charris/numpy,seberg/numpy,pbrod/numpy,mattip/numpy,numpy/numpy,pbrod/numpy,jakirkham/numpy,pdebuyl/numpy,seberg/numpy,pbrod/numpy,jakirkham/numpy,pdebuyl/numpy,simongibbons/numpy,charris/numpy,endolith/numpy,grlee77/numpy,grlee77/numpy,mhvk/numpy,jakirkham/numpy,numpy/numpy,anntzer/numpy,grlee77/numpy,rgommers/numpy,endolith/numpy,grlee77/numpy,madphysicist/numpy,charris/numpy,pbrod/numpy,numpy/numpy,rgommers/numpy,madphysicist/numpy,abalkin/numpy,numpy/numpy,pbrod/numpy,simongibbons/numpy,mattip/numpy,simongibbons/numpy,grlee77/numpy,anntzer/numpy,charris/numpy,simongibbons/numpy,madphysicist/numpy,pdebuyl/numpy,jakirkham/numpy,rgommers/numpy,jakirkham/numpy,rgommers/numpy,mattip/numpy,madphysicist/numpy,simongibbons/numpy,seberg/numpy,endolith/numpy,abalkin/numpy,seberg/numpy,anntzer/numpy,mhvk/numpy,madphysicist/numpy,mhvk/numpy,anntzer/numpy
--- +++ @@ -0,0 +1,79 @@ +#!/usr/bin/env python +""" +Download NumPy wheels from Anaconda staging area. + +""" +import sys +import os +import re +import shutil +import argparse + +import urllib3 +from bs4 import BeautifulSoup + +__version__ = '0.1' + +ANACONDA_INDEX = 'https://anaconda.org/multibuild-wheels-staging/numpy/files' +ANACONDA_FILES = 'https://anaconda.org/multibuild-wheels-staging/numpy/simple' + + +def get_wheel_names(version): + """ Get wheel names from Anaconda HTML directory. + + This looks in the Anaconda multibuild-wheels-staging page and + parses the HTML to get all the wheel names for a release version. + + Parameters + ---------- + version : str + The release version. For instance, "1.18.3". + + """ + tmpl = re.compile('^.*numpy-' + version + '.*\.whl$') + http = urllib3.PoolManager(cert_reqs='CERT_REQUIRED') + indx = http.request('GET', ANACONDA_INDEX) + soup = BeautifulSoup(indx.data, 'html.parser') + return soup.findAll(text=tmpl) + + +def download_wheels(version, wheelhouse): + """Download release wheels. + + The release wheels for the given NumPy version are downloaded + into the given directory. + + Parameters + ---------- + version : str + The release version. For instance, "1.18.3". + wheelhouse : str + Directory in which to download the wheels. + + """ + wheel_names = get_wheel_names(version[0]) + http = urllib3.PoolManager(cert_reqs='CERT_REQUIRED') + for wheel_name in wheel_names: + wheel_url = os.path.join(ANACONDA_FILES, wheel_name) + wheel_path = os.path.join(wheelhouse, wheel_name) + with open(wheel_path, 'wb') as f: + with http.request('GET', wheel_url, preload_content=False,) as r: + print(f"Downloading {wheel_name}") + shutil.copyfileobj(r, f) + + +if __name__ == '__main__': + parser = argparse.ArgumentParser() + parser.add_argument( + "version", + help="NumPy version to download.") + parser.add_argument( + "-w", "--wheelhouse", + default=os.path.join(os.getcwd(), "release", "installers"), + help="Directory in which to store downloaded wheels\n" + "[defaults to <cwd>/release/installers]") + + args = parser.parse_args() + + wheelhouse = os.path.expanduser(args.wheelhouse) + download_wheels(args.version, wheelhouse)
a49ace410d239a0e8ccce8b2780b0ff2eb9e6dfb
protogeni/test/consoleurl.py
protogeni/test/consoleurl.py
#! /usr/bin/env python # # Copyright (c) 2008-2014 University of Utah and the Flux Group. # # {{{GENIPUBLIC-LICENSE # # GENI Public License # # Permission is hereby granted, free of charge, to any person obtaining # a copy of this software and/or hardware specification (the "Work") to # deal in the Work without restriction, including without limitation the # rights to use, copy, modify, merge, publish, distribute, sublicense, # and/or sell copies of the Work, and to permit persons to whom the Work # is furnished to do so, subject to the following conditions: # # The above copyright notice and this permission notice shall be # included in all copies or substantial portions of the Work. # # THE WORK IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS # OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF # MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND # NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT # HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, # WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, # OUT OF OR IN CONNECTION WITH THE WORK OR THE USE OR OTHER DEALINGS # IN THE WORK. # # }}} # # # import sys import pwd import getopt import os import re ACCEPTSLICENAME=1 sliver_urn = None execfile( "test-common.py" ) if len(REQARGS) != 1: print >> sys.stderr, "Must provide a sliver urn" sys.exit(1) else: sliver_urn = REQARGS[0] pass # # Get a credential for myself, that allows me to do things at the SA. # mycredential = get_self_credential() print "Got my SA credential. Looking for slice ..." # # Lookup slice. # myslice = resolve_slice( SLICENAME, mycredential ) print "Found the slice, asking for a credential ..." # # Get the slice credential. # slicecred = get_slice_credential( myslice, mycredential ) print "Got the slice credential, asking for the sliver console url ..." # # Get the console url # params = {} params["slice_urn"] = myslice["urn"] params["sliver_urn"] = sliver_urn params["credentials"] = (slicecred,) rval,response = do_method("cm", "ConsoleURL", params, version="2.0") if rval: Fatal("Could not get the console URL") pass print str(response["value"])
Test script to get console url for a sliver.
Test script to get console url for a sliver.
Python
agpl-3.0
nmc-probe/emulab-nome,nmc-probe/emulab-nome,nmc-probe/emulab-nome,nmc-probe/emulab-nome,nmc-probe/emulab-nome,nmc-probe/emulab-nome,nmc-probe/emulab-nome,nmc-probe/emulab-nome,nmc-probe/emulab-nome,nmc-probe/emulab-nome,nmc-probe/emulab-nome
--- +++ @@ -0,0 +1,81 @@ +#! /usr/bin/env python +# +# Copyright (c) 2008-2014 University of Utah and the Flux Group. +# +# {{{GENIPUBLIC-LICENSE +# +# GENI Public License +# +# Permission is hereby granted, free of charge, to any person obtaining +# a copy of this software and/or hardware specification (the "Work") to +# deal in the Work without restriction, including without limitation the +# rights to use, copy, modify, merge, publish, distribute, sublicense, +# and/or sell copies of the Work, and to permit persons to whom the Work +# is furnished to do so, subject to the following conditions: +# +# The above copyright notice and this permission notice shall be +# included in all copies or substantial portions of the Work. +# +# THE WORK IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS +# OR IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF +# MERCHANTABILITY, FITNESS FOR A PARTICULAR PURPOSE AND +# NONINFRINGEMENT. IN NO EVENT SHALL THE AUTHORS OR COPYRIGHT +# HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER LIABILITY, +# WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, +# OUT OF OR IN CONNECTION WITH THE WORK OR THE USE OR OTHER DEALINGS +# IN THE WORK. +# +# }}} +# + +# +# +import sys +import pwd +import getopt +import os +import re + +ACCEPTSLICENAME=1 +sliver_urn = None + +execfile( "test-common.py" ) + +if len(REQARGS) != 1: + print >> sys.stderr, "Must provide a sliver urn" + sys.exit(1) +else: + sliver_urn = REQARGS[0] + pass + +# +# Get a credential for myself, that allows me to do things at the SA. +# +mycredential = get_self_credential() +print "Got my SA credential. Looking for slice ..." + +# +# Lookup slice. +# +myslice = resolve_slice( SLICENAME, mycredential ) +print "Found the slice, asking for a credential ..." + +# +# Get the slice credential. +# +slicecred = get_slice_credential( myslice, mycredential ) +print "Got the slice credential, asking for the sliver console url ..." + +# +# Get the console url +# +params = {} +params["slice_urn"] = myslice["urn"] +params["sliver_urn"] = sliver_urn +params["credentials"] = (slicecred,) +rval,response = do_method("cm", "ConsoleURL", params, version="2.0") +if rval: + Fatal("Could not get the console URL") + pass +print str(response["value"]) +
93feee9e6d4221e6c876413055ee8be659ddeb36
pythran/tests/rosetta/yin_and_yang.py
pythran/tests/rosetta/yin_and_yang.py
# from http://rosettacode.org/wiki/Yin_and_yang#Python import math def yinyang(n=3): radii = [i * n for i in [1, 3, 6]] ranges = [list(range(-r, r+1)) for r in radii] squares = [[ (x,y) for x in rnge for y in rnge] for rnge in ranges] circles = [[ (x,y) for x,y in sqrpoints if math.hypot(x,y) <= radius ] for sqrpoints, radius in zip(squares, radii)] m = {(x,y):' ' for x,y in squares[-1]} for x,y in circles[-1]: m[x,y] = '*' for x,y in circles[-1]: if x>0: m[(x,y)] = '.' for x,y in circles[-2]: m[(x,y+3*n)] = '*' m[(x,y-3*n)] = '.' for x,y in circles[-3]: m[(x,y+3*n)] = '.' m[(x,y-3*n)] = '*' return '\n'.join(''.join(m[(x,y)] for x in reversed(ranges[-1])) for y in ranges[-1]) def test(): return yinyang()
Handle Y' rosetta code series
Handle Y' rosetta code series A very beautiful one, that once again raised a lot of errors...
Python
bsd-3-clause
pombredanne/pythran,pbrunet/pythran,hainm/pythran,artas360/pythran,pbrunet/pythran,serge-sans-paille/pythran,serge-sans-paille/pythran,pombredanne/pythran,pbrunet/pythran,hainm/pythran,pombredanne/pythran,hainm/pythran,artas360/pythran,artas360/pythran
--- +++ @@ -0,0 +1,25 @@ +# from http://rosettacode.org/wiki/Yin_and_yang#Python +import math +def yinyang(n=3): + radii = [i * n for i in [1, 3, 6]] + ranges = [list(range(-r, r+1)) for r in radii] + squares = [[ (x,y) for x in rnge for y in rnge] + for rnge in ranges] + circles = [[ (x,y) for x,y in sqrpoints + if math.hypot(x,y) <= radius ] + for sqrpoints, radius in zip(squares, radii)] + m = {(x,y):' ' for x,y in squares[-1]} + for x,y in circles[-1]: + m[x,y] = '*' + for x,y in circles[-1]: + if x>0: m[(x,y)] = '.' + for x,y in circles[-2]: + m[(x,y+3*n)] = '*' + m[(x,y-3*n)] = '.' + for x,y in circles[-3]: + m[(x,y+3*n)] = '.' + m[(x,y-3*n)] = '*' + return '\n'.join(''.join(m[(x,y)] for x in reversed(ranges[-1])) for y in ranges[-1]) + +def test(): + return yinyang()
39bb9d8408cde59525a576119d28399c49e092f9
mars.py
mars.py
import requests rover_url = 'https://api.nasa.gov/mars-photos/api/v1/rovers/curiosity/photos' parameters = {'api_key': 'DEMO_KEY', 'sol': '1324'} response = requests.get(rover_url, params=parameters).json() print(len(response['photos'])) for photo in response['photos']: print(photo['img_src'])
Print out image URLs from Mars
Print out image URLs from Mars
Python
mit
sagnew/PhoningHome
--- +++ @@ -0,0 +1,10 @@ +import requests + +rover_url = 'https://api.nasa.gov/mars-photos/api/v1/rovers/curiosity/photos' +parameters = {'api_key': 'DEMO_KEY', 'sol': '1324'} + +response = requests.get(rover_url, params=parameters).json() + +print(len(response['photos'])) +for photo in response['photos']: + print(photo['img_src'])
a066aba57d828d25a692ae1dda31797ea7046ddf
test.py
test.py
from ringcentral import SDK from config import USERNAME, EXTENSION, PASSWORD, APP_KEY, APP_SECRET, SERVER # Before you start # Rename credentials-sample.ini to credentials.ini # Edit credentials.ini with information about your app and your creds sdk = SDK(APP_KEY, APP_SECRET, SERVER) platform = sdk.platform() platform.login(USERNAME, EXTENSION, PASSWORD) res = platform.get('/account/~/extension/~') print('User loaded ' + res.json().name)
Test file to get started with APIs
Test file to get started with APIs
Python
mit
ringcentral/ringcentral-python
--- +++ @@ -0,0 +1,13 @@ +from ringcentral import SDK +from config import USERNAME, EXTENSION, PASSWORD, APP_KEY, APP_SECRET, SERVER + +# Before you start +# Rename credentials-sample.ini to credentials.ini +# Edit credentials.ini with information about your app and your creds + +sdk = SDK(APP_KEY, APP_SECRET, SERVER) +platform = sdk.platform() +platform.login(USERNAME, EXTENSION, PASSWORD) + +res = platform.get('/account/~/extension/~') +print('User loaded ' + res.json().name)
f03734d273914ab5ce31cf0e28f86732ddb35b0c
tests/test_casefold.py
tests/test_casefold.py
# -*- coding: utf-8 -*- from irc3.testing import BotTestCase class TestCasefold(BotTestCase): config = dict(includes=['irc3.plugins.casefold']) def test_ascii(self): bot = self.callFTU(server_config={'CASEMAPPING': 'ascii'}) self.assertEquals(bot.casefold('#testchan\\123[]56'), '#testchan\\123[]56') self.assertEquals(bot.casefold('#tESt[]chAn'), '#test[]chan') self.assertEquals(bot.casefold('#TEsT\\CHaN'), '#test\\chan') self.assertEquals(bot.casefold(u'#TEsT\\CHaN'), u'#test\\chan') def test_rfc1459(self): bot = self.callFTU(server_config={'CASEMAPPING': 'rfc1459'}) self.assertEquals(bot.casefold('#testchan\\123[]56'), '#testchan|123{}56') self.assertEquals(bot.casefold('#tESt[]chAn'), '#test{}chan') self.assertEquals(bot.casefold('#TEsT\\CHaN'), '#test|chan') self.assertEquals(bot.casefold(u'#TEsT\\CHaN'), u'#test|chan')
Add testcase for casefold plugin
Add testcase for casefold plugin
Python
mit
mrhanky17/irc3,mrhanky17/irc3,gawel/irc3
--- +++ @@ -0,0 +1,27 @@ +# -*- coding: utf-8 -*- +from irc3.testing import BotTestCase + + +class TestCasefold(BotTestCase): + + config = dict(includes=['irc3.plugins.casefold']) + + def test_ascii(self): + bot = self.callFTU(server_config={'CASEMAPPING': 'ascii'}) + + self.assertEquals(bot.casefold('#testchan\\123[]56'), + '#testchan\\123[]56') + self.assertEquals(bot.casefold('#tESt[]chAn'), '#test[]chan') + self.assertEquals(bot.casefold('#TEsT\\CHaN'), '#test\\chan') + + self.assertEquals(bot.casefold(u'#TEsT\\CHaN'), u'#test\\chan') + + def test_rfc1459(self): + bot = self.callFTU(server_config={'CASEMAPPING': 'rfc1459'}) + + self.assertEquals(bot.casefold('#testchan\\123[]56'), + '#testchan|123{}56') + self.assertEquals(bot.casefold('#tESt[]chAn'), '#test{}chan') + self.assertEquals(bot.casefold('#TEsT\\CHaN'), '#test|chan') + + self.assertEquals(bot.casefold(u'#TEsT\\CHaN'), u'#test|chan')
980cf50a2bf6c42cbd689c1b7e1abcadd45f0f1e
tests/test_frontend.py
tests/test_frontend.py
# -*- coding: utf-8 -*- # # This file is part of pypuppetdbquery. # Copyright © 2016 Chris Boot <bootc@bootc.net> # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import json import unittest from pypuppetdbquery import parse class TestFrontend(unittest.TestCase): """ Test suite for `pypuppetdbquery.parse`. """ def _parse(self, s, **kwargs): return parse(s, parser_opts={ 'lex_options': { 'debug': False, 'optimize': False, }, 'yacc_options': { 'debug': False, 'optimize': False, 'write_tables': False, }, }, **kwargs) def test_empty_queries(self): out = self._parse('') self.assertTrue(out is None) def test_simple_json(self): out = self._parse('foo=bar') expect = json.dumps([ 'in', 'certname', ['extract', 'certname', ['select_fact_contents', ['and', ['=', 'path', ['foo']], ['=', 'value', 'bar']]]]]) self.assertEqual(out, expect) def test_simple_raw(self): out = self._parse('foo=bar', json=False) expect = [ 'in', 'certname', ['extract', 'certname', ['select_fact_contents', ['and', ['=', 'path', ['foo']], ['=', 'value', 'bar']]]]] self.assertEqual(out, expect)
Add unit tests for pypuppetdbquery.parse()
Add unit tests for pypuppetdbquery.parse()
Python
apache-2.0
bootc/pypuppetdbquery
--- +++ @@ -0,0 +1,65 @@ +# -*- coding: utf-8 -*- +# +# This file is part of pypuppetdbquery. +# Copyright © 2016 Chris Boot <bootc@bootc.net> +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +import json +import unittest + +from pypuppetdbquery import parse + + +class TestFrontend(unittest.TestCase): + """ + Test suite for `pypuppetdbquery.parse`. + """ + def _parse(self, s, **kwargs): + return parse(s, parser_opts={ + 'lex_options': { + 'debug': False, + 'optimize': False, + }, + 'yacc_options': { + 'debug': False, + 'optimize': False, + 'write_tables': False, + }, + }, **kwargs) + + def test_empty_queries(self): + out = self._parse('') + self.assertTrue(out is None) + + def test_simple_json(self): + out = self._parse('foo=bar') + expect = json.dumps([ + 'in', 'certname', + ['extract', 'certname', + ['select_fact_contents', + ['and', + ['=', 'path', ['foo']], + ['=', 'value', 'bar']]]]]) + self.assertEqual(out, expect) + + def test_simple_raw(self): + out = self._parse('foo=bar', json=False) + expect = [ + 'in', 'certname', + ['extract', 'certname', + ['select_fact_contents', + ['and', + ['=', 'path', ['foo']], + ['=', 'value', 'bar']]]]] + self.assertEqual(out, expect)
443456f94a92f1844b99965ffcf3679bd939d42c
tests/test_moving_eddies.py
tests/test_moving_eddies.py
from parcels import Particle, ParticleSet, JITParticle, JITParticleSet from parcels import NEMOGrid, ParticleFile, AdvectionRK4 from argparse import ArgumentParser pclasses = {'scipy': (Particle, ParticleSet), 'jit': (JITParticle, JITParticleSet)} def moving_eddies(grid, npart, mode='jit', verbose=False): """Configuration of a particle set that follows two moving eddies :arg grid: :class NEMOGrid: that defines the flow field :arg npart: Number of particles to intialise""" # Determine particle and set classes according to mode ParticleClass, ParticleSetClass = pclasses[mode] lon = [3.3, 3.3] lat = [46., 47.8] pset = ParticleSetClass(2, grid, lon=lon, lat=lat) if verbose: print("Initial particle positions:") for p in pset: print(p) out = ParticleFile(name="EddyParticle", particleset=pset) out.write(pset, 0.) current = 0. for _ in range(24 * 25): pset.execute(AdvectionRK4, time=current, timesteps=12, dt=300.) out.write(pset, current) current += 3600. if verbose: print("Final particle positions:") for p in pset: print(p) if __name__ == "__main__": p = ArgumentParser(description=""" Example of particle advection around an idealised peninsula""") p.add_argument('mode', choices=('scipy', 'jit'), nargs='?', default='jit', help='Execution mode for performing RK4 computation') p.add_argument('-p', '--particles', type=int, default=20, help='Number of particles to advect') p.add_argument('-v', '--verbose', action='store_true', default=False, help='Print particle information before and after execution') p.add_argument('--profiling', action='store_true', default=False, help='Print profiling information after run') args = p.parse_args() # Open grid file set grid = NEMOGrid.from_file('moving_eddies') if args.profiling: from cProfile import runctx from pstats import Stats runctx("moving_eddies(grid, args.particles, mode=args.mode, \ verbose=args.verbose)", globals(), locals(), "Profile.prof") Stats("Profile.prof").strip_dirs().sort_stats("time").print_stats(10) else: moving_eddies(grid, args.particles, mode=args.mode, verbose=args.verbose)
Add simple test setup that follows two particles
MovingEddies: Add simple test setup that follows two particles
Python
mit
OceanPARCELS/parcels,OceanPARCELS/parcels
--- +++ @@ -0,0 +1,66 @@ +from parcels import Particle, ParticleSet, JITParticle, JITParticleSet +from parcels import NEMOGrid, ParticleFile, AdvectionRK4 +from argparse import ArgumentParser + + +pclasses = {'scipy': (Particle, ParticleSet), + 'jit': (JITParticle, JITParticleSet)} + + +def moving_eddies(grid, npart, mode='jit', verbose=False): + """Configuration of a particle set that follows two moving eddies + + :arg grid: :class NEMOGrid: that defines the flow field + :arg npart: Number of particles to intialise""" + + # Determine particle and set classes according to mode + ParticleClass, ParticleSetClass = pclasses[mode] + + lon = [3.3, 3.3] + lat = [46., 47.8] + pset = ParticleSetClass(2, grid, lon=lon, lat=lat) + + if verbose: + print("Initial particle positions:") + for p in pset: + print(p) + + out = ParticleFile(name="EddyParticle", particleset=pset) + out.write(pset, 0.) + current = 0. + for _ in range(24 * 25): + pset.execute(AdvectionRK4, time=current, + timesteps=12, dt=300.) + out.write(pset, current) + current += 3600. + + if verbose: + print("Final particle positions:") + for p in pset: + print(p) + +if __name__ == "__main__": + p = ArgumentParser(description=""" +Example of particle advection around an idealised peninsula""") + p.add_argument('mode', choices=('scipy', 'jit'), nargs='?', default='jit', + help='Execution mode for performing RK4 computation') + p.add_argument('-p', '--particles', type=int, default=20, + help='Number of particles to advect') + p.add_argument('-v', '--verbose', action='store_true', default=False, + help='Print particle information before and after execution') + p.add_argument('--profiling', action='store_true', default=False, + help='Print profiling information after run') + args = p.parse_args() + + # Open grid file set + grid = NEMOGrid.from_file('moving_eddies') + + if args.profiling: + from cProfile import runctx + from pstats import Stats + runctx("moving_eddies(grid, args.particles, mode=args.mode, \ + verbose=args.verbose)", + globals(), locals(), "Profile.prof") + Stats("Profile.prof").strip_dirs().sort_stats("time").print_stats(10) + else: + moving_eddies(grid, args.particles, mode=args.mode, verbose=args.verbose)
5735595d4a465b2cc9f52bdd1fc524c0d26d60cb
tests/test_task_statuses.py
tests/test_task_statuses.py
from taiga.requestmaker import RequestMaker, RequestMakerException from taiga.models.base import InstanceResource, ListResource from taiga.models import TaskStatus, TaskStatuses from taiga import TaigaAPI import taiga.exceptions import json import requests import unittest from mock import patch from .tools import create_mock_json from .tools import MockResponse class TestTaskStatuses(unittest.TestCase): @patch('taiga.models.base.ListResource._new_resource') def test_create_task_status(self, mock_new_resource): rm = RequestMaker('/api/v1', 'fakehost', 'faketoken') mock_new_resource.return_value = TaskStatus(rm) ts = TaskStatuses(rm).create(1, 'TS 1') mock_new_resource.assert_called_with( payload={'project': 1, 'name': 'TS 1'} )
Add tests for task statuses
Add tests for task statuses
Python
mit
bameda/python-taiga,mlq/python-taiga,jespino/python-taiga,bameda/python-taiga,mlq/python-taiga,erikw/python-taiga,jespino/python-taiga,erikw/python-taiga,nephila/python-taiga
--- +++ @@ -0,0 +1,24 @@ +from taiga.requestmaker import RequestMaker, RequestMakerException +from taiga.models.base import InstanceResource, ListResource +from taiga.models import TaskStatus, TaskStatuses +from taiga import TaigaAPI +import taiga.exceptions +import json +import requests +import unittest +from mock import patch +from .tools import create_mock_json +from .tools import MockResponse + + +class TestTaskStatuses(unittest.TestCase): + + @patch('taiga.models.base.ListResource._new_resource') + def test_create_task_status(self, mock_new_resource): + rm = RequestMaker('/api/v1', 'fakehost', 'faketoken') + mock_new_resource.return_value = TaskStatus(rm) + ts = TaskStatuses(rm).create(1, 'TS 1') + mock_new_resource.assert_called_with( + payload={'project': 1, 'name': 'TS 1'} + ) +
96ed9d514541b18419b97d5187237f6fdba7b3c7
tohu/v3/dependency_graph.py
tohu/v3/dependency_graph.py
import logging import re from graphviz import Digraph from IPython.display import SVG __all__ = ['DependencyGraph'] logger = logging.getLogger('tohu') class DependencyGraph: NODE_ATTR = dict(shape='box', style='filled', fillcolor='white') def __init__(self, *, scale=1.0, name=None, graph_attr=None): graph_attr = graph_attr or dict() graph_attr['splines'] = 'ortho' self.graph = Digraph(name=name, node_attr=self.NODE_ATTR, graph_attr=graph_attr) self.scale = scale def __repr__(self): return f'<DependencyGraph>' def _repr_svg_(self): return self.get_svg(scale=self.scale) def add_node(self, g): self.graph.node(g.tohu_id, label=f'{g:long}') def add_edge(self, g1, g2, *, color='/paired3/2', style='solid', constraint=None): self.graph.edge(g1.tohu_id, g2.tohu_id, color=color, style=style, constraint=constraint) def add_subgraph(self, sg): assert isinstance(sg, DependencyGraph) self.graph.subgraph(sg.graph) def get_svg(self, scale=1.0): """ Return string with an SVG representation of the graph. """ svg = self.graph._repr_svg_() width, height = re.search('svg width="(\d+)pt" height="(\d+)pt"', svg).groups() width_new = int(scale * int(width)) height_new = int(scale * int(height)) svg_scaled = re.sub('svg width="\d+pt" height="\d+pt"', f'svg width="{width_new}pt" height="{height_new}pt"', svg) return svg_scaled def draw(self, scale=1.0): """ Convenience method to draw a - possibly scaled - version of the graph in Jupyter notebooks. Returns an SVG object which is rendered automatically by the notebook. """ return SVG(self.get_svg(scale=scale))
Add class to support dependency graph visualisation
Add class to support dependency graph visualisation
Python
mit
maxalbert/tohu
--- +++ @@ -0,0 +1,53 @@ +import logging +import re +from graphviz import Digraph +from IPython.display import SVG + +__all__ = ['DependencyGraph'] + +logger = logging.getLogger('tohu') + + +class DependencyGraph: + + NODE_ATTR = dict(shape='box', style='filled', fillcolor='white') + + def __init__(self, *, scale=1.0, name=None, graph_attr=None): + graph_attr = graph_attr or dict() + graph_attr['splines'] = 'ortho' + self.graph = Digraph(name=name, node_attr=self.NODE_ATTR, graph_attr=graph_attr) + self.scale = scale + + def __repr__(self): + return f'<DependencyGraph>' + + def _repr_svg_(self): + return self.get_svg(scale=self.scale) + + def add_node(self, g): + self.graph.node(g.tohu_id, label=f'{g:long}') + + def add_edge(self, g1, g2, *, color='/paired3/2', style='solid', constraint=None): + self.graph.edge(g1.tohu_id, g2.tohu_id, color=color, style=style, constraint=constraint) + + def add_subgraph(self, sg): + assert isinstance(sg, DependencyGraph) + self.graph.subgraph(sg.graph) + + def get_svg(self, scale=1.0): + """ + Return string with an SVG representation of the graph. + """ + svg = self.graph._repr_svg_() + width, height = re.search('svg width="(\d+)pt" height="(\d+)pt"', svg).groups() + width_new = int(scale * int(width)) + height_new = int(scale * int(height)) + svg_scaled = re.sub('svg width="\d+pt" height="\d+pt"', f'svg width="{width_new}pt" height="{height_new}pt"', svg) + return svg_scaled + + def draw(self, scale=1.0): + """ + Convenience method to draw a - possibly scaled - version of the graph in Jupyter notebooks. + Returns an SVG object which is rendered automatically by the notebook. + """ + return SVG(self.get_svg(scale=scale))
b54f1755d6f792a97eaa06a83cadbb72d045402f
testtube/decorators.py
testtube/decorators.py
class RequireModule(object): """Decorator that raises import error if specified module isn't found.""" def __init__(self, module_name): self.module_name = module_name def _require_module(self): try: __import__(self.module_name) except ImportError: raise ImportError( '%s must be installed to use this helper.' % self.module_name) def __call__(self, func, *args, **kwargs): def wrapper(*args, **kwargs): self._require_module() return func(*args, **kwargs) return wrapper
Add a RequireModule decorator for use by helpers
Add a RequireModule decorator for use by helpers This eliminates the need for _require in helpers.
Python
mit
blaix/testtube,thomasw/testtube,beck/testtube
--- +++ @@ -0,0 +1,19 @@ +class RequireModule(object): + """Decorator that raises import error if specified module isn't found.""" + + def __init__(self, module_name): + self.module_name = module_name + + def _require_module(self): + try: + __import__(self.module_name) + except ImportError: + raise ImportError( + '%s must be installed to use this helper.' % self.module_name) + + def __call__(self, func, *args, **kwargs): + def wrapper(*args, **kwargs): + self._require_module() + return func(*args, **kwargs) + + return wrapper
55ee1a6609a5dfd922cce78fd4faeb64cd2d3c7d
run.py
run.py
"""Hacked-together development server for feedreader. Runs the feedreader server under the /api prefix, serves URI not containing a dot public/index.html, servers everything else to public. """ import tornado.ioloop import tornado.web import feedreader.main class PrefixedFallbackHandler(tornado.web.FallbackHandler): """FallbackHandler that removes the given prefix from requests.""" def prepare(self): # hacky way of removing /api/ self.request.uri = self.request.uri[4:] self.request.path = self.request.path[4:] super(PrefixedFallbackHandler, self).prepare() class SingleFileHandler(tornado.web.StaticFileHandler): """FileHandler that only reads a single static file.""" @classmethod def get_absolute_path(cls, root, path): return tornado.web.StaticFileHandler.get_absolute_path(root, "index.html") def main(): feedreader_app = feedreader.main.get_application() application = tornado.web.Application([ (r"/api/(.*)", PrefixedFallbackHandler, dict(fallback=feedreader_app)), (r"/(.*\..*)", tornado.web.StaticFileHandler, {"path": "public"}), (r"/(.*)", SingleFileHandler, {"path": "public"}), ]) application.listen(8080) tornado.ioloop.IOLoop.instance().start() if __name__ == "__main__": main()
Add new dev server to replace tape
Add new dev server to replace tape
Python
mit
tdryer/feeder,tdryer/feeder
--- +++ @@ -0,0 +1,44 @@ +"""Hacked-together development server for feedreader. + +Runs the feedreader server under the /api prefix, serves URI not containing a +dot public/index.html, servers everything else to public. +""" + +import tornado.ioloop +import tornado.web + +import feedreader.main + + +class PrefixedFallbackHandler(tornado.web.FallbackHandler): + """FallbackHandler that removes the given prefix from requests.""" + + def prepare(self): + # hacky way of removing /api/ + self.request.uri = self.request.uri[4:] + self.request.path = self.request.path[4:] + super(PrefixedFallbackHandler, self).prepare() + + +class SingleFileHandler(tornado.web.StaticFileHandler): + """FileHandler that only reads a single static file.""" + + @classmethod + def get_absolute_path(cls, root, path): + return tornado.web.StaticFileHandler.get_absolute_path(root, + "index.html") + + +def main(): + feedreader_app = feedreader.main.get_application() + application = tornado.web.Application([ + (r"/api/(.*)", PrefixedFallbackHandler, dict(fallback=feedreader_app)), + (r"/(.*\..*)", tornado.web.StaticFileHandler, {"path": "public"}), + (r"/(.*)", SingleFileHandler, {"path": "public"}), + ]) + application.listen(8080) + tornado.ioloop.IOLoop.instance().start() + + +if __name__ == "__main__": + main()
3764459fb9a8995a1476cc7a36c88a98dc9d4685
scipy-example/fitsincos.py
scipy-example/fitsincos.py
""" Given the set of points generated by f(x) = 0.5 * cos(2 * x) + 2 * sin(0.5 * x) with some noise, use Levenberg-Marquardt algorithm to find the model of the form f(x) = a * cos(b * x) + b * sin(a * x) to fit all the points. """ import numpy as np import scipy.optimize as scipy_opt def sincos_func(x_data, a, b): """ Computes the function a * sin(b * x) + b * cos(a * x) Args: x_data : A Numpy array of input data a : Real-valued argument of the function b : Real-valued argument of the function Returns: A Numpy array of values of the function a * sin(b * x) + b * cos(a * x) evaluated at each x in xData """ return a * np.cos(b * x_data) + b * np.sin(a * x_data) def main(): """ Main function to set up data points and calls Scipy curve fitting routine (whose underlying algorithm is Levenberg-Marquardt) """ x_data = np.array([ 1.0, 1.5, -1.0, 2.0, 1.8, 2.5, -0.5, -0.8, -1.1, 2.2, 2.6, 2.8, -2.0, -2.2, -1.7, -1.4, 0.05, 0.0, 1.570796, -1.570796, 0.6, -0.6, 1.67, 2.4, 0.1 ]) y_data = np.array([ 0.76, 0.860000, -1.18, 1.356, 1.118, 2.039, -0.224, -0.7934, -1.339, 1.63, 2.1613, 2.35, -2.009, -1.936, -1.985, -1.759, 0.55, 0.5, 0.914, -1.9142, 0.77, -0.4, 1.0, 1.9, 0.59 ]) guess_abs = [[0.25, 1.5], [1.7, 3], [10, 5], [0.0, 3.0]] for guess_ab in guess_abs: ab, covariance = scipy_opt.curve_fit( sincos_func, x_data, y_data, guess_ab) print 'Intial guess: %s' % str(guess_ab) print 'LM results: %s' % str(ab) if __name__ == "__main__": main()
Add scipy sin cos example
Add scipy sin cos example
Python
mit
truongduy134/levenberg-marquardt,truongduy134/levenberg-marquardt,truongduy134/levenberg-marquardt
--- +++ @@ -0,0 +1,46 @@ +""" Given the set of points generated by +f(x) = 0.5 * cos(2 * x) + 2 * sin(0.5 * x) with some noise, +use Levenberg-Marquardt algorithm to find the model of the form +f(x) = a * cos(b * x) + b * sin(a * x) to fit all the points. +""" + +import numpy as np +import scipy.optimize as scipy_opt + +def sincos_func(x_data, a, b): + """ Computes the function a * sin(b * x) + b * cos(a * x) + Args: + x_data : A Numpy array of input data + a : Real-valued argument of the function + b : Real-valued argument of the function + + Returns: + A Numpy array of values of the function a * sin(b * x) + b * cos(a * x) + evaluated at each x in xData + """ + return a * np.cos(b * x_data) + b * np.sin(a * x_data) + +def main(): + """ Main function to set up data points and calls Scipy curve fitting + routine (whose underlying algorithm is Levenberg-Marquardt) + """ + x_data = np.array([ + 1.0, 1.5, -1.0, 2.0, 1.8, 2.5, -0.5, -0.8, -1.1, 2.2, 2.6, 2.8, -2.0, + -2.2, -1.7, -1.4, 0.05, 0.0, 1.570796, -1.570796, 0.6, -0.6, + 1.67, 2.4, 0.1 + ]) + y_data = np.array([ + 0.76, 0.860000, -1.18, 1.356, 1.118, 2.039, -0.224, -0.7934, -1.339, + 1.63, 2.1613, 2.35, -2.009, -1.936, -1.985, -1.759, 0.55, 0.5, 0.914, + -1.9142, 0.77, -0.4, 1.0, 1.9, 0.59 + ]) + guess_abs = [[0.25, 1.5], [1.7, 3], [10, 5], [0.0, 3.0]] + + for guess_ab in guess_abs: + ab, covariance = scipy_opt.curve_fit( + sincos_func, x_data, y_data, guess_ab) + print 'Intial guess: %s' % str(guess_ab) + print 'LM results: %s' % str(ab) + +if __name__ == "__main__": + main()
4091fa177dcaeaac0a36fefee8f5b3cac4c6202b
oscar/apps/dashboard/catalogue/fields.py
oscar/apps/dashboard/catalogue/fields.py
from django import forms from django.template import Context, Template class ProductImageMultipleChoiceField(forms.ModelMultipleChoiceField): """ Field that renders a ProductImage as a thumbnail and text instead of just as text. """ # Using the low-level Template API instead of storing it in a separate file # A user might want to override this, so perhaps it should be a 'partial' template _template = Template(""" {% load thumbnail %} {% thumbnail image.original "50x50" crop="center" as thumb %} <img src="{{ thumb.url }}" alt="{{ image }}" /> {{ image.original }} {% endthumbnail %} """) def __init__(self, *args, **kwargs): if 'widget' not in kwargs: kwargs['widget'] = forms.CheckboxSelectMultiple() if 'required' not in kwargs: kwargs['required'] = False super(ProductImageMultipleChoiceField, self).__init__(*args, **kwargs) def label_from_instance(self, obj): return self._template.render(Context({ 'image': obj }))
Add field to render image many-to-many as checkboxes with thumbnails.
Add field to render image many-to-many as checkboxes with thumbnails.
Python
bsd-3-clause
sonofatailor/django-oscar,sonofatailor/django-oscar,sonofatailor/django-oscar,sonofatailor/django-oscar
--- +++ @@ -0,0 +1,33 @@ +from django import forms +from django.template import Context, Template + +class ProductImageMultipleChoiceField(forms.ModelMultipleChoiceField): + """ + Field that renders a ProductImage as a thumbnail and text instead of + just as text. + """ + + # Using the low-level Template API instead of storing it in a separate file + # A user might want to override this, so perhaps it should be a 'partial' template + _template = Template(""" + {% load thumbnail %} + + {% thumbnail image.original "50x50" crop="center" as thumb %} + <img src="{{ thumb.url }}" alt="{{ image }}" /> {{ image.original }} + {% endthumbnail %} + """) + + + def __init__(self, *args, **kwargs): + if 'widget' not in kwargs: + kwargs['widget'] = forms.CheckboxSelectMultiple() + + if 'required' not in kwargs: + kwargs['required'] = False + + super(ProductImageMultipleChoiceField, self).__init__(*args, **kwargs) + + def label_from_instance(self, obj): + return self._template.render(Context({ + 'image': obj + }))
767f241a930ac0dea5fd3cb7fcb3dc6837d5648b
tests/test_cmatrices.py
tests/test_cmatrices.py
# to run this test, from directory above: # setenv PYTHONPATH /path/to/pyradiomics/radiomics # nosetests --nocapture -v tests/test_features.py import logging from nose_parameterized import parameterized import numpy import six from radiomics import cMatsEnabled, getFeatureClasses from testUtils import custom_name_func, RadiomicsTestUtils testUtils = RadiomicsTestUtils() defaultTestCases = testUtils.getTestCases() testCases = defaultTestCases featureClasses = getFeatureClasses() class TestFeatures: def generate_scenarios(): global testCases, featureClasses for testCase in testCases: for className, featureClass in six.iteritems(featureClasses): assert(featureClass is not None) if "_calculateCMatrix" in dir(featureClass) or className == "shape": logging.debug('generate_scenarios: featureClass = %s', className) yield testCase, className global testUtils @parameterized.expand(generate_scenarios(), testcase_func_name=custom_name_func) def test_scenario(self, testCase, featureClassName): print("") global testUtils, featureClasses logging.debug('test_scenario: testCase = %s, featureClassName = %s', testCase, featureClassName) assert cMatsEnabled() testUtils.setFeatureClassAndTestCase(featureClassName, testCase) testImage = testUtils.getImage() testMask = testUtils.getMask() featureClass = featureClasses[featureClassName](testImage, testMask, **testUtils.getKwargs()) if featureClassName == 'shape': cSA = getattr(featureClass, 'SurfaceArea') # pre-calculated value by C extension assert (cSA is not None) pySA = getattr(featureClass, '_calculateSurfaceArea')() # Function, call to calculate SA in full-python mode assert (pySA is not None) # Check if the calculated values match assert (numpy.abs(pySA - cSA)) < 1e-3 else: assert "_calculateMatrix" in dir(featureClass) cMat = getattr(featureClass, 'P_%s' % featureClassName) # matrix calculated at initialization by C extension assert cMat is not None pyMat = featureClass._calculateMatrix() assert pyMat is not None # Check if the calculated arrays match assert numpy.max(numpy.abs(pyMat - cMat)) < 1e-3
Add testing for C extensions
ENH: Add testing for C extensions Testing compares matrices calculated by python algorithms to those calculated by the C extension. Tests the calculation of surface area in a similar manner. Testing fails if C extension is not available or if any element in the matrix differs more than 1e-3 from the corresponding element in the other matrix.
Python
bsd-3-clause
Radiomics/pyradiomics,Radiomics/pyradiomics,Radiomics/pyradiomics,Radiomics/pyradiomics
--- +++ @@ -0,0 +1,72 @@ +# to run this test, from directory above: +# setenv PYTHONPATH /path/to/pyradiomics/radiomics +# nosetests --nocapture -v tests/test_features.py + +import logging + +from nose_parameterized import parameterized +import numpy +import six + +from radiomics import cMatsEnabled, getFeatureClasses +from testUtils import custom_name_func, RadiomicsTestUtils + + +testUtils = RadiomicsTestUtils() +defaultTestCases = testUtils.getTestCases() + +testCases = defaultTestCases + +featureClasses = getFeatureClasses() + +class TestFeatures: + + def generate_scenarios(): + global testCases, featureClasses + + for testCase in testCases: + for className, featureClass in six.iteritems(featureClasses): + assert(featureClass is not None) + if "_calculateCMatrix" in dir(featureClass) or className == "shape": + logging.debug('generate_scenarios: featureClass = %s', className) + yield testCase, className + + global testUtils + + @parameterized.expand(generate_scenarios(), testcase_func_name=custom_name_func) + def test_scenario(self, testCase, featureClassName): + print("") + global testUtils, featureClasses + + logging.debug('test_scenario: testCase = %s, featureClassName = %s', testCase, featureClassName) + + assert cMatsEnabled() + + testUtils.setFeatureClassAndTestCase(featureClassName, testCase) + + testImage = testUtils.getImage() + testMask = testUtils.getMask() + + featureClass = featureClasses[featureClassName](testImage, testMask, **testUtils.getKwargs()) + + if featureClassName == 'shape': + cSA = getattr(featureClass, 'SurfaceArea') # pre-calculated value by C extension + assert (cSA is not None) + + pySA = getattr(featureClass, '_calculateSurfaceArea')() # Function, call to calculate SA in full-python mode + assert (pySA is not None) + + # Check if the calculated values match + assert (numpy.abs(pySA - cSA)) < 1e-3 + + else: + assert "_calculateMatrix" in dir(featureClass) + + cMat = getattr(featureClass, 'P_%s' % featureClassName) # matrix calculated at initialization by C extension + assert cMat is not None + + pyMat = featureClass._calculateMatrix() + assert pyMat is not None + + # Check if the calculated arrays match + assert numpy.max(numpy.abs(pyMat - cMat)) < 1e-3
c49f193c1ad516bfec52eab64c7dc2508d35f53d
app/grandchallenge/retina_core/migrations/0002_auto_20190225_1228.py
app/grandchallenge/retina_core/migrations/0002_auto_20190225_1228.py
# Generated by Django 2.1.7 on 2019-02-25 12:28 from django.db import migrations from django.contrib.auth.models import User, Group from django.conf import settings from guardian.shortcuts import assign_perm, remove_perm # Permission types PERMISSION_TYPES = ("view", "add", "change", "delete") # Existing annotation (name, codename) as of annotations.0001_initial ANNOTATION_CODENAMES = ( ("BooleanClassificationAnnotation", "booleanclassificationannotation"), ("CoordinateListAnnotation", "coordinatelistannotation"), ("IntegerClassificationAnnotation", "integerclassificationannotation"), ("LandmarkAnnotationSet", "landmarkannotationset"), ("MeasurementAnnotation", "measurementannotation"), ("PolygonAnnotationSet", "polygonannotationset"), ("SingleLandmarkAnnotation", "singlelandmarkannotation"), ("SinglePolygonAnnotation", "singlepolygonannotation"), ("ETDRSGridAnnotation", "etdrsgridannotation"), ) def change_retina_permissions(apps, change_permission_func): retina_admins = Group.objects.get(name=settings.RETINA_ADMINS_GROUP_NAME) for (annotation_name, annotation_codename) in ANNOTATION_CODENAMES: # Change user level object permissions to owners of annotations Annotation = apps.get_model('annotations', annotation_name) for annotation in Annotation.objects.all(): if annotation_name.startswith('single'): owner = annotation.annotation_set.grader else: owner = annotation.grader if owner.groups.filter(name=settings.RETINA_GRADERS_GROUP_NAME).exists(): for permission_type in PERMISSION_TYPES: change_permission_func( f"annotations.{permission_type}_{annotation_codename}", owner, annotation ) # Change group level permissions for permission_type in PERMISSION_TYPES: change_permission_func( f"annotations.{permission_type}_{annotation_codename}", retina_admins, ) def set_retina_permissions_forward(apps, schema_editor): change_retina_permissions(apps, assign_perm) def set_retina_permissions_backward(apps, schema_editor): change_retina_permissions(apps, remove_perm) class Migration(migrations.Migration): dependencies = [ ("retina_core", "0001_initial"), migrations.swappable_dependency(settings.AUTH_USER_MODEL), ("annotations", "0001_initial"), ] operations = [ migrations.RunPython( set_retina_permissions_forward, set_retina_permissions_backward ) ]
Add migration for setting retina annotation permissions
Add migration for setting retina annotation permissions
Python
apache-2.0
comic/comic-django,comic/comic-django,comic/comic-django,comic/comic-django,comic/comic-django
--- +++ @@ -0,0 +1,72 @@ +# Generated by Django 2.1.7 on 2019-02-25 12:28 + +from django.db import migrations +from django.contrib.auth.models import User, Group +from django.conf import settings +from guardian.shortcuts import assign_perm, remove_perm + +# Permission types +PERMISSION_TYPES = ("view", "add", "change", "delete") + +# Existing annotation (name, codename) as of annotations.0001_initial +ANNOTATION_CODENAMES = ( + ("BooleanClassificationAnnotation", "booleanclassificationannotation"), + ("CoordinateListAnnotation", "coordinatelistannotation"), + ("IntegerClassificationAnnotation", "integerclassificationannotation"), + ("LandmarkAnnotationSet", "landmarkannotationset"), + ("MeasurementAnnotation", "measurementannotation"), + ("PolygonAnnotationSet", "polygonannotationset"), + ("SingleLandmarkAnnotation", "singlelandmarkannotation"), + ("SinglePolygonAnnotation", "singlepolygonannotation"), + ("ETDRSGridAnnotation", "etdrsgridannotation"), +) + + +def change_retina_permissions(apps, change_permission_func): + retina_admins = Group.objects.get(name=settings.RETINA_ADMINS_GROUP_NAME) + for (annotation_name, annotation_codename) in ANNOTATION_CODENAMES: + # Change user level object permissions to owners of annotations + Annotation = apps.get_model('annotations', annotation_name) + for annotation in Annotation.objects.all(): + if annotation_name.startswith('single'): + owner = annotation.annotation_set.grader + else: + owner = annotation.grader + + if owner.groups.filter(name=settings.RETINA_GRADERS_GROUP_NAME).exists(): + for permission_type in PERMISSION_TYPES: + change_permission_func( + f"annotations.{permission_type}_{annotation_codename}", + owner, + annotation + ) + + # Change group level permissions + for permission_type in PERMISSION_TYPES: + change_permission_func( + f"annotations.{permission_type}_{annotation_codename}", + retina_admins, + ) + + +def set_retina_permissions_forward(apps, schema_editor): + change_retina_permissions(apps, assign_perm) + + +def set_retina_permissions_backward(apps, schema_editor): + change_retina_permissions(apps, remove_perm) + + +class Migration(migrations.Migration): + + dependencies = [ + ("retina_core", "0001_initial"), + migrations.swappable_dependency(settings.AUTH_USER_MODEL), + ("annotations", "0001_initial"), + ] + + operations = [ + migrations.RunPython( + set_retina_permissions_forward, set_retina_permissions_backward + ) + ]
2a25e583933325c032c53e885b6dcac04e8061bb
tests/queue/test_file.py
tests/queue/test_file.py
''' Unit tests for `tilequeue.queue.file`. ''' from tilequeue.queue import OutputFileQueue from tilequeue import tile from ModestMaps.Core import Coordinate import unittest import StringIO class TestQueue(unittest.TestCase): def setUp(self): self.test_tile_coords = [ (0, 0, 0), (1, 2, 3), (4, 5, 6), (9, 3, 1), (4, 7, 1) ] self.test_tile_objs = [ Coordinate(*coords) for coords in self.test_tile_coords] self.tile_coords_str = '\n'.join( map(tile.serialize_coord, self.test_tile_objs)) + '\n' self.tiles_fp = StringIO.StringIO() self.queue = OutputFileQueue(self.tiles_fp) def test_read(self): self._write_tiles_to_file() # Test `.read() for multiple records.` num_to_read = 3 actual_coords = [ msg.coord for msg in self.queue.read(max_to_read=num_to_read)] expected = self.test_tile_objs[:num_to_read] self.assertEqual( actual_coords, expected, 'Reading multiple records failed') # Test `.read()` for just 1 record at a time. for expected in self.test_tile_objs[num_to_read:]: [actual_msg] = self.queue.read() self.assertEqual( actual_msg.coord, expected, 'Reading 1 record failed') def test_enqueue_and_enqueue_batch(self): # Test `.enqueue_batch()`. num_to_enqueue = 3 self.assertEqual( self.queue.enqueue_batch(self.test_tile_objs[:num_to_enqueue]), (num_to_enqueue, 0), 'Return value of `enqueue_batch()` does not match expected' ) # Test `.enqueue()`. for coords in self.test_tile_objs[num_to_enqueue:]: self.queue.enqueue(coords) self.assertEqual( self.tiles_fp.getvalue(), self.tile_coords_str, 'Contents of file do not match expected') def test_clear(self): self._write_tiles_to_file() self.assertEqual( self.queue.clear(), -1, 'Return value of `clear()` does not match expected.') self.assertEqual( self.tiles_fp.getvalue(), '', '`clear()` did not clear the file!') def test_close(self): self.assertFalse( self.tiles_fp.closed, 'Sanity check failed: the test runner\'s file pointer appears to ' 'be closed. This shouldn\'t ever happen.') self.queue.close() self.assertTrue(self.tiles_fp.closed, 'File pointer was not closed!') def _write_tiles_to_file(self): self.tiles_fp.write(self.tile_coords_str) self.tiles_fp.seek(0)
Add unit tests for queue.file
Add unit tests for queue.file tests/queue/test_file.py -Add a bunch of `unittest` unit tests for most/all of the functions in `tilequeue.queue.file`, as per all of the recent work there.
Python
mit
tilezen/tilequeue,mapzen/tilequeue
--- +++ @@ -0,0 +1,82 @@ +''' +Unit tests for `tilequeue.queue.file`. +''' + +from tilequeue.queue import OutputFileQueue +from tilequeue import tile +from ModestMaps.Core import Coordinate +import unittest +import StringIO + +class TestQueue(unittest.TestCase): + + def setUp(self): + self.test_tile_coords = [ + (0, 0, 0), + (1, 2, 3), + (4, 5, 6), + (9, 3, 1), + (4, 7, 1) + ] + self.test_tile_objs = [ + Coordinate(*coords) for coords in self.test_tile_coords] + self.tile_coords_str = '\n'.join( + map(tile.serialize_coord, self.test_tile_objs)) + '\n' + self.tiles_fp = StringIO.StringIO() + self.queue = OutputFileQueue(self.tiles_fp) + + def test_read(self): + self._write_tiles_to_file() + + # Test `.read() for multiple records.` + num_to_read = 3 + actual_coords = [ + msg.coord for msg in self.queue.read(max_to_read=num_to_read)] + expected = self.test_tile_objs[:num_to_read] + self.assertEqual( + actual_coords, expected, 'Reading multiple records failed') + + # Test `.read()` for just 1 record at a time. + for expected in self.test_tile_objs[num_to_read:]: + [actual_msg] = self.queue.read() + self.assertEqual( + actual_msg.coord, expected, + 'Reading 1 record failed') + + def test_enqueue_and_enqueue_batch(self): + # Test `.enqueue_batch()`. + num_to_enqueue = 3 + self.assertEqual( + self.queue.enqueue_batch(self.test_tile_objs[:num_to_enqueue]), + (num_to_enqueue, 0), + 'Return value of `enqueue_batch()` does not match expected' + ) + + # Test `.enqueue()`. + for coords in self.test_tile_objs[num_to_enqueue:]: + self.queue.enqueue(coords) + + self.assertEqual( + self.tiles_fp.getvalue(), + self.tile_coords_str, + 'Contents of file do not match expected') + + def test_clear(self): + self._write_tiles_to_file() + self.assertEqual( + self.queue.clear(), -1, + 'Return value of `clear()` does not match expected.') + self.assertEqual( + self.tiles_fp.getvalue(), '', '`clear()` did not clear the file!') + + def test_close(self): + self.assertFalse( + self.tiles_fp.closed, + 'Sanity check failed: the test runner\'s file pointer appears to ' + 'be closed. This shouldn\'t ever happen.') + self.queue.close() + self.assertTrue(self.tiles_fp.closed, 'File pointer was not closed!') + + def _write_tiles_to_file(self): + self.tiles_fp.write(self.tile_coords_str) + self.tiles_fp.seek(0)
f89ba4c5a9ada2ce84c9d46d6100fc5aec13f758
examples/stories/movie_lister/apps_db_csv.py
examples/stories/movie_lister/apps_db_csv.py
"""A naive example of dependency injection in Python. Example implementation of dependency injection in Python from Martin Fowler's article about dependency injection and inversion of control: http://www.martinfowler.com/articles/injection.html This mini application uses ``movies`` library, that is configured to work with csv file movies database. """ import sqlite3 from dependency_injector import catalogs from dependency_injector import providers from dependency_injector import injections from movies import MoviesModule from movies import finders from settings import MOVIES_CSV_PATH from settings import MOVIES_DB_PATH class ApplicationModule(catalogs.DeclarativeCatalog): """Catalog of application component providers.""" database = providers.Singleton(sqlite3.connect, MOVIES_DB_PATH) @catalogs.copy(MoviesModule) class DbMoviesModule(MoviesModule): """Customized catalog of movies module component providers.""" movie_finder = providers.Factory(finders.SqliteMovieFinder, *MoviesModule.movie_finder.injections, database=ApplicationModule.database) @catalogs.copy(MoviesModule) class CsvMoviesModule(MoviesModule): """Customized catalog of movies module component providers.""" movie_finder = providers.Factory(finders.CsvMovieFinder, *MoviesModule.movie_finder.injections, csv_file=MOVIES_CSV_PATH, delimeter=',') @injections.inject(db_movie_lister=DbMoviesModule.movie_lister) @injections.inject(csv_movie_lister=CsvMoviesModule.movie_lister) def main(db_movie_lister, csv_movie_lister): """Main function. This program prints info about all movies that were directed by different persons and then prints all movies that were released in 2015. :param db_movie_lister: Database movie lister instance :type db_movie_lister: movies.listers.MovieLister :param csv_movie_lister: Database movie lister instance :type csv_movie_lister: movies.listers.MovieLister """ print db_movie_lister.movies_directed_by('Francis Lawrence') print db_movie_lister.movies_directed_by('Patricia Riggen') print db_movie_lister.movies_directed_by('JJ Abrams') print db_movie_lister.movies_released_in(2015) print csv_movie_lister.movies_directed_by('Francis Lawrence') print csv_movie_lister.movies_directed_by('Patricia Riggen') print csv_movie_lister.movies_directed_by('JJ Abrams') print csv_movie_lister.movies_released_in(2015) if __name__ == '__main__': main()
Add example of catalog copying into MovieLister example
Add example of catalog copying into MovieLister example
Python
bsd-3-clause
rmk135/dependency_injector,rmk135/objects,ets-labs/python-dependency-injector,ets-labs/dependency_injector
--- +++ @@ -0,0 +1,76 @@ +"""A naive example of dependency injection in Python. + +Example implementation of dependency injection in Python from Martin Fowler's +article about dependency injection and inversion of control: + +http://www.martinfowler.com/articles/injection.html + +This mini application uses ``movies`` library, that is configured to work with +csv file movies database. +""" + +import sqlite3 + +from dependency_injector import catalogs +from dependency_injector import providers +from dependency_injector import injections + +from movies import MoviesModule +from movies import finders + +from settings import MOVIES_CSV_PATH +from settings import MOVIES_DB_PATH + + +class ApplicationModule(catalogs.DeclarativeCatalog): + """Catalog of application component providers.""" + + database = providers.Singleton(sqlite3.connect, MOVIES_DB_PATH) + + +@catalogs.copy(MoviesModule) +class DbMoviesModule(MoviesModule): + """Customized catalog of movies module component providers.""" + + movie_finder = providers.Factory(finders.SqliteMovieFinder, + *MoviesModule.movie_finder.injections, + database=ApplicationModule.database) + + +@catalogs.copy(MoviesModule) +class CsvMoviesModule(MoviesModule): + """Customized catalog of movies module component providers.""" + + movie_finder = providers.Factory(finders.CsvMovieFinder, + *MoviesModule.movie_finder.injections, + csv_file=MOVIES_CSV_PATH, + delimeter=',') + + +@injections.inject(db_movie_lister=DbMoviesModule.movie_lister) +@injections.inject(csv_movie_lister=CsvMoviesModule.movie_lister) +def main(db_movie_lister, csv_movie_lister): + """Main function. + + This program prints info about all movies that were directed by different + persons and then prints all movies that were released in 2015. + + :param db_movie_lister: Database movie lister instance + :type db_movie_lister: movies.listers.MovieLister + + :param csv_movie_lister: Database movie lister instance + :type csv_movie_lister: movies.listers.MovieLister + """ + print db_movie_lister.movies_directed_by('Francis Lawrence') + print db_movie_lister.movies_directed_by('Patricia Riggen') + print db_movie_lister.movies_directed_by('JJ Abrams') + print db_movie_lister.movies_released_in(2015) + + print csv_movie_lister.movies_directed_by('Francis Lawrence') + print csv_movie_lister.movies_directed_by('Patricia Riggen') + print csv_movie_lister.movies_directed_by('JJ Abrams') + print csv_movie_lister.movies_released_in(2015) + + +if __name__ == '__main__': + main()
7fc5855aa3aa7e554aeb6faf9ae05567c7a4910c
monitor_website_status.py
monitor_website_status.py
#!/usr/bin/python import urllib2 from BeautifulSoup import BeautifulSoup from StringIO import StringIO from datetime import datetime, timedelta import subprocess WEBSITE = 'http://www.newsdiffs.org/browse/' MAX_TIME = timedelta(hours=1) EMAILS = 'ecprice@mit.edu jenny8lee@gmail.com price@mit.edu'.split() def send_alert_email(subject, body): email = 'Subject: %s\n\n%s' % (subject, body) p = subprocess.Popen(['/usr/bin/msmtp', '-t'] + EMAILS, stdin=subprocess.PIPE) p.communicate(email) if p.wait(): print 'Bad return code:', p.returncode def get_update_time(): html = urllib2.urlopen(WEBSITE) soup = BeautifulSoup(html) datestr = soup.findAll('td')[1].findChild('a').getText() date = datetime.strptime(datestr.replace('.', ''), '%B %d, %Y, %I:%M %p') return date if __name__ == '__main__': try: update_time = get_update_time() time_since_update = datetime.now() - update_time print 'Update time:', time_since_update if time_since_update > MAX_TIME: send_alert_email('Trouble with newsdiffs.org', 'No updates since %s\n%s is too long' % (update_time, time_since_update)) except Exception, e: import traceback traceback.print_exc() send_alert_email('Trouble with newsdiffs.org', 'Cannot check website\n%s' % traceback.format_exc())
Add monitor script for website.
Add monitor script for website.
Python
mit
bjowi/newsdiffs,flupzor/newsdiffs,amandabee/newsdiffs,flupzor/bijgeschaafd,bjowi/newsdiffs,flupzor/bijgeschaafd,catcosmo/newsdiffs,flupzor/newsdiffs,catcosmo/newsdiffs,catcosmo/newsdiffs,flupzor/bijgeschaafd,COLABORATI/newsdiffs,COLABORATI/newsdiffs,flupzor/newsdiffs,bjowi/newsdiffs,flupzor/bijgeschaafd,flupzor/newsdiffs,amandabee/newsdiffs,COLABORATI/newsdiffs,amandabee/newsdiffs
--- +++ @@ -0,0 +1,44 @@ +#!/usr/bin/python + +import urllib2 +from BeautifulSoup import BeautifulSoup +from StringIO import StringIO +from datetime import datetime, timedelta +import subprocess + +WEBSITE = 'http://www.newsdiffs.org/browse/' +MAX_TIME = timedelta(hours=1) +EMAILS = 'ecprice@mit.edu jenny8lee@gmail.com price@mit.edu'.split() + +def send_alert_email(subject, body): + email = 'Subject: %s\n\n%s' % (subject, body) + + p = subprocess.Popen(['/usr/bin/msmtp', '-t'] + EMAILS, + stdin=subprocess.PIPE) + p.communicate(email) + if p.wait(): + print 'Bad return code:', p.returncode + + +def get_update_time(): + html = urllib2.urlopen(WEBSITE) + soup = BeautifulSoup(html) + datestr = soup.findAll('td')[1].findChild('a').getText() + date = datetime.strptime(datestr.replace('.', ''), '%B %d, %Y, %I:%M %p') + return date + +if __name__ == '__main__': + try: + update_time = get_update_time() + time_since_update = datetime.now() - update_time + print 'Update time:', time_since_update + if time_since_update > MAX_TIME: + send_alert_email('Trouble with newsdiffs.org', + 'No updates since %s\n%s is too long' % + (update_time, time_since_update)) + except Exception, e: + import traceback + traceback.print_exc() + send_alert_email('Trouble with newsdiffs.org', + 'Cannot check website\n%s' % traceback.format_exc()) +
1340ff25daef84e66841476eaff9446886cda0b5
minsem.py
minsem.py
#!/usr/bin/env python from collections import OrderedDict class DataToken: def __init__(self, line: str): ( self.offset, self.word, self.lowercase_lemma, self.pos_tag, self.mwe_tag, self.parent_offset, self.strength, self.supersense, self.sentence_id ) = line.split('\t') class DataSentence: def __init__(self): self._tokens: OrderedDict = OrderedDict() self.sentence_id = None def __iter__(self): return iter(self._tokens) def __getitem__(self, item): return self._tokens[item] def __bool__(self): return bool(self._tokens) def append(self, token: DataToken): if not self.sentence_id: self.sentence_id = token.sentence_id else: if token.sentence_id != self.sentence_id: raise ValueError("sentence IDs do not match: {} (sentence) != {} (token)".format(self.sentence_id, token.sentence_id)) self._tokens[int(token.offset)] = token @property def sentence(self): return ' '.join(map(lambda t: t.word, self._tokens.values())) def read_data_file(datafile: str): data_sentences = {} with open(datafile) as df: sentence = DataSentence() for line in df: line = line.strip(' \n') if not line: # An empty line indicates the end of a sentence. if sentence: data_sentences[sentence.sentence_id] = sentence sentence = DataSentence() else: # A new token to be accumulated! token = DataToken(line) sentence.append(token) # Check if there is a valid sentence; this would happen if the file does not end in a newline. if sentence: data_sentences[sentence.sentence_id] = sentence return data_sentences if __name__ == '__main__': import argparse parser = argparse.ArgumentParser() parser.add_argument('training-data', help='the datafile to train with') parser.add_argument('testing-data', help='the datafile to test with') args = parser.parse_args() # Read the data. training_sentences = read_data_file(args.training_data) testing_sentences = read_data_file(args.testing_data)
Implement reading from data files
Implement reading from data files Reads files and stores contents in new DataSentence/DataToken constructions.
Python
mit
pdarragh/MinSem
--- +++ @@ -0,0 +1,79 @@ +#!/usr/bin/env python + +from collections import OrderedDict + + +class DataToken: + def __init__(self, line: str): + ( + self.offset, + self.word, + self.lowercase_lemma, + self.pos_tag, + self.mwe_tag, + self.parent_offset, + self.strength, + self.supersense, + self.sentence_id + ) = line.split('\t') + + +class DataSentence: + def __init__(self): + self._tokens: OrderedDict = OrderedDict() + self.sentence_id = None + + def __iter__(self): + return iter(self._tokens) + + def __getitem__(self, item): + return self._tokens[item] + + def __bool__(self): + return bool(self._tokens) + + def append(self, token: DataToken): + if not self.sentence_id: + self.sentence_id = token.sentence_id + else: + if token.sentence_id != self.sentence_id: + raise ValueError("sentence IDs do not match: {} (sentence) != {} (token)".format(self.sentence_id, + token.sentence_id)) + self._tokens[int(token.offset)] = token + + @property + def sentence(self): + return ' '.join(map(lambda t: t.word, self._tokens.values())) + + +def read_data_file(datafile: str): + data_sentences = {} + with open(datafile) as df: + sentence = DataSentence() + for line in df: + line = line.strip(' \n') + if not line: + # An empty line indicates the end of a sentence. + if sentence: + data_sentences[sentence.sentence_id] = sentence + sentence = DataSentence() + else: + # A new token to be accumulated! + token = DataToken(line) + sentence.append(token) + # Check if there is a valid sentence; this would happen if the file does not end in a newline. + if sentence: + data_sentences[sentence.sentence_id] = sentence + return data_sentences + + +if __name__ == '__main__': + import argparse + parser = argparse.ArgumentParser() + parser.add_argument('training-data', help='the datafile to train with') + parser.add_argument('testing-data', help='the datafile to test with') + args = parser.parse_args() + + # Read the data. + training_sentences = read_data_file(args.training_data) + testing_sentences = read_data_file(args.testing_data)
d95181d8de55c77e10aca390a40074d5ef0a7bb2
angr/procedures/java_lang/string_equals.py
angr/procedures/java_lang/string_equals.py
from ..java import JavaSimProcedure import logging l = logging.getLogger('angr.procedures.java.string.equals') class StringEquals(JavaSimProcedure): NO_RET = True __provides__ = ( ("java.lang.String", "equals(java.lang.String)"), ) def run(self, str_1, str_2): l.info("Called SimProcedure java.string.equals with args: %s (%r), %s (%r)", str_1, str_1, str_2, str_2)
Add dummy Simprocedure for java.lang.string.equals
Add dummy Simprocedure for java.lang.string.equals
Python
bsd-2-clause
schieb/angr,schieb/angr,iamahuman/angr,angr/angr,angr/angr,schieb/angr,angr/angr,iamahuman/angr,iamahuman/angr
--- +++ @@ -0,0 +1,16 @@ +from ..java import JavaSimProcedure +import logging + +l = logging.getLogger('angr.procedures.java.string.equals') + +class StringEquals(JavaSimProcedure): + + NO_RET = True + + __provides__ = ( + ("java.lang.String", "equals(java.lang.String)"), + ) + + def run(self, str_1, str_2): + l.info("Called SimProcedure java.string.equals with args: %s (%r), %s (%r)", str_1, str_1, str_2, str_2) +
390430b676d88dbe26d28a5551257b3013aee19b
copy_selected_photos.py
copy_selected_photos.py
import pandas as pd def main(): # create a directory for selected photos print 'mkdir photos_las_vegas_food_drinks' # load data about selected photos selected_photos_business = pd.read_csv('las_vegas_food_drinks.csv') # for each photo generate copy command for photo_id in selected_photos_business['photo_id']: print 'cp -v photos/' + photo_id + '.jpg photos_las_vegas_food_drinks' if __name__ == "__main__": main()
Add copying of selected photos to a new directory
Add copying of selected photos to a new directory
Python
mit
aysent/yelp-photo-explorer
--- +++ @@ -0,0 +1,16 @@ +import pandas as pd + +def main(): + + # create a directory for selected photos + print 'mkdir photos_las_vegas_food_drinks' + + # load data about selected photos + selected_photos_business = pd.read_csv('las_vegas_food_drinks.csv') + + # for each photo generate copy command + for photo_id in selected_photos_business['photo_id']: + print 'cp -v photos/' + photo_id + '.jpg photos_las_vegas_food_drinks' + +if __name__ == "__main__": + main()
868c29b23d74ef2dced64eed545995d43f0256ef
registries/serializers.py
registries/serializers.py
from rest_framework import serializers from registries.models import Organization from gwells.models import ProvinceState class DrillerListSerializer(serializers.ModelSerializer): province_state = serializers.ReadOnlyField() class Meta: model = Organization # Using all fields for now fields = ( #'who_created', #'when_created', #'who_updated', #'when_updated', 'name', 'street_address', 'city', 'province_state', 'postal_code', 'main_tel', 'fax_tel', 'website_url', 'certificate_authority', )
Add simple driller list serializer
Add simple driller list serializer
Python
apache-2.0
rstens/gwells,rstens/gwells,bcgov/gwells,rstens/gwells,rstens/gwells,bcgov/gwells,bcgov/gwells,bcgov/gwells
--- +++ @@ -0,0 +1,26 @@ +from rest_framework import serializers +from registries.models import Organization +from gwells.models import ProvinceState + +class DrillerListSerializer(serializers.ModelSerializer): + province_state = serializers.ReadOnlyField() + + class Meta: + model = Organization + + # Using all fields for now + fields = ( + #'who_created', + #'when_created', + #'who_updated', + #'when_updated', + 'name', + 'street_address', + 'city', + 'province_state', + 'postal_code', + 'main_tel', + 'fax_tel', + 'website_url', + 'certificate_authority', + )
dd0fd45f1324696144267da78a6fe62a8fc346cf
lpthw/ex24.py
lpthw/ex24.py
print "Let's practice everything." print 'You\'d need to know \'bout escapes with \\ that do \n newlines and \t tabs." poem = """ \t the lovely world wtih logic so firmly planted cannot discern \n the needs of love nor comprehend passion from intuition and requires an explanation \n\t\twhere there is none. """ print "--------------" print poem print "--------------" five = 10 - 2 + 3 - 6 print "This should be five: %s" % five def secret_formula(started): """ This is not the Krabby Patty Secret Formula (tm) """ jelly_beans = started * 500 jars = jelly_beans / 1000 crates = jars / 100 return jelly_beans, jars, crates start_point = 10000 beans, jars, crates = secret_formula(start_point) print "With a starting point of %d" % start_point print "We'd have %d beans %d jars, and %d crates." % (beans, jars, crates) start_point = start_point / 10 print "We can also do that this way:" # This part is pretty darn cool. \/ print "We'd have %d beans, %d jars, and %d crates." % secret_formula(start_point)
Add work for Exercise 24
Add work for Exercise 24
Python
mit
jaredmanning/learning,jaredmanning/learning
--- +++ @@ -0,0 +1,43 @@ +print "Let's practice everything." +print 'You\'d need to know \'bout escapes with \\ that do \n newlines and \t tabs." + +poem = """ +\t the lovely world +wtih logic so firmly planted +cannot discern \n the needs of love +nor comprehend passion from intuition +and requires an explanation +\n\t\twhere there is none. +""" + +print "--------------" +print poem +print "--------------" + + +five = 10 - 2 + 3 - 6 +print "This should be five: %s" % five + +def secret_formula(started): + """ + This is not the Krabby + Patty Secret Formula (tm) + """ + jelly_beans = started * 500 + jars = jelly_beans / 1000 + crates = jars / 100 + return jelly_beans, jars, crates + + +start_point = 10000 +beans, jars, crates = secret_formula(start_point) + +print "With a starting point of %d" % start_point +print "We'd have %d beans %d jars, and %d crates." % (beans, jars, crates) + +start_point = start_point / 10 + +print "We can also do that this way:" +# This part is pretty darn cool. \/ +print "We'd have %d beans, %d jars, and %d crates." % secret_formula(start_point) +
610396a7caa1c23532369d4bc5e382dbc21cba60
tests/chainer_tests/training_tests/extensions_tests/test_plot_report.py
tests/chainer_tests/training_tests/extensions_tests/test_plot_report.py
import unittest from chainer.training import extensions class TestPlotReport(unittest.TestCase): def test_available(self): try: from matplotlib import pyplot # NOQA available = True except ImportError: available = False self.assertEqual(extensions.PlotReport.available(), available)
Fix PlotReport.available and write test
Fix PlotReport.available and write test
Python
mit
ktnyt/chainer,hvy/chainer,jnishi/chainer,anaruse/chainer,chainer/chainer,okuta/chainer,cupy/cupy,niboshi/chainer,cupy/cupy,ysekky/chainer,niboshi/chainer,keisuke-umezawa/chainer,tkerola/chainer,okuta/chainer,chainer/chainer,ktnyt/chainer,rezoo/chainer,wkentaro/chainer,jnishi/chainer,aonotas/chainer,keisuke-umezawa/chainer,okuta/chainer,keisuke-umezawa/chainer,kashif/chainer,wkentaro/chainer,pfnet/chainer,chainer/chainer,jnishi/chainer,ktnyt/chainer,chainer/chainer,jnishi/chainer,ronekko/chainer,cupy/cupy,wkentaro/chainer,hvy/chainer,niboshi/chainer,ktnyt/chainer,wkentaro/chainer,delta2323/chainer,okuta/chainer,hvy/chainer,keisuke-umezawa/chainer,niboshi/chainer,hvy/chainer,cupy/cupy,kiyukuta/chainer
--- +++ @@ -0,0 +1,15 @@ +import unittest + +from chainer.training import extensions + + +class TestPlotReport(unittest.TestCase): + + def test_available(self): + try: + from matplotlib import pyplot # NOQA + available = True + except ImportError: + available = False + + self.assertEqual(extensions.PlotReport.available(), available)
4dd4557796acca3046806a9504d8d6bcb78ca16d
test/test_UrbanDict.py
test/test_UrbanDict.py
#!/usr/bin/env python ### # Copyright (c) 2004, Kevin Murphy # All rights reserved. # # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions are met: # # * Redistributions of source code must retain the above copyright notice, # this list of conditions, and the following disclaimer. # * Redistributions in binary form must reproduce the above copyright notice, # this list of conditions, and the following disclaimer in the # documentation and/or other materials provided with the distribution. # * Neither the name of the author of this software nor the name of # contributors to this software may be used to endorse or promote products # derived from this software without specific prior written consent. # # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" # AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE # IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE # ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE # LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR # CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF # SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS # INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN # CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) # ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE # POSSIBILITY OF SUCH DAMAGE. ### from testsupport import * class UrbanDictTestCase(ChannelPluginTestCase, PluginDocumentation): plugins = ('UrbanDict',) def setUp(self): ChannelPluginTestCase.setUp(self) if network: def testUrbanDict(self): self.assertNotError('ud') self.assertRegexp('ud somethinginvalid', 'Error: No definition found.') self.assertRegexp('ud yo', 'a word that died a horrible death in 1993') self.assertRegexp('ud bozac', '"Druid Diallect"') # vim:set shiftwidth=4 tabstop=8 expandtab textwidth=78:
Test cases for UrbanDict plugin
Test cases for UrbanDict plugin
Python
bsd-3-clause
kblin/supybot-gsoc,ProgVal/Limnoria-test,raboof/supybot,Ban3/Limnoria,Ban3/Limnoria,jeffmahoney/supybot,mazaclub/mazabot-core,buildbot/supybot,mazaclub/mazabot-core,frumiousbandersnatch/supybot-code,prashantpawar/supybot-rothbot,ProgVal/Limnoria-test,haxwithaxe/supybot
--- +++ @@ -0,0 +1,50 @@ +#!/usr/bin/env python + +### +# Copyright (c) 2004, Kevin Murphy +# All rights reserved. +# +# Redistribution and use in source and binary forms, with or without +# modification, are permitted provided that the following conditions are met: +# +# * Redistributions of source code must retain the above copyright notice, +# this list of conditions, and the following disclaimer. +# * Redistributions in binary form must reproduce the above copyright notice, +# this list of conditions, and the following disclaimer in the +# documentation and/or other materials provided with the distribution. +# * Neither the name of the author of this software nor the name of +# contributors to this software may be used to endorse or promote products +# derived from this software without specific prior written consent. +# +# THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" +# AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE +# IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE +# ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR CONTRIBUTORS BE +# LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR +# CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF +# SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS +# INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN +# CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) +# ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE +# POSSIBILITY OF SUCH DAMAGE. +### + +from testsupport import * + +class UrbanDictTestCase(ChannelPluginTestCase, PluginDocumentation): + plugins = ('UrbanDict',) + def setUp(self): + ChannelPluginTestCase.setUp(self) + + if network: + def testUrbanDict(self): + self.assertNotError('ud') + self.assertRegexp('ud somethinginvalid', + 'Error: No definition found.') + self.assertRegexp('ud yo', + 'a word that died a horrible death in 1993') + self.assertRegexp('ud bozac', + '"Druid Diallect"') + +# vim:set shiftwidth=4 tabstop=8 expandtab textwidth=78: +
263194056dbd749681f36823b77bb0ba2c76e379
tests/test_es.py
tests/test_es.py
# -*- coding: utf-8 -*- """Elasticsearch client test cases.""" import unittest from sqlalchemy import types as sql_types from esis.es import Mapping class MappingTest(unittest.TestCase): """Test translation from SQL schema to Elasticsearch mapping.""" def test_mapping_types(self): """Test mapping from sql to Elasticsearch index types.""" table_name = 'my_table' table_schema = { 'my_bigint': sql_types.BIGINT(), 'my_boolean': sql_types.BOOLEAN(), 'my_char': sql_types.CHAR(16), 'my_clob': sql_types.CLOB(), 'my_date': sql_types.DATE(), 'my_datetime': sql_types.DATETIME(), 'my_decimal': sql_types.DECIMAL(10, 5), 'my_float': sql_types.FLOAT(), 'my_integer': sql_types.INTEGER(), 'my_nchar': sql_types.NCHAR(16), 'my_nvarchar': sql_types.NVARCHAR(16), 'my_null': sql_types.NullType(), 'my_numeric': sql_types.NUMERIC(), 'my_real': sql_types.REAL(), 'my_smallint': sql_types.SMALLINT(), 'my_text': sql_types.TEXT(), 'my_timestamp': sql_types.TIMESTAMP(), 'my_varchar': sql_types.VARCHAR(16), } mapping = Mapping(table_name, table_schema) self.assertDictEqual( mapping.mapping, { table_name: { 'properties': { 'my_bigint': {'type': 'long'}, 'my_boolean': {'type': 'boolean'}, 'my_char': {'type': 'string'}, 'my_clob': {'type': 'string'}, 'my_datetime': {'type': 'date'}, 'my_float': {'type': 'float'}, 'my_integer': {'type': 'long'}, 'my_nchar': {'type': 'string'}, 'my_nvarchar': {'type': 'string'}, 'my_real': {'type': 'double'}, 'my_smallint': {'type': 'integer'}, 'my_text': {'type': 'string'}, 'my_timestamp': {'type': 'date'}, 'my_varchar': {'type': 'string'}, }, }, }, )
Add Elasticsearch mapping test case
Add Elasticsearch mapping test case
Python
mit
jcollado/esis
--- +++ @@ -0,0 +1,61 @@ +# -*- coding: utf-8 -*- +"""Elasticsearch client test cases.""" + +import unittest + +from sqlalchemy import types as sql_types + +from esis.es import Mapping + + +class MappingTest(unittest.TestCase): + + """Test translation from SQL schema to Elasticsearch mapping.""" + + def test_mapping_types(self): + """Test mapping from sql to Elasticsearch index types.""" + table_name = 'my_table' + table_schema = { + 'my_bigint': sql_types.BIGINT(), + 'my_boolean': sql_types.BOOLEAN(), + 'my_char': sql_types.CHAR(16), + 'my_clob': sql_types.CLOB(), + 'my_date': sql_types.DATE(), + 'my_datetime': sql_types.DATETIME(), + 'my_decimal': sql_types.DECIMAL(10, 5), + 'my_float': sql_types.FLOAT(), + 'my_integer': sql_types.INTEGER(), + 'my_nchar': sql_types.NCHAR(16), + 'my_nvarchar': sql_types.NVARCHAR(16), + 'my_null': sql_types.NullType(), + 'my_numeric': sql_types.NUMERIC(), + 'my_real': sql_types.REAL(), + 'my_smallint': sql_types.SMALLINT(), + 'my_text': sql_types.TEXT(), + 'my_timestamp': sql_types.TIMESTAMP(), + 'my_varchar': sql_types.VARCHAR(16), + } + mapping = Mapping(table_name, table_schema) + self.assertDictEqual( + mapping.mapping, + { + table_name: { + 'properties': { + 'my_bigint': {'type': 'long'}, + 'my_boolean': {'type': 'boolean'}, + 'my_char': {'type': 'string'}, + 'my_clob': {'type': 'string'}, + 'my_datetime': {'type': 'date'}, + 'my_float': {'type': 'float'}, + 'my_integer': {'type': 'long'}, + 'my_nchar': {'type': 'string'}, + 'my_nvarchar': {'type': 'string'}, + 'my_real': {'type': 'double'}, + 'my_smallint': {'type': 'integer'}, + 'my_text': {'type': 'string'}, + 'my_timestamp': {'type': 'date'}, + 'my_varchar': {'type': 'string'}, + }, + }, + }, + )
32f48a7f16fe5657ac02ea5c8521ecb068e4ac77
pontoon/base/migrations/0013_add_en_US.py
pontoon/base/migrations/0013_add_en_US.py
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import models, migrations def add_back_source_locales(apps, schema_editor): Locale = apps.get_model('base', 'Locale') Locale.objects.create( code='en', name='English', nplurals=2, plural_rule='(n != 1)', cldr_plurals='1,5' ) Locale.objects.create( code='en-US', name='English', nplurals=2, plural_rule='(n != 1)', cldr_plurals='1,5' ) class Migration(migrations.Migration): dependencies = [ ('base', '0012_auto_20150804_0859'), ] operations = [ migrations.RunPython(add_back_source_locales) ]
Add en and en-US locales back to locale list.
Add en and en-US locales back to locale list. Both are source locales and shouldn't have been removed.
Python
bsd-3-clause
mozilla/pontoon,Osmose/pontoon,mastizada/pontoon,mathjazz/pontoon,vivekanand1101/pontoon,sudheesh001/pontoon,mastizada/pontoon,participedia/pontoon,participedia/pontoon,jotes/pontoon,m8ttyB/pontoon,yfdyh000/pontoon,sudheesh001/pontoon,jotes/pontoon,vivekanand1101/pontoon,Osmose/pontoon,Osmose/pontoon,sudheesh001/pontoon,yfdyh000/pontoon,mastizada/pontoon,mathjazz/pontoon,yfdyh000/pontoon,m8ttyB/pontoon,m8ttyB/pontoon,participedia/pontoon,yfdyh000/pontoon,sudheesh001/pontoon,vivekanand1101/pontoon,participedia/pontoon,mathjazz/pontoon,mozilla/pontoon,mastizada/pontoon,mozilla/pontoon,m8ttyB/pontoon,mathjazz/pontoon,jotes/pontoon,jotes/pontoon,mathjazz/pontoon,vivekanand1101/pontoon,Osmose/pontoon,mozilla/pontoon,mozilla/pontoon
--- +++ @@ -0,0 +1,33 @@ +# -*- coding: utf-8 -*- +from __future__ import unicode_literals + +from django.db import models, migrations + + +def add_back_source_locales(apps, schema_editor): + Locale = apps.get_model('base', 'Locale') + Locale.objects.create( + code='en', + name='English', + nplurals=2, + plural_rule='(n != 1)', + cldr_plurals='1,5' + ) + Locale.objects.create( + code='en-US', + name='English', + nplurals=2, + plural_rule='(n != 1)', + cldr_plurals='1,5' + ) + + +class Migration(migrations.Migration): + + dependencies = [ + ('base', '0012_auto_20150804_0859'), + ] + + operations = [ + migrations.RunPython(add_back_source_locales) + ]
82943db5942f6d0749776efb1b1acf928af97a1c
example.py
example.py
#!/usr/bin/env python3 import numpy as np import rust_sorting as rs array = np.zeros((5,), dtype=np.int8) rs.quicksort(array) print("Python done!")
Call quicksort from Python; it works!
Call quicksort from Python; it works!
Python
bsd-3-clause
nbigaouette/rust-sorting,nbigaouette/rust-sorting,nbigaouette/rust-sorting
--- +++ @@ -0,0 +1,12 @@ +#!/usr/bin/env python3 + +import numpy as np +import rust_sorting as rs + + +array = np.zeros((5,), dtype=np.int8) + + +rs.quicksort(array) + +print("Python done!")
b12dd992da93d43e5acf1d42cee0f7d34f8367bf
spam/tests/fake_params.py
spam/tests/fake_params.py
#!/usr/bin/env python # -*- coding: utf-8 -*- import os DATASET_PATH = os.path.join( os.path.dirname(__file__), 'test_dataset', ) DATASET_SUBDIRS = [ { 'name': 'enron1', 'total_count': 5, 'ham_count': 3, 'spam_count': 2, 'path': os.path.join(DATASET_PATH, 'enron1'), 'ham_path': os.path.join( DATASET_PATH, 'enron1', 'ham' ), 'spam_path': os.path.join( DATASET_PATH, 'enron1', 'spam' ), }, { 'name': 'enron2', 'total_count': 6, 'ham_count': 2, 'spam_count': 4, 'path': os.path.join(DATASET_PATH, 'enron2'), 'ham_path': os.path.join( DATASET_PATH, 'enron2', 'ham' ), 'spam_path': os.path.join( DATASET_PATH, 'enron2', 'spam' ), }, ]
Add fake params from test.
Add fake params from test.
Python
mit
benigls/spam,benigls/spam
--- +++ @@ -0,0 +1,47 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- + +import os + + +DATASET_PATH = os.path.join( + os.path.dirname(__file__), + 'test_dataset', +) + +DATASET_SUBDIRS = [ + { + 'name': 'enron1', + 'total_count': 5, + 'ham_count': 3, + 'spam_count': 2, + 'path': os.path.join(DATASET_PATH, 'enron1'), + 'ham_path': os.path.join( + DATASET_PATH, + 'enron1', + 'ham' + ), + 'spam_path': os.path.join( + DATASET_PATH, + 'enron1', + 'spam' + ), + }, + { + 'name': 'enron2', + 'total_count': 6, + 'ham_count': 2, + 'spam_count': 4, + 'path': os.path.join(DATASET_PATH, 'enron2'), + 'ham_path': os.path.join( + DATASET_PATH, + 'enron2', + 'ham' + ), + 'spam_path': os.path.join( + DATASET_PATH, + 'enron2', + 'spam' + ), + }, +]
0cdf73f158e425db0ab65d70966b18ef9f4af1c6
test/test_services_view.py
test/test_services_view.py
#!/usr/bin/env python # -*- coding: utf-8 -*- # Copyright (c) 2015-2016: # Matthieu Estrada, ttamalfor@gmail.com # # This file is part of (AlignakApp). # # (AlignakApp) is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # (AlignakApp) is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with (AlignakApp). If not, see <http://www.gnu.org/licenses/>. import sys import unittest2 from alignak_app.core.utils import set_app_config from alignak_app.synthesis.services_view import ServicesView try: __import__('PyQt5') from PyQt5.QtWidgets import QApplication except ImportError: from PyQt4.Qt import QApplication class TestServicesView(unittest2.TestCase): """ This file test the ServicesView class. """ set_app_config() @classmethod def setUpClass(cls): """Create QApplication""" try: cls.app = QApplication(sys.argv) except: pass def test_display_services(self): """Inititalize ServicesView""" under_test = ServicesView() self.assertIsNotNone(under_test.layout) under_test.display_services(None, 'name') self.assertIsNotNone(under_test.layout)
Add unit tests for services_view.py
Add unit tests for services_view.py
Python
agpl-3.0
Alignak-monitoring-contrib/alignak-app,Alignak-monitoring-contrib/alignak-app
--- +++ @@ -0,0 +1,61 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- + +# Copyright (c) 2015-2016: +# Matthieu Estrada, ttamalfor@gmail.com +# +# This file is part of (AlignakApp). +# +# (AlignakApp) is free software: you can redistribute it and/or modify +# it under the terms of the GNU Affero General Public License as published by +# the Free Software Foundation, either version 3 of the License, or +# (at your option) any later version. +# +# (AlignakApp) is distributed in the hope that it will be useful, +# but WITHOUT ANY WARRANTY; without even the implied warranty of +# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the +# GNU Affero General Public License for more details. +# +# You should have received a copy of the GNU Affero General Public License +# along with (AlignakApp). If not, see <http://www.gnu.org/licenses/>. + +import sys + +import unittest2 + +from alignak_app.core.utils import set_app_config +from alignak_app.synthesis.services_view import ServicesView + +try: + __import__('PyQt5') + from PyQt5.QtWidgets import QApplication +except ImportError: + from PyQt4.Qt import QApplication + + +class TestServicesView(unittest2.TestCase): + """ + This file test the ServicesView class. + """ + + set_app_config() + + @classmethod + def setUpClass(cls): + """Create QApplication""" + try: + cls.app = QApplication(sys.argv) + except: + pass + + def test_display_services(self): + """Inititalize ServicesView""" + + under_test = ServicesView() + + self.assertIsNotNone(under_test.layout) + + under_test.display_services(None, 'name') + + self.assertIsNotNone(under_test.layout) +
78e84abea9b0cc41bdb1aa8b9ea5c9953a37941d
bdbcontrib/contrib_exp.py
bdbcontrib/contrib_exp.py
from bayeslite.shell.hook import bayesdb_shell_cmd from bdbcontrib.general_utils import ArgparseError, ArgumentParser import bdbexp experiments = { 'haystacks' : bdbexp.haystacks, 'hyperparams' : bdbexp.hyperparams, 'infer' : bdbexp.infer, 'kl_divergence' : bdbexp.kl_divergence, 'permute' : bdbexp.permute, 'predictive_pdf' : bdbexp.predictive_pdf, 'recover' : bdbexp.recover } @bayesdb_shell_cmd('experiment') def experiment(self, argin): ''' Launch an experimental inference quality test. USAGE: .experiment <exp_name> [exp_args ...] <exp_name> permute kl_divergence haystacks predictive_pdf hyperparams recover infer [exp_args] To see experiment specific arguments, use .experiment <exp_name> --help Examples: bayeslite> .experiment predictive_pdf --help bayeslite> .experiment haystacks --n_iter=200 --n_distractors=4 ''' expname = argin.split()[0] if argin != '' else argin if expname not in experiments: print 'Invalid experiment {}'.format(expname) print 'For help use: .help experiment' return try: experiments[argin.split()[0]].main(argin.split()[1:], halt=True) except ArgparseError as e: self.stdout.write('%s' % (e.message,)) return except SystemExit: return # This happens when --help is invoked.
Add hooks for inference experiments.
Add hooks for inference experiments.
Python
apache-2.0
probcomp/bdbcontrib,probcomp/bdbcontrib
--- +++ @@ -0,0 +1,48 @@ +from bayeslite.shell.hook import bayesdb_shell_cmd +from bdbcontrib.general_utils import ArgparseError, ArgumentParser + +import bdbexp + +experiments = { + 'haystacks' : bdbexp.haystacks, + 'hyperparams' : bdbexp.hyperparams, + 'infer' : bdbexp.infer, + 'kl_divergence' : bdbexp.kl_divergence, + 'permute' : bdbexp.permute, + 'predictive_pdf' : bdbexp.predictive_pdf, + 'recover' : bdbexp.recover + } + +@bayesdb_shell_cmd('experiment') +def experiment(self, argin): + ''' + Launch an experimental inference quality test. + USAGE: .experiment <exp_name> [exp_args ...] + + <exp_name> + permute kl_divergence + haystacks predictive_pdf + hyperparams recover + infer + + [exp_args] + To see experiment specific arguments, use + .experiment <exp_name> --help + + Examples: + bayeslite> .experiment predictive_pdf --help + bayeslite> .experiment haystacks --n_iter=200 --n_distractors=4 + ''' + expname = argin.split()[0] if argin != '' else argin + if expname not in experiments: + print 'Invalid experiment {}'.format(expname) + print 'For help use: .help experiment' + return + + try: + experiments[argin.split()[0]].main(argin.split()[1:], halt=True) + except ArgparseError as e: + self.stdout.write('%s' % (e.message,)) + return + except SystemExit: + return # This happens when --help is invoked.
c860d536710d2d9546924b00aef7de4f516e38a5
tools/fixup_io_tilegrid.py
tools/fixup_io_tilegrid.py
import database import tiles import json from os import path """ Despite Lattice assigning them the same tile type; "odd" and "even" top/left/right IO locations have slightly different routing - swapped output tristate and data This script fixes this by patching tile names """ for f, d in [("LIFCL", "LIFCL-40")]: tgp = path.join(database.get_db_root(), f, d, "tilegrid.json") with open(tgp, "r") as infile: tg = json.load(infile)["tiles"] tiles_by_xy = [[]] max_row = 0 max_col = 0 for tile in sorted(tg.keys()): r, c = tiles.pos_from_name(tile) max_row = max(r, max_row) max_col = max(c, max_col) while r >= len(tiles_by_xy): tiles_by_xy.append([]) while c >= len(tiles_by_xy[r]): tiles_by_xy[r].append([]) tiles_by_xy[r][c].append(tile) # Top tiles is_odd = False for col in tiles_by_xy[0]: for tile in col: tt = tiles.type_from_fullname(tile) if not tt.startswith("SYSIO"): continue # Don't rename special or already-renamed tiles if tt[-1].isdigit(): new_name = tile + ("_ODD" if is_odd else "_EVEN") assert new_name not in tg tg[new_name] = dict(tg[tile]) tg[new_name]["tiletype"] = tg[new_name]["tiletype"] + ("_ODD" if is_odd else "_EVEN") del tg[tile] is_odd = not is_odd # Left/right tiles for tc in (0, max_col): is_odd = False bank = "" for row in tiles_by_xy: for tile in row[tc]: tt = tiles.type_from_fullname(tile) if not tt.startswith("SYSIO"): continue if tt.endswith("REM"): continue tile_bank = tt[tt.find("B")+1] if tile_bank != bank: is_odd = False bank = tile_bank if tt[-1].isdigit(): new_name = tile + ("_ODD" if is_odd else "_EVEN") assert new_name not in tg tg[new_name] = dict(tg[tile]) tg[new_name]["tiletype"] = tg[new_name]["tiletype"] + ("_ODD" if is_odd else "_EVEN") del tg[tile] is_odd = not is_odd with open(tgp, "w") as outfile: json.dump({"tiles": tg}, outfile, sort_keys=True, indent=4)
Add tool to fix IO tilegrid
Add tool to fix IO tilegrid Signed-off-by: David Shah <bfcdf3e6ca6cef45543bfbb57509c92aec9a39fb@ds0.me>
Python
isc
gatecat/prjoxide,gatecat/prjoxide,gatecat/prjoxide
--- +++ @@ -0,0 +1,71 @@ +import database +import tiles +import json +from os import path + +""" +Despite Lattice assigning them the same tile type; "odd" and "even" top/left/right IO +locations have slightly different routing - swapped output tristate and data + +This script fixes this by patching tile names +""" + +for f, d in [("LIFCL", "LIFCL-40")]: + tgp = path.join(database.get_db_root(), f, d, "tilegrid.json") + with open(tgp, "r") as infile: + tg = json.load(infile)["tiles"] + + tiles_by_xy = [[]] + max_row = 0 + max_col = 0 + for tile in sorted(tg.keys()): + r, c = tiles.pos_from_name(tile) + max_row = max(r, max_row) + max_col = max(c, max_col) + while r >= len(tiles_by_xy): + tiles_by_xy.append([]) + while c >= len(tiles_by_xy[r]): + tiles_by_xy[r].append([]) + tiles_by_xy[r][c].append(tile) + + # Top tiles + is_odd = False + for col in tiles_by_xy[0]: + for tile in col: + tt = tiles.type_from_fullname(tile) + if not tt.startswith("SYSIO"): + continue + # Don't rename special or already-renamed tiles + if tt[-1].isdigit(): + new_name = tile + ("_ODD" if is_odd else "_EVEN") + assert new_name not in tg + tg[new_name] = dict(tg[tile]) + tg[new_name]["tiletype"] = tg[new_name]["tiletype"] + ("_ODD" if is_odd else "_EVEN") + del tg[tile] + is_odd = not is_odd + + + # Left/right tiles + for tc in (0, max_col): + is_odd = False + bank = "" + for row in tiles_by_xy: + for tile in row[tc]: + tt = tiles.type_from_fullname(tile) + if not tt.startswith("SYSIO"): + continue + if tt.endswith("REM"): + continue + tile_bank = tt[tt.find("B")+1] + if tile_bank != bank: + is_odd = False + bank = tile_bank + if tt[-1].isdigit(): + new_name = tile + ("_ODD" if is_odd else "_EVEN") + assert new_name not in tg + tg[new_name] = dict(tg[tile]) + tg[new_name]["tiletype"] = tg[new_name]["tiletype"] + ("_ODD" if is_odd else "_EVEN") + del tg[tile] + is_odd = not is_odd + with open(tgp, "w") as outfile: + json.dump({"tiles": tg}, outfile, sort_keys=True, indent=4)
c0f06b64c15d74be26be2cd9e6d593e5c5cae2a9
tools/xml_select_minmax.py
tools/xml_select_minmax.py
#! /usr/bin/python3 import sys import argparse import xml_utils as u import os from argparse import RawTextHelpFormatter ##---------------------------------------------------------- ## for each label that has more than the mininum count, select the ## largest subset less than the maxinum count. ## writes out to a new xml file. ##---------------------------------------------------------- def main (argv) : parser = argparse.ArgumentParser(description='Select a subset if label count falls between min and max.', formatter_class=RawTextHelpFormatter) parser.add_argument ('image_db') parser.add_argument ('min', default=0) parser.add_argument ('max', default=0) parser.add_argument ('files', nargs='+') parser.add_argument ('-o', '-out', '--output') parser.add_argument ('-filetype', '--filetype', default="chips", help='Type of input file <images|faces|chips>. Defaults to "chips".') parser.add_argument ('-v', '--verbosity', type=int, default=1, choices=[0, 1, 2, 3], help='') # help="increase output verbosity" u.set_argv (argv) args = parser.parse_args() u.set_verbosity (args.verbosity) u.set_argv (argv) u.set_filetype ('faces') verbose = 0 if verbose > 0: print("files: ", args.files) filetypes = ['images', 'chips', 'faces'] filetype = args.filetype if filetype not in filetypes : print('unrecognized filetype :', filetype, 'should be one of:', filetypes) return output_file = 'selected_' + u.current_datetime () + '.xml' if args.output : # user specified if not os.path.exists (args.output) : output_file = args.output else : print ('output file exists, writing to', output_file) u.select_labels_minmax (args.files, args.image_db, args.min, args.max, output_file, filetype) if __name__ == "__main__": main (sys.argv)
Select a subset for each label. If the label image count is greater than min, return the greatest number of images less than or equal to the max. The images will be evenly distributed across years, seasons and days.
Select a subset for each label. If the label image count is greater than min, return the greatest number of images less than or equal to the max. The images will be evenly distributed across years, seasons and days.
Python
mit
hypraptive/bearid,hypraptive/bearid,hypraptive/bearid
--- +++ @@ -0,0 +1,51 @@ +#! /usr/bin/python3 + +import sys +import argparse +import xml_utils as u +import os +from argparse import RawTextHelpFormatter + +##---------------------------------------------------------- +## for each label that has more than the mininum count, select the +## largest subset less than the maxinum count. +## writes out to a new xml file. +##---------------------------------------------------------- +def main (argv) : + parser = argparse.ArgumentParser(description='Select a subset if label count falls between min and max.', + formatter_class=RawTextHelpFormatter) + parser.add_argument ('image_db') + parser.add_argument ('min', default=0) + parser.add_argument ('max', default=0) + parser.add_argument ('files', nargs='+') + parser.add_argument ('-o', '-out', '--output') + parser.add_argument ('-filetype', '--filetype', default="chips", + help='Type of input file <images|faces|chips>. Defaults to "chips".') + parser.add_argument ('-v', '--verbosity', type=int, default=1, + choices=[0, 1, 2, 3], help='') + # help="increase output verbosity" + u.set_argv (argv) + args = parser.parse_args() + u.set_verbosity (args.verbosity) + u.set_argv (argv) + u.set_filetype ('faces') + verbose = 0 + if verbose > 0: + print("files: ", args.files) + + filetypes = ['images', 'chips', 'faces'] + filetype = args.filetype + if filetype not in filetypes : + print('unrecognized filetype :', filetype, 'should be one of:', filetypes) + return + output_file = 'selected_' + u.current_datetime () + '.xml' + if args.output : # user specified + if not os.path.exists (args.output) : + output_file = args.output + else : + print ('output file exists, writing to', output_file) + u.select_labels_minmax (args.files, args.image_db, args.min, args.max, output_file, filetype) + +if __name__ == "__main__": + main (sys.argv) +
5e1c8f5bbe3cd8927b60a08b2784f7d2ea8263f2
src/backend/live_classify_local_camera.py
src/backend/live_classify_local_camera.py
import openface import numpy as np import os import cv2 import pickle id_name = ["Alec", "Greg", "Phong", "Emil"] def classify(aligned_face, net, clf, le): rep = net.forward(aligned_face) predictions = clf.predict_proba(rep.reshape((1, len(rep)))).ravel() maxI = np.argmax(predictions) person = le.inverse_transform(maxI) confidence = predictions[maxI] print("Predict {} with {:.2f} confidence.".format(person, confidence)) return person # path to the face alignment model dLib_predictor = "../../resource/shaqpe_predictor_68_face_landmarks.dat" # construct the face alignment model align = openface.AlignDlib(dLib_predictor) # path to deep neural network for feature representation network_model = "../../resource/nn4.small2.v1.t7" # construct the network for feature represenation net = openface.TorchNeuralNet(network_model, 96) classifier_model = "../../resource/svm.pkl" with open(classifier_model, 'r') as f: (le, clf) = pickle.load(f) print("Successfully loaded SVM model") video = cv2.VideoCapture(0) if video is None: exit() while True: # grab image ret, cameraFrame = video.read() if not ret: exit() try: bbs = align.getAllFaceBoundingBoxes(cameraFrame) print("Found {} face".format(len(bbs))) for bb2 in bbs: alignedFace = align.align(96, cameraFrame, bb2, landmarkIndices=openface.AlignDlib.OUTER_EYES_AND_NOSE) id = classify(alignedFace, net, clf, le) person_name = id_name[id] print(person_name) rectColor = (0,255, 0) textColor = (255, 0, 0) face_top_left = (bb2.left(), bb2.top()) face_bottom_right = (bb2.right(), bb2.bottom()) cv2.rectangle(cameraFrame, face_top_left, face_bottom_right, rectColor) cv2.putText(cameraFrame, person_name, face_top_left, fontFace=cv2.FONT_HERSHEY_SIMPLEX, fontScale=1, color=textColor, thickness=2) cv2.imshow('FaceRecognizer', cameraFrame) if cv2.waitKey(1) & 0xFF == ord('q'): break except: cv2.imshow('FaceRecognize', cameraFrame) continue
Use the model to classify face in local camera.
Use the model to classify face in local camera.
Python
apache-2.0
xphongvn/smart-attendance-system-ta,xphongvn/smart-attendance-system-ta,xphongvn/smart-attendance-system-ta
--- +++ @@ -0,0 +1,65 @@ +import openface +import numpy as np +import os +import cv2 +import pickle + +id_name = ["Alec", "Greg", "Phong", "Emil"] + + +def classify(aligned_face, net, clf, le): + rep = net.forward(aligned_face) + predictions = clf.predict_proba(rep.reshape((1, len(rep)))).ravel() + maxI = np.argmax(predictions) + person = le.inverse_transform(maxI) + confidence = predictions[maxI] + print("Predict {} with {:.2f} confidence.".format(person, confidence)) + return person + +# path to the face alignment model +dLib_predictor = "../../resource/shaqpe_predictor_68_face_landmarks.dat" +# construct the face alignment model +align = openface.AlignDlib(dLib_predictor) +# path to deep neural network for feature representation +network_model = "../../resource/nn4.small2.v1.t7" +# construct the network for feature represenation +net = openface.TorchNeuralNet(network_model, 96) + +classifier_model = "../../resource/svm.pkl" + +with open(classifier_model, 'r') as f: + (le, clf) = pickle.load(f) + print("Successfully loaded SVM model") + video = cv2.VideoCapture(0) + if video is None: + exit() + + while True: + # grab image + ret, cameraFrame = video.read() + if not ret: + exit() + try: + bbs = align.getAllFaceBoundingBoxes(cameraFrame) + print("Found {} face".format(len(bbs))) + for bb2 in bbs: + alignedFace = align.align(96, cameraFrame, bb2, + landmarkIndices=openface.AlignDlib.OUTER_EYES_AND_NOSE) + id = classify(alignedFace, net, clf, le) + person_name = id_name[id] + print(person_name) + + rectColor = (0,255, 0) + textColor = (255, 0, 0) + face_top_left = (bb2.left(), bb2.top()) + face_bottom_right = (bb2.right(), bb2.bottom()) + cv2.rectangle(cameraFrame, face_top_left, face_bottom_right, rectColor) + cv2.putText(cameraFrame, person_name, face_top_left, + fontFace=cv2.FONT_HERSHEY_SIMPLEX, fontScale=1, color=textColor, thickness=2) + + cv2.imshow('FaceRecognizer', cameraFrame) + if cv2.waitKey(1) & 0xFF == ord('q'): + break + except: + cv2.imshow('FaceRecognize', cameraFrame) + continue
d2c83bf007a36a47754ec862d592c7c97b3145b9
plugins/callbacks/timer.py
plugins/callbacks/timer.py
import os import datetime from datetime import datetime, timedelta class CallbackModule(object): """ This callback module tells you how long your plays ran for. """ start_time = datetime.now() def __init__(self): start_time = datetime.now() print "Timer plugin is active." def days_hours_minutes_seconds(self, timedelta): minutes = (timedelta.seconds//60)%60 r_seconds = timedelta.seconds - (minutes * 60) return timedelta.days, timedelta.seconds//3600, minutes, r_seconds def playbook_on_stats(self, stats): end_time = datetime.now() timedelta = end_time - self.start_time print "Playbook run took %s days, %s hours, %s minutes, %s seconds" % (self.days_hours_minutes_seconds(timedelta))
Add simple plugin that times ansible-playbook runs.
Add simple plugin that times ansible-playbook runs.
Python
mit
thaim/ansible,thaim/ansible
--- +++ @@ -0,0 +1,27 @@ +import os +import datetime +from datetime import datetime, timedelta + + +class CallbackModule(object): + """ + This callback module tells you how long your plays ran for. + """ + + start_time = datetime.now() + + def __init__(self): + start_time = datetime.now() + print "Timer plugin is active." + + def days_hours_minutes_seconds(self, timedelta): + minutes = (timedelta.seconds//60)%60 + r_seconds = timedelta.seconds - (minutes * 60) + return timedelta.days, timedelta.seconds//3600, minutes, r_seconds + + def playbook_on_stats(self, stats): + end_time = datetime.now() + timedelta = end_time - self.start_time + print "Playbook run took %s days, %s hours, %s minutes, %s seconds" % (self.days_hours_minutes_seconds(timedelta)) + +
f4e0254eada3a6dd3aaa794926c5cc82e993b180
setup/bin/swc-nano-installer.py
setup/bin/swc-nano-installer.py
#!/usr/bin/env python """Software Carpentry Nano Installer for Windows Installs nano and makes it the default editor in msysgit To use: 1. Install Python 2. Install msysgit http://code.google.com/p/msysgit/downloads/list?q=full+installer+official+git 3. Run swc_nano_installer.py You should be able to simply double click the file in Windows This is a stripped down version of swc_windows_installer.py originally written by Ethan White and W. Trevor Price. """ try: # Python 3 from io import BytesIO as _BytesIO except ImportError: # Python 2 from StringIO import StringIO as _BytesIO import os try: # Python 3 from urllib.request import urlopen as _urlopen except ImportError: # Python 2 from urllib2 import urlopen as _urlopen import zipfile def install_nano(install_dir): """Download and install the nano text editor""" url = "http://www.nano-editor.org/dist/v2.2/NT/nano-2.2.6.zip" r = _urlopen(url) nano_zip_content = _BytesIO(r.read()) nano_zip = zipfile.ZipFile(nano_zip_content) nano_files = ['nano.exe', 'cygwin1.dll', 'cygintl-8.dll', 'cygiconv-2.dll', 'cyggcc_s-1.dll'] for file_name in nano_files: nano_zip.extract(file_name, install_dir) def make_bashrc(home_dir, nano_dir): """Creates a .bashrc file for nano setup Adds nano to the path and sets the default editor to nano """ nano_path = make_posix_path(nano_dir) contents = '\n'.join([ 'export PATH=$PATH:%s' % nano_path, 'export EDITOR=nano', '', ]) with open(os.path.join(home_dir, '.bashrc'), 'w') as f: f.write(contents) def make_posix_path(windows_path): """Convert a Windows path to a posix path""" return windows_path.replace('\\', '/').replace('C:', '/c') def main(): home_dir = os.path.expanduser("~") nano_dir = os.path.join(home_dir, '.nano') #home_dir = "/home/ethan/swc-nano-test" if not os.path.exists(nano_dir): os.makedirs(nano_dir) install_nano(nano_dir) make_bashrc(home_dir, nano_dir) if __name__ == '__main__': main()
Add a Nano installer for Windows
Add a Nano installer for Windows 1. Downloads and installs Nano into the users home directory 2. Adds Nano to the path 3. Makes Nano the default editor
Python
bsd-2-clause
selimnairb/2014-02-25-swctest,selimnairb/2014-02-25-swctest,selimnairb/2014-02-25-swctest,selimnairb/2014-02-25-swctest,selimnairb/2014-02-25-swctest,selimnairb/2014-02-25-swctest
--- +++ @@ -0,0 +1,73 @@ +#!/usr/bin/env python + +"""Software Carpentry Nano Installer for Windows + +Installs nano and makes it the default editor in msysgit + +To use: + +1. Install Python +2. Install msysgit + http://code.google.com/p/msysgit/downloads/list?q=full+installer+official+git +3. Run swc_nano_installer.py + You should be able to simply double click the file in Windows + +This is a stripped down version of swc_windows_installer.py +originally written by Ethan White and W. Trevor Price. + +""" + +try: # Python 3 + from io import BytesIO as _BytesIO +except ImportError: # Python 2 + from StringIO import StringIO as _BytesIO +import os +try: # Python 3 + from urllib.request import urlopen as _urlopen +except ImportError: # Python 2 + from urllib2 import urlopen as _urlopen +import zipfile + + +def install_nano(install_dir): + """Download and install the nano text editor""" + url = "http://www.nano-editor.org/dist/v2.2/NT/nano-2.2.6.zip" + r = _urlopen(url) + nano_zip_content = _BytesIO(r.read()) + nano_zip = zipfile.ZipFile(nano_zip_content) + nano_files = ['nano.exe', 'cygwin1.dll', 'cygintl-8.dll', + 'cygiconv-2.dll', 'cyggcc_s-1.dll'] + for file_name in nano_files: + nano_zip.extract(file_name, install_dir) + +def make_bashrc(home_dir, nano_dir): + """Creates a .bashrc file for nano setup + + Adds nano to the path and sets the default editor to nano + + """ + + nano_path = make_posix_path(nano_dir) + contents = '\n'.join([ + 'export PATH=$PATH:%s' % nano_path, + 'export EDITOR=nano', + '', + ]) + with open(os.path.join(home_dir, '.bashrc'), 'w') as f: + f.write(contents) + +def make_posix_path(windows_path): + """Convert a Windows path to a posix path""" + return windows_path.replace('\\', '/').replace('C:', '/c') + +def main(): + home_dir = os.path.expanduser("~") + nano_dir = os.path.join(home_dir, '.nano') + #home_dir = "/home/ethan/swc-nano-test" + if not os.path.exists(nano_dir): + os.makedirs(nano_dir) + install_nano(nano_dir) + make_bashrc(home_dir, nano_dir) + +if __name__ == '__main__': + main()
910793402dccc02c853618f93917c5f42c42a3da
GatewayMonitor/GatewayMonitorService.py
GatewayMonitor/GatewayMonitorService.py
# # A sample service to be 'compiled' into an exe-file with py2exe. # # See also # setup.py - the distutils' setup script # setup.cfg - the distutils' config file for this # README.txt - detailed usage notes # # A minimal service, doing nothing else than # - write 'start' and 'stop' entries into the NT event log # - when started, waits to be stopped again. # import win32serviceutil import win32service import win32event import win32evtlogutil import gatewaymonitor.GatewayMonitor class GatewayMonitorService(win32serviceutil.ServiceFramework): _svc_name_ = "GatewayMonitorService" _svc_display_name_ = "AMMO Gateway Monitor Service" _svc_deps_ = ["EventLog"] def __init__(self, args): win32serviceutil.ServiceFramework.__init__(self, args) self.hWaitStop = win32event.CreateEvent(None, 0, 0, None) def SvcStop(self): self.ReportServiceStatus(win32service.SERVICE_STOP_PENDING) win32event.SetEvent(self.hWaitStop) def SvcDoRun(self): import servicemanager mon = gatewaymonitor.GatewayMonitor.GatewayMonitor() mon.start() # Write a 'started' event to the event log... win32evtlogutil.ReportEvent(self._svc_name_, servicemanager.PYS_SERVICE_STARTED, 0, # category servicemanager.EVENTLOG_INFORMATION_TYPE, (self._svc_name_, '')) # wait for being stopped... win32event.WaitForSingleObject(self.hWaitStop, win32event.INFINITE) mon.cancel() mon.join() # and write a 'stopped' event to the event log. win32evtlogutil.ReportEvent(self._svc_name_, servicemanager.PYS_SERVICE_STOPPED, 0, # category servicemanager.EVENTLOG_INFORMATION_TYPE, (self._svc_name_, '')) if __name__ == '__main__': # Note that this code will not be run in the 'frozen' exe-file!!! win32serviceutil.HandleCommandLine(GatewayMonitorService)
Add gateway monitor Windows service
Add gateway monitor Windows service
Python
mit
isis-ammo/ammo-gateway,isis-ammo/ammo-gateway,isis-ammo/ammo-gateway,isis-ammo/ammo-gateway,isis-ammo/ammo-gateway,isis-ammo/ammo-gateway
--- +++ @@ -0,0 +1,60 @@ +# +# A sample service to be 'compiled' into an exe-file with py2exe. +# +# See also +# setup.py - the distutils' setup script +# setup.cfg - the distutils' config file for this +# README.txt - detailed usage notes +# +# A minimal service, doing nothing else than +# - write 'start' and 'stop' entries into the NT event log +# - when started, waits to be stopped again. +# +import win32serviceutil +import win32service +import win32event +import win32evtlogutil +import gatewaymonitor.GatewayMonitor + +class GatewayMonitorService(win32serviceutil.ServiceFramework): + _svc_name_ = "GatewayMonitorService" + _svc_display_name_ = "AMMO Gateway Monitor Service" + _svc_deps_ = ["EventLog"] + def __init__(self, args): + win32serviceutil.ServiceFramework.__init__(self, args) + self.hWaitStop = win32event.CreateEvent(None, 0, 0, None) + + def SvcStop(self): + self.ReportServiceStatus(win32service.SERVICE_STOP_PENDING) + win32event.SetEvent(self.hWaitStop) + + def SvcDoRun(self): + import servicemanager + + mon = gatewaymonitor.GatewayMonitor.GatewayMonitor() + mon.start() + + # Write a 'started' event to the event log... + win32evtlogutil.ReportEvent(self._svc_name_, + servicemanager.PYS_SERVICE_STARTED, + 0, # category + servicemanager.EVENTLOG_INFORMATION_TYPE, + (self._svc_name_, '')) + + # wait for being stopped... + win32event.WaitForSingleObject(self.hWaitStop, win32event.INFINITE) + + mon.cancel() + mon.join() + + # and write a 'stopped' event to the event log. + win32evtlogutil.ReportEvent(self._svc_name_, + servicemanager.PYS_SERVICE_STOPPED, + 0, # category + servicemanager.EVENTLOG_INFORMATION_TYPE, + (self._svc_name_, '')) + + +if __name__ == '__main__': + # Note that this code will not be run in the 'frozen' exe-file!!! + win32serviceutil.HandleCommandLine(GatewayMonitorService)
c73b8a7503f21e16171ea1b0b40180bd1624f4d3
social/apps/flask_app/routes.py
social/apps/flask_app/routes.py
from flask import g, Blueprint from flask.ext.login import login_required, login_user from social.actions import do_auth, do_complete, do_disconnect from social.apps.flask_app.utils import strategy social_auth = Blueprint('social', __name__) @social_auth.route('/login/<string:backend>/', methods=('GET', 'POST')) @strategy('social.complete') def auth(backend): return do_auth(g.strategy) @social_auth.route('/complete/<string:backend>/', methods=('GET', 'POST')) @strategy('social.complete') def complete(backend, *args, **kwargs): """Authentication complete view, override this view if transaction management doesn't suit your needs.""" return do_complete(g.strategy, login=lambda strat, user: login_user(user), user=g.user, *args, **kwargs) @social_auth.route('/disconnect/<string:backend>/', methods=('POST',)) @social_auth.route('/disconnect/<string:backend>/<int:association_id>/', methods=('POST',)) @login_required @strategy() def disconnect(backend, association_id=None): """Disconnects given backend from current logged in user.""" return do_disconnect(g.strategy, g.user, association_id)
from flask import g, Blueprint, request from flask.ext.login import login_required, login_user from social.actions import do_auth, do_complete, do_disconnect from social.apps.flask_app.utils import strategy social_auth = Blueprint('social', __name__) @social_auth.route('/login/<string:backend>/', methods=('GET', 'POST')) @strategy('social.complete') def auth(backend): return do_auth(g.strategy) @social_auth.route('/complete/<string:backend>/', methods=('GET', 'POST')) @strategy('social.complete') def complete(backend, *args, **kwargs): """Authentication complete view, override this view if transaction management doesn't suit your needs.""" return do_complete(g.strategy, login=do_login, user=g.user, *args, **kwargs) @social_auth.route('/disconnect/<string:backend>/', methods=('POST',)) @social_auth.route('/disconnect/<string:backend>/<int:association_id>/', methods=('POST',)) @login_required @strategy() def disconnect(backend, association_id=None): """Disconnects given backend from current logged in user.""" return do_disconnect(g.strategy, g.user, association_id) def do_login(strategy, user): return login_user(user, remember=request.cookies.get('remember') or request.args.get('remember') or request.form.get('remember') or False)
Support remember flag when calling login on flask app
Support remember flag when calling login on flask app
Python
bsd-3-clause
lamby/python-social-auth,lneoe/python-social-auth,henocdz/python-social-auth,ononeor12/python-social-auth,cjltsod/python-social-auth,JJediny/python-social-auth,henocdz/python-social-auth,henocdz/python-social-auth,ariestiyansyah/python-social-auth,mathspace/python-social-auth,rsteca/python-social-auth,jneves/python-social-auth,tutumcloud/python-social-auth,alrusdi/python-social-auth,mrwags/python-social-auth,tkajtoch/python-social-auth,barseghyanartur/python-social-auth,san-mate/python-social-auth,yprez/python-social-auth,S01780/python-social-auth,python-social-auth/social-core,frankier/python-social-auth,ByteInternet/python-social-auth,jeyraof/python-social-auth,joelstanner/python-social-auth,rsalmaso/python-social-auth,webjunkie/python-social-auth,bjorand/python-social-auth,tutumcloud/python-social-auth,jameslittle/python-social-auth,Andygmb/python-social-auth,S01780/python-social-auth,iruga090/python-social-auth,lawrence34/python-social-auth,robbiet480/python-social-auth,degs098/python-social-auth,cjltsod/python-social-auth,DhiaEddineSaidi/python-social-auth,MSOpenTech/python-social-auth,merutak/python-social-auth,mchdks/python-social-auth,tkajtoch/python-social-auth,rsalmaso/python-social-auth,hsr-ba-fs15-dat/python-social-auth,degs098/python-social-auth,ByteInternet/python-social-auth,jeyraof/python-social-auth,degs098/python-social-auth,hsr-ba-fs15-dat/python-social-auth,frankier/python-social-auth,yprez/python-social-auth,SeanHayes/python-social-auth,alrusdi/python-social-auth,fearlessspider/python-social-auth,drxos/python-social-auth,joelstanner/python-social-auth,python-social-auth/social-app-django,JJediny/python-social-auth,jameslittle/python-social-auth,contracode/python-social-auth,wildtetris/python-social-auth,michael-borisov/python-social-auth,hsr-ba-fs15-dat/python-social-auth,imsparsh/python-social-auth,webjunkie/python-social-auth,contracode/python-social-auth,nirmalvp/python-social-auth,chandolia/python-social-auth,duoduo369/python-social-auth,falcon1kr/python-social-auth,SeanHayes/python-social-auth,clef/python-social-auth,mrwags/python-social-auth,imsparsh/python-social-auth,joelstanner/python-social-auth,falcon1kr/python-social-auth,mathspace/python-social-auth,robbiet480/python-social-auth,ariestiyansyah/python-social-auth,iruga090/python-social-auth,mark-adams/python-social-auth,chandolia/python-social-auth,mrwags/python-social-auth,cmichal/python-social-auth,lneoe/python-social-auth,python-social-auth/social-app-django,garrett-schlesinger/python-social-auth,ByteInternet/python-social-auth,JJediny/python-social-auth,JerzySpendel/python-social-auth,cmichal/python-social-auth,drxos/python-social-auth,yprez/python-social-auth,fearlessspider/python-social-auth,contracode/python-social-auth,jneves/python-social-auth,MSOpenTech/python-social-auth,nirmalvp/python-social-auth,tkajtoch/python-social-auth,python-social-auth/social-app-cherrypy,mark-adams/python-social-auth,mathspace/python-social-auth,muhammad-ammar/python-social-auth,barseghyanartur/python-social-auth,lawrence34/python-social-auth,S01780/python-social-auth,msampathkumar/python-social-auth,Andygmb/python-social-auth,python-social-auth/social-app-django,noodle-learns-programming/python-social-auth,mark-adams/python-social-auth,merutak/python-social-auth,san-mate/python-social-auth,daniula/python-social-auth,bjorand/python-social-auth,ononeor12/python-social-auth,daniula/python-social-auth,clef/python-social-auth,drxos/python-social-auth,robbiet480/python-social-auth,jeyraof/python-social-auth,falcon1kr/python-social-auth,lamby/python-social-auth,MSOpenTech/python-social-auth,firstjob/python-social-auth,VishvajitP/python-social-auth,VishvajitP/python-social-auth,chandolia/python-social-auth,rsteca/python-social-auth,Andygmb/python-social-auth,VishvajitP/python-social-auth,python-social-auth/social-core,JerzySpendel/python-social-auth,barseghyanartur/python-social-auth,webjunkie/python-social-auth,DhiaEddineSaidi/python-social-auth,wildtetris/python-social-auth,lamby/python-social-auth,firstjob/python-social-auth,msampathkumar/python-social-auth,clef/python-social-auth,lneoe/python-social-auth,michael-borisov/python-social-auth,muhammad-ammar/python-social-auth,jameslittle/python-social-auth,bjorand/python-social-auth,daniula/python-social-auth,firstjob/python-social-auth,tobias47n9e/social-core,rsteca/python-social-auth,lawrence34/python-social-auth,michael-borisov/python-social-auth,DhiaEddineSaidi/python-social-auth,nirmalvp/python-social-auth,cmichal/python-social-auth,duoduo369/python-social-auth,noodle-learns-programming/python-social-auth,merutak/python-social-auth,iruga090/python-social-auth,ononeor12/python-social-auth,msampathkumar/python-social-auth,noodle-learns-programming/python-social-auth,python-social-auth/social-storage-sqlalchemy,python-social-auth/social-docs,fearlessspider/python-social-auth,mchdks/python-social-auth,mchdks/python-social-auth,ariestiyansyah/python-social-auth,jneves/python-social-auth,alrusdi/python-social-auth,garrett-schlesinger/python-social-auth,muhammad-ammar/python-social-auth,imsparsh/python-social-auth,san-mate/python-social-auth,wildtetris/python-social-auth,JerzySpendel/python-social-auth
--- +++ @@ -1,4 +1,4 @@ -from flask import g, Blueprint +from flask import g, Blueprint, request from flask.ext.login import login_required, login_user from social.actions import do_auth, do_complete, do_disconnect @@ -19,8 +19,8 @@ def complete(backend, *args, **kwargs): """Authentication complete view, override this view if transaction management doesn't suit your needs.""" - return do_complete(g.strategy, login=lambda strat, user: login_user(user), - user=g.user, *args, **kwargs) + return do_complete(g.strategy, login=do_login, user=g.user, + *args, **kwargs) @social_auth.route('/disconnect/<string:backend>/', methods=('POST',)) @@ -31,3 +31,9 @@ def disconnect(backend, association_id=None): """Disconnects given backend from current logged in user.""" return do_disconnect(g.strategy, g.user, association_id) + + +def do_login(strategy, user): + return login_user(user, remember=request.cookies.get('remember') or + request.args.get('remember') or + request.form.get('remember') or False)
50b2dccd24c756f5f09e6b9f4e2442abeed7f5c7
st2api/tests/unit/controllers/v1/test_alias_execution_rbac.py
st2api/tests/unit/controllers/v1/test_alias_execution_rbac.py
# Licensed to the StackStorm, Inc ('StackStorm') under one or more # contributor license agreements. See the NOTICE file distributed with # this work for additional information regarding copyright ownership. # The ASF licenses this file to You under the Apache License, Version 2.0 # (the "License"); you may not use this file except in compliance with # the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import mock from st2common.services import action as action_service from st2tests.fixturesloader import FixturesLoader from tests.base import APIControllerWithRBACTestCase from tests.unit.controllers.v1.test_alias_execution import DummyActionExecution FIXTURES_PACK = 'aliases' TEST_MODELS = { 'aliases': ['alias1.yaml', 'alias2.yaml'], 'actions': ['action1.yaml'], 'runners': ['runner1.yaml'] } TEST_LOAD_MODELS = { 'aliases': ['alias3.yaml'] } __all__ = [ 'AliasExecutionWithRBACTestCase' ] class AliasExecutionWithRBACTestCase(APIControllerWithRBACTestCase): def setUp(self): super(AliasExecutionWithRBACTestCase, self).setUp() self.models = FixturesLoader().save_fixtures_to_db(fixtures_pack=FIXTURES_PACK, fixtures_dict=TEST_MODELS) self.alias1 = self.models['aliases']['alias1.yaml'] self.alias2 = self.models['aliases']['alias2.yaml'] @mock.patch.object(action_service, 'request', return_value=(None, DummyActionExecution(id_=1))) def test_live_action_context_user_is_set_to_authenticated_user(self, request): # Verify that the user inside the context of live action is set to authenticated user # which hit the endpoint. This is important for RBAC and many other things. user_db = self.users['admin'] self.use_user(user_db) command = 'Lorem ipsum value1 dolor sit "value2, value3" amet.' post_resp = self._do_post(alias_execution=self.alias2, command=command) self.assertEqual(post_resp.status_int, 200) live_action_db = request.call_args[0][0] self.assertEquals(live_action_db.context['user'], 'admin') def _do_post(self, alias_execution, command, expect_errors=False): execution = {'name': alias_execution.name, 'format': alias_execution.formats[0], 'command': command, 'user': 'stanley', 'source_channel': 'test', 'notification_route': 'test'} return self.app.post_json('/v1/aliasexecution', execution, expect_errors=expect_errors)
Add a regression test case for alias execution and live action context user.
Add a regression test case for alias execution and live action context user. Verify that the user inside the context is correctly set to the authenticated user which triggered the alias execution.
Python
apache-2.0
Plexxi/st2,StackStorm/st2,lakshmi-kannan/st2,nzlosh/st2,pixelrebel/st2,dennybaa/st2,pixelrebel/st2,pixelrebel/st2,peak6/st2,StackStorm/st2,peak6/st2,punalpatel/st2,dennybaa/st2,lakshmi-kannan/st2,armab/st2,emedvedev/st2,tonybaloney/st2,dennybaa/st2,tonybaloney/st2,armab/st2,lakshmi-kannan/st2,punalpatel/st2,peak6/st2,Plexxi/st2,nzlosh/st2,armab/st2,nzlosh/st2,emedvedev/st2,Plexxi/st2,StackStorm/st2,punalpatel/st2,nzlosh/st2,emedvedev/st2,Plexxi/st2,tonybaloney/st2,StackStorm/st2
--- +++ @@ -0,0 +1,73 @@ +# Licensed to the StackStorm, Inc ('StackStorm') under one or more +# contributor license agreements. See the NOTICE file distributed with +# this work for additional information regarding copyright ownership. +# The ASF licenses this file to You under the Apache License, Version 2.0 +# (the "License"); you may not use this file except in compliance with +# the License. You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +import mock + +from st2common.services import action as action_service +from st2tests.fixturesloader import FixturesLoader +from tests.base import APIControllerWithRBACTestCase +from tests.unit.controllers.v1.test_alias_execution import DummyActionExecution + +FIXTURES_PACK = 'aliases' + +TEST_MODELS = { + 'aliases': ['alias1.yaml', 'alias2.yaml'], + 'actions': ['action1.yaml'], + 'runners': ['runner1.yaml'] +} + +TEST_LOAD_MODELS = { + 'aliases': ['alias3.yaml'] +} + +__all__ = [ + 'AliasExecutionWithRBACTestCase' +] + + +class AliasExecutionWithRBACTestCase(APIControllerWithRBACTestCase): + + def setUp(self): + super(AliasExecutionWithRBACTestCase, self).setUp() + + self.models = FixturesLoader().save_fixtures_to_db(fixtures_pack=FIXTURES_PACK, + fixtures_dict=TEST_MODELS) + self.alias1 = self.models['aliases']['alias1.yaml'] + self.alias2 = self.models['aliases']['alias2.yaml'] + + @mock.patch.object(action_service, 'request', + return_value=(None, DummyActionExecution(id_=1))) + def test_live_action_context_user_is_set_to_authenticated_user(self, request): + # Verify that the user inside the context of live action is set to authenticated user + # which hit the endpoint. This is important for RBAC and many other things. + user_db = self.users['admin'] + self.use_user(user_db) + + command = 'Lorem ipsum value1 dolor sit "value2, value3" amet.' + post_resp = self._do_post(alias_execution=self.alias2, command=command) + self.assertEqual(post_resp.status_int, 200) + + live_action_db = request.call_args[0][0] + self.assertEquals(live_action_db.context['user'], 'admin') + + def _do_post(self, alias_execution, command, expect_errors=False): + execution = {'name': alias_execution.name, + 'format': alias_execution.formats[0], + 'command': command, + 'user': 'stanley', + 'source_channel': 'test', + 'notification_route': 'test'} + return self.app.post_json('/v1/aliasexecution', execution, + expect_errors=expect_errors)
b1083460378166b23bfe379dfb228d4cf5236255
app/grandchallenge/profiles/utils.py
app/grandchallenge/profiles/utils.py
from django.conf import settings from grandchallenge.subdomains.utils import reverse def signin_redirect(redirect=None, user=None): """ Redirect user after successful sign in. First looks for a ``requested_redirect``. If not supplied will fall-back to the user specific account page. If all fails, will fall-back to redirect to the homepage. Returns a string defining the URI to go next. :param redirect: A value normally supplied by ``next`` form field. Gets preference before the default view which requires the user. :param user: A ``User`` object specifying the user who has just signed in. :return: String containing the URI to redirect to. """ if redirect and settings.LOGOUT_URL not in redirect: return redirect elif user is not None: return reverse("profile_redirect") else: return reverse("home")
Create method for redirect after signin
Create method for redirect after signin
Python
apache-2.0
comic/comic-django,comic/comic-django,comic/comic-django,comic/comic-django,comic/comic-django
--- +++ @@ -0,0 +1,24 @@ +from django.conf import settings + +from grandchallenge.subdomains.utils import reverse + + +def signin_redirect(redirect=None, user=None): + """ + Redirect user after successful sign in. + First looks for a ``requested_redirect``. If not supplied will fall-back to + the user specific account page. If all fails, will fall-back to redirect to + the homepage. Returns a string defining the URI to go next. + :param redirect: + A value normally supplied by ``next`` form field. Gets preference + before the default view which requires the user. + :param user: + A ``User`` object specifying the user who has just signed in. + :return: String containing the URI to redirect to. + """ + if redirect and settings.LOGOUT_URL not in redirect: + return redirect + elif user is not None: + return reverse("profile_redirect") + else: + return reverse("home")
64f820b18190034a2285e996cd2f66fc100c2f54
cogbot/extensions/jira.py
cogbot/extensions/jira.py
import re import urllib.parse from discord.ext import commands from discord.ext.commands import Context from cogbot.cog_bot import CogBot class Jira: REPORT_PATTERN = re.compile('^(mc-)?(\d+)$', re.IGNORECASE) def __init__(self, bot: CogBot, ext: str): self.bot = bot @commands.command(pass_context=True) async def jira(self, ctx: Context, *, query: str): rmatch = self.REPORT_PATTERN.match(query) if rmatch: rgroups = rmatch.groups() report_no = rgroups[1] url = 'https://bugs.mojang.com/browse/MC-' + report_no else: search_url = urllib.parse.urlencode({'searchString': query}) url = 'https://bugs.mojang.com/secure/QuickSearch.jspa?' + search_url await self.bot.say(url) await self.bot.react_success(ctx) def setup(bot): bot.add_cog(Jira(bot, __name__))
Add extension to search JIRA for bug reports
Add extension to search JIRA for bug reports
Python
mit
Arcensoth/cogbot
--- +++ @@ -0,0 +1,31 @@ +import re +import urllib.parse + +from discord.ext import commands +from discord.ext.commands import Context + +from cogbot.cog_bot import CogBot + + +class Jira: + REPORT_PATTERN = re.compile('^(mc-)?(\d+)$', re.IGNORECASE) + + def __init__(self, bot: CogBot, ext: str): + self.bot = bot + + @commands.command(pass_context=True) + async def jira(self, ctx: Context, *, query: str): + rmatch = self.REPORT_PATTERN.match(query) + if rmatch: + rgroups = rmatch.groups() + report_no = rgroups[1] + url = 'https://bugs.mojang.com/browse/MC-' + report_no + else: + search_url = urllib.parse.urlencode({'searchString': query}) + url = 'https://bugs.mojang.com/secure/QuickSearch.jspa?' + search_url + await self.bot.say(url) + await self.bot.react_success(ctx) + + +def setup(bot): + bot.add_cog(Jira(bot, __name__))
201e63722b102ee453610a955e3014d3c772a8fa
ceph_deploy/tests/parser/test_purge.py
ceph_deploy/tests/parser/test_purge.py
import pytest from ceph_deploy.cli import get_parser class TestParserPurge(object): def setup(self): self.parser = get_parser() def test_purge_help(self, capsys): with pytest.raises(SystemExit): self.parser.parse_args('purge --help'.split()) out, err = capsys.readouterr() assert 'usage: ceph-deploy purge' in out assert 'positional arguments:' in out assert 'optional arguments:' in out def test_purge_host_required(self, capsys): with pytest.raises(SystemExit): self.parser.parse_args('purge'.split()) out, err = capsys.readouterr() assert "error: too few arguments" in err def test_purge_one_host(self): args = self.parser.parse_args('purge host1'.split()) assert args.host == ['host1'] def test_purge_multiple_hosts(self): hostnames = ['host1', 'host2', 'host3'] args = self.parser.parse_args(['purge'] + hostnames) assert frozenset(args.host) == frozenset(hostnames)
Add tests for argparse purge
[RM-11742] Add tests for argparse purge Signed-off-by: Travis Rhoden <e5e44d6dbac12e32e01c3bb8b67940d8b42e225b@redhat.com>
Python
mit
osynge/ceph-deploy,codenrhoden/ceph-deploy,Vicente-Cheng/ceph-deploy,SUSE/ceph-deploy,branto1/ceph-deploy,imzhulei/ceph-deploy,SUSE/ceph-deploy-to-be-deleted,ceph/ceph-deploy,SUSE/ceph-deploy-to-be-deleted,Vicente-Cheng/ceph-deploy,isyippee/ceph-deploy,zhouyuan/ceph-deploy,ceph/ceph-deploy,zhouyuan/ceph-deploy,shenhequnying/ceph-deploy,isyippee/ceph-deploy,codenrhoden/ceph-deploy,imzhulei/ceph-deploy,SUSE/ceph-deploy,shenhequnying/ceph-deploy,branto1/ceph-deploy,ghxandsky/ceph-deploy,ghxandsky/ceph-deploy,trhoden/ceph-deploy,trhoden/ceph-deploy,osynge/ceph-deploy
--- +++ @@ -0,0 +1,32 @@ +import pytest + +from ceph_deploy.cli import get_parser + + +class TestParserPurge(object): + + def setup(self): + self.parser = get_parser() + + def test_purge_help(self, capsys): + with pytest.raises(SystemExit): + self.parser.parse_args('purge --help'.split()) + out, err = capsys.readouterr() + assert 'usage: ceph-deploy purge' in out + assert 'positional arguments:' in out + assert 'optional arguments:' in out + + def test_purge_host_required(self, capsys): + with pytest.raises(SystemExit): + self.parser.parse_args('purge'.split()) + out, err = capsys.readouterr() + assert "error: too few arguments" in err + + def test_purge_one_host(self): + args = self.parser.parse_args('purge host1'.split()) + assert args.host == ['host1'] + + def test_purge_multiple_hosts(self): + hostnames = ['host1', 'host2', 'host3'] + args = self.parser.parse_args(['purge'] + hostnames) + assert frozenset(args.host) == frozenset(hostnames)
fbbd0a3f55c1e79ebf6ae7b872697611740edb24
foobar.py
foobar.py
# -*- coding: utf-8 -*- class TreeElement(object): def __repr__(self): return u'{}'.format(self.__class__.__name__) class Token(TreeElement): def __init__(self, value): if isinstance(value, TreeElement): raise TypeError self._value = value def __repr__(self): return u'{}({})'.format(self.__class__.__name__, self._value) def __gt__(self, other): if not isinstance(other, self.__class__): return hash(self) > hash(other) return self._value > other._value def __ge__(self, other): if not isinstance(other, self.__class__): return hash(self) >= hash(other) return self._value >= other._value def __eq__(self, other): if not isinstance(other, self.__class__): return False return self._value == other._value def __hash__(self): return hash((self.__class__.__name__, self._value)) def tokens(self): yield self._value class Not(TreeElement): def __init__(self, child): if not isinstance(child, TreeElement): raise TypeError self._child = child def __repr__(self): return u'{}({})'.format(self.__class__.__name__, self._child) def __gt__(self, other): if not isinstance(other, self.__class__): return hash(self) > hash(other) return self._child > other._child def __ge__(self, other): if not isinstance(other, self.__class__): return hash(self) >= hash(other) return self._child >= other._child def __eq__(self, other): if not isinstance(other, self.__class__): return False return self._child == other._child def __hash__(self): return hash((self.__class__.__name__, self._child)) def tokens(self): for token in self._child.tokens(): yield token class GroupMixin(object): def __init__(self, a, b, *others): children = (a, b) + others for child in children: if not isinstance(child, TreeElement): raise TypeError self._children = tuple(sorted(children)) def __repr__(self): return u'{}({})'.format( self.__class__.__name__, u', '.join(repr(child) for child in self._children) ) def __gt__(self, other): if not isinstance(other, self.__class__): return hash(self) > hash(other) return self._children > other._children def __ge__(self, other): if not isinstance(other, self.__class__): return hash(self) >= hash(other) return self._children >= other._children def __eq__(self, other): if not isinstance(other, self.__class__): return False return self._children == other._children def __hash__(self): return hash(tuple([self.__class__.__name__]) + self._children) def tokens(self): for child in self._children: for token in child.tokens(): yield token class And(GroupMixin, TreeElement): pass class Or(GroupMixin, TreeElement): pass
Implement Token and Not, And, Or classes
Implement Token and Not, And, Or classes
Python
mit
hackebrot/i-am-bool
--- +++ @@ -0,0 +1,116 @@ +# -*- coding: utf-8 -*- + + +class TreeElement(object): + def __repr__(self): + return u'{}'.format(self.__class__.__name__) + + +class Token(TreeElement): + def __init__(self, value): + if isinstance(value, TreeElement): + raise TypeError + self._value = value + + def __repr__(self): + return u'{}({})'.format(self.__class__.__name__, self._value) + + def __gt__(self, other): + if not isinstance(other, self.__class__): + return hash(self) > hash(other) + return self._value > other._value + + def __ge__(self, other): + if not isinstance(other, self.__class__): + return hash(self) >= hash(other) + return self._value >= other._value + + def __eq__(self, other): + if not isinstance(other, self.__class__): + return False + return self._value == other._value + + def __hash__(self): + return hash((self.__class__.__name__, self._value)) + + def tokens(self): + yield self._value + + +class Not(TreeElement): + def __init__(self, child): + if not isinstance(child, TreeElement): + raise TypeError + self._child = child + + def __repr__(self): + return u'{}({})'.format(self.__class__.__name__, self._child) + + def __gt__(self, other): + if not isinstance(other, self.__class__): + return hash(self) > hash(other) + return self._child > other._child + + def __ge__(self, other): + if not isinstance(other, self.__class__): + return hash(self) >= hash(other) + return self._child >= other._child + + def __eq__(self, other): + if not isinstance(other, self.__class__): + return False + return self._child == other._child + + def __hash__(self): + return hash((self.__class__.__name__, self._child)) + + def tokens(self): + for token in self._child.tokens(): + yield token + + +class GroupMixin(object): + def __init__(self, a, b, *others): + children = (a, b) + others + + for child in children: + if not isinstance(child, TreeElement): + raise TypeError + self._children = tuple(sorted(children)) + + def __repr__(self): + return u'{}({})'.format( + self.__class__.__name__, + u', '.join(repr(child) for child in self._children) + ) + + def __gt__(self, other): + if not isinstance(other, self.__class__): + return hash(self) > hash(other) + return self._children > other._children + + def __ge__(self, other): + if not isinstance(other, self.__class__): + return hash(self) >= hash(other) + return self._children >= other._children + + def __eq__(self, other): + if not isinstance(other, self.__class__): + return False + return self._children == other._children + + def __hash__(self): + return hash(tuple([self.__class__.__name__]) + self._children) + + def tokens(self): + for child in self._children: + for token in child.tokens(): + yield token + + +class And(GroupMixin, TreeElement): + pass + + +class Or(GroupMixin, TreeElement): + pass
cf4dc3f65049c23f4a20140cf5092f3c5a771a6a
administrator/tests/models/test_user.py
administrator/tests/models/test_user.py
from django.test import TestCase from administrator.models import User, UserManager class UserManagerTestCase(TestCase): valid_email = 'example@example.com' valid_username = 'foobar' valid_password = 'qwerty123' def setUp(self): self.user_manager = UserManager() self.user_manager.model = User def test_user_is_not_created_without_email(self): with self.assertRaises(ValueError): self.user_manager.create_user('', '', '') def test_user_is_created(self): user = self.user_manager.create_user(self.valid_email, self.valid_username, self.valid_password) self.assertEqual(user.is_superuser, False) def test_superuser_is_not_created_with_is_superuser_field_set_to_false(self): with self.assertRaises(ValueError): self.user_manager.create_superuser(self.valid_email, self.valid_username, self.valid_password, is_superuser=False) def test_superuser_is_created(self): superuser = self.user_manager.create_user(self.valid_email, self.valid_username, self.valid_password) self.assertEqual(superuser.is_superuser, True) class UserTestCase(TestCase): valid_email = 'example@example.com' def setUp(self): self.user = User() self.user.email = self.valid_email def test_user_status_can_be_set_to_active(self): self.user.set_status_and_is_active(0) self.assertEqual(self.user.status, 0) self.assertEqual(self.user.is_active, True) def test_user_status_can_be_set_to_inactive(self): self.user.set_status_and_is_active(1) self.assertEqual(self.user.status, 1) self.assertEqual(self.user.is_active, False) def test_user_has_email_as_his_short_name(self): short_name = self.user.get_short_name() self.assertEqual(short_name, self.user.email)
Add tests for User model
Add tests for User model
Python
mit
Social-projects-Rivne/Rv-025.Python,Social-projects-Rivne/Rv-025.Python,Social-projects-Rivne/Rv-025.Python
--- +++ @@ -0,0 +1,52 @@ +from django.test import TestCase + +from administrator.models import User, UserManager + + +class UserManagerTestCase(TestCase): + valid_email = 'example@example.com' + valid_username = 'foobar' + valid_password = 'qwerty123' + + def setUp(self): + self.user_manager = UserManager() + self.user_manager.model = User + + def test_user_is_not_created_without_email(self): + with self.assertRaises(ValueError): + self.user_manager.create_user('', '', '') + + def test_user_is_created(self): + user = self.user_manager.create_user(self.valid_email, self.valid_username, self.valid_password) + self.assertEqual(user.is_superuser, False) + + def test_superuser_is_not_created_with_is_superuser_field_set_to_false(self): + with self.assertRaises(ValueError): + self.user_manager.create_superuser(self.valid_email, self.valid_username, self.valid_password, + is_superuser=False) + + def test_superuser_is_created(self): + superuser = self.user_manager.create_user(self.valid_email, self.valid_username, self.valid_password) + self.assertEqual(superuser.is_superuser, True) + + +class UserTestCase(TestCase): + valid_email = 'example@example.com' + + def setUp(self): + self.user = User() + self.user.email = self.valid_email + + def test_user_status_can_be_set_to_active(self): + self.user.set_status_and_is_active(0) + self.assertEqual(self.user.status, 0) + self.assertEqual(self.user.is_active, True) + + def test_user_status_can_be_set_to_inactive(self): + self.user.set_status_and_is_active(1) + self.assertEqual(self.user.status, 1) + self.assertEqual(self.user.is_active, False) + + def test_user_has_email_as_his_short_name(self): + short_name = self.user.get_short_name() + self.assertEqual(short_name, self.user.email)
b9a8be97bc445edac024a35d644d5c86419a7f3a
ReactAndroid/src/main/third-party/android-support-for-standalone-apps/v7/appcompat/res-unpacker.py
ReactAndroid/src/main/third-party/android-support-for-standalone-apps/v7/appcompat/res-unpacker.py
import contextlib import os import shutil import sys import tempfile import zipfile # Helper that unpacks the contents of the res folder of an .aar file # into given destination. @contextlib.contextmanager def cleanup(path): yield path shutil.rmtree(path) if __name__ == '__main__': with zipfile.ZipFile(sys.argv[1], 'r') as z: with cleanup(tempfile.mkdtemp()) as temp_path: z.extractall(temp_path, filter(lambda n: n.startswith('res/'), z.namelist())) shutil.move(os.path.join(temp_path, 'res'), sys.argv[2])
Build code depending on resources from the appcompat library with both Buck and Gradle
Build code depending on resources from the appcompat library with both Buck and Gradle Reviewed By: sdwilsh Differential Revision: D2844961 fb-gh-sync-id: 686a9f253eb370a9dc8cc33ca1c4e8453f89210a
Python
bsd-3-clause
hammerandchisel/react-native,jaggs6/react-native,ultralame/react-native,alin23/react-native,yamill/react-native,doochik/react-native,iodine/react-native,Swaagie/react-native,facebook/react-native,exponentjs/react-native,corbt/react-native,makadaw/react-native,arthuralee/react-native,jevakallio/react-native,forcedotcom/react-native,sghiassy/react-native,peterp/react-native,hammerandchisel/react-native,negativetwelve/react-native,gre/react-native,mrspeaker/react-native,a2/react-native,charlesvinette/react-native,aljs/react-native,foghina/react-native,Swaagie/react-native,clozr/react-native,mrspeaker/react-native,hoangpham95/react-native,nathanajah/react-native,chirag04/react-native,shrutic/react-native,xiayz/react-native,miracle2k/react-native,dabit3/react-native,jasonnoahchoi/react-native,ultralame/react-native,farazs/react-native,shinate/react-native,ultralame/react-native,chnfeeeeeef/react-native,aljs/react-native,yamill/react-native,apprennet/react-native,dabit3/react-native,ndejesus1227/react-native,CntChen/react-native,skatpgusskat/react-native,jasonnoahchoi/react-native,tsjing/react-native,BretJohnson/react-native,CntChen/react-native,salanki/react-native,jasonnoahchoi/react-native,urvashi01/react-native,cdlewis/react-native,tadeuzagallo/react-native,formatlos/react-native,formatlos/react-native,BretJohnson/react-native,Bhullnatik/react-native,brentvatne/react-native,exponent/react-native,makadaw/react-native,aljs/react-native,esauter5/react-native,exponent/react-native,compulim/react-native,mrspeaker/react-native,Maxwell2022/react-native,ptomasroos/react-native,compulim/react-native,exponent/react-native,PlexChat/react-native,negativetwelve/react-native,imjerrybao/react-native,hoastoolshop/react-native,hayeah/react-native,imjerrybao/react-native,charlesvinette/react-native,brentvatne/react-native,naoufal/react-native,gilesvangruisen/react-native,callstack-io/react-native,christopherdro/react-native,kesha-antonov/react-native,imDangerous/react-native,CntChen/react-native,skatpgusskat/react-native,miracle2k/react-native,ptomasroos/react-native,ankitsinghania94/react-native,Maxwell2022/react-native,shrutic/react-native,doochik/react-native,hoastoolshop/react-native,alin23/react-native,tgoldenberg/react-native,Andreyco/react-native,chnfeeeeeef/react-native,jhen0409/react-native,skevy/react-native,Swaagie/react-native,formatlos/react-native,browniefed/react-native,Emilios1995/react-native,MattFoley/react-native,christopherdro/react-native,Guardiannw/react-native,thotegowda/react-native,jevakallio/react-native,mironiasty/react-native,clozr/react-native,Guardiannw/react-native,CodeLinkIO/react-native,cpunion/react-native,browniefed/react-native,facebook/react-native,forcedotcom/react-native,rebeccahughes/react-native,Purii/react-native,arbesfeld/react-native,htc2u/react-native,hammerandchisel/react-native,Guardiannw/react-native,Andreyco/react-native,ankitsinghania94/react-native,negativetwelve/react-native,corbt/react-native,mironiasty/react-native,imjerrybao/react-native,facebook/react-native,myntra/react-native,shrutic/react-native,happypancake/react-native,ankitsinghania94/react-native,tarkus/react-native-appletv,ptmt/react-native-macos,rebeccahughes/react-native,InterfaceInc/react-native,Ehesp/react-native,exponent/react-native,skevy/react-native,a2/react-native,iodine/react-native,dubert/react-native,adamjmcgrath/react-native,Swaagie/react-native,peterp/react-native,apprennet/react-native,naoufal/react-native,tgoldenberg/react-native,farazs/react-native,gitim/react-native,lelandrichardson/react-native,satya164/react-native,jadbox/react-native,rebeccahughes/react-native,thotegowda/react-native,myntra/react-native,DerayGa/react-native,shrutic123/react-native,hoastoolshop/react-native,jaggs6/react-native,Purii/react-native,sghiassy/react-native,philikon/react-native,nathanajah/react-native,tszajna0/react-native,formatlos/react-native,imDangerous/react-native,Guardiannw/react-native,janicduplessis/react-native,aaron-goshine/react-native,gre/react-native,martinbigio/react-native,ndejesus1227/react-native,rickbeerendonk/react-native,tgoldenberg/react-native,Livyli/react-native,skevy/react-native,jadbox/react-native,chnfeeeeeef/react-native,CntChen/react-native,gre/react-native,kesha-antonov/react-native,Guardiannw/react-native,arthuralee/react-native,callstack-io/react-native,wesley1001/react-native,shrutic/react-native,arthuralee/react-native,aljs/react-native,gitim/react-native,doochik/react-native,chnfeeeeeef/react-native,CodeLinkIO/react-native,a2/react-native,corbt/react-native,ptmt/react-native-macos,jhen0409/react-native,esauter5/react-native,a2/react-native,doochik/react-native,Andreyco/react-native,ptmt/react-native-macos,jadbox/react-native,iodine/react-native,Purii/react-native,charlesvinette/react-native,CodeLinkIO/react-native,shrutic123/react-native,DannyvanderJagt/react-native,dikaiosune/react-native,jadbox/react-native,eduardinni/react-native,mrspeaker/react-native,dubert/react-native,jevakallio/react-native,ankitsinghania94/react-native,yamill/react-native,hoangpham95/react-native,skatpgusskat/react-native,clozr/react-native,apprennet/react-native,luqin/react-native,gre/react-native,tsjing/react-native,thotegowda/react-native,tgoldenberg/react-native,skatpgusskat/react-native,javache/react-native,shrutic/react-native,hayeah/react-native,kesha-antonov/react-native,thotegowda/react-native,ptomasroos/react-native,eduardinni/react-native,tarkus/react-native-appletv,pglotov/react-native,yamill/react-native,ptomasroos/react-native,machard/react-native,miracle2k/react-native,eduardinni/react-native,InterfaceInc/react-native,shinate/react-native,esauter5/react-native,shrutic123/react-native,happypancake/react-native,corbt/react-native,thotegowda/react-native,DanielMSchmidt/react-native,DanielMSchmidt/react-native,shinate/react-native,Livyli/react-native,chirag04/react-native,skevy/react-native,rebeccahughes/react-native,janicduplessis/react-native,peterp/react-native,chirag04/react-native,dikaiosune/react-native,jeffchienzabinet/react-native,machard/react-native,wesley1001/react-native,Livyli/react-native,ndejesus1227/react-native,InterfaceInc/react-native,jasonnoahchoi/react-native,adamjmcgrath/react-native,wenpkpk/react-native,tadeuzagallo/react-native,eduardinni/react-native,javache/react-native,csatf/react-native,jadbox/react-native,BretJohnson/react-native,janicduplessis/react-native,imDangerous/react-native,urvashi01/react-native,catalinmiron/react-native,negativetwelve/react-native,javache/react-native,martinbigio/react-native,yamill/react-native,brentvatne/react-native,catalinmiron/react-native,htc2u/react-native,corbt/react-native,alin23/react-native,tadeuzagallo/react-native,chirag04/react-native,browniefed/react-native,philikon/react-native,iodine/react-native,foghina/react-native,Livyli/react-native,christopherdro/react-native,orenklein/react-native,mironiasty/react-native,htc2u/react-native,apprennet/react-native,hoastoolshop/react-native,DerayGa/react-native,Ehesp/react-native,dubert/react-native,rickbeerendonk/react-native,gre/react-native,dubert/react-native,foghina/react-native,ultralame/react-native,luqin/react-native,DerayGa/react-native,dikaiosune/react-native,Tredsite/react-native,orenklein/react-native,martinbigio/react-native,luqin/react-native,forcedotcom/react-native,ptomasroos/react-native,eduardinni/react-native,ndejesus1227/react-native,DerayGa/react-native,orenklein/react-native,dabit3/react-native,Tredsite/react-native,tgoldenberg/react-native,htc2u/react-native,facebook/react-native,cosmith/react-native,philikon/react-native,hoangpham95/react-native,mrspeaker/react-native,tsjing/react-native,farazs/react-native,Emilios1995/react-native,foghina/react-native,clozr/react-native,ndejesus1227/react-native,ptomasroos/react-native,tarkus/react-native-appletv,cosmith/react-native,gre/react-native,kesha-antonov/react-native,esauter5/react-native,browniefed/react-native,myntra/react-native,tsjing/react-native,aaron-goshine/react-native,xiayz/react-native,satya164/react-native,imDangerous/react-native,Tredsite/react-native,charlesvinette/react-native,PlexChat/react-native,martinbigio/react-native,hayeah/react-native,tadeuzagallo/react-native,pglotov/react-native,javache/react-native,sghiassy/react-native,exponentjs/react-native,adamjmcgrath/react-native,chirag04/react-native,urvashi01/react-native,xiayz/react-native,ptmt/react-native-macos,lprhodes/react-native,gilesvangruisen/react-native,corbt/react-native,csatf/react-native,PlexChat/react-native,csatf/react-native,MattFoley/react-native,gilesvangruisen/react-native,formatlos/react-native,jaggs6/react-native,frantic/react-native,Emilios1995/react-native,philikon/react-native,shrutic123/react-native,peterp/react-native,jaggs6/react-native,cpunion/react-native,Emilios1995/react-native,frantic/react-native,browniefed/react-native,makadaw/react-native,exponent/react-native,Purii/react-native,cosmith/react-native,nickhudkins/react-native,satya164/react-native,dabit3/react-native,alin23/react-native,chirag04/react-native,hammerandchisel/react-native,myntra/react-native,Ehesp/react-native,cpunion/react-native,Tredsite/react-native,shinate/react-native,doochik/react-native,nickhudkins/react-native,urvashi01/react-native,brentvatne/react-native,tsjing/react-native,jevakallio/react-native,jevakallio/react-native,makadaw/react-native,Maxwell2022/react-native,Purii/react-native,myntra/react-native,foghina/react-native,DannyvanderJagt/react-native,gilesvangruisen/react-native,arbesfeld/react-native,Guardiannw/react-native,aaron-goshine/react-native,dabit3/react-native,exponentjs/react-native,DannyvanderJagt/react-native,eduardinni/react-native,aljs/react-native,jhen0409/react-native,tadeuzagallo/react-native,Tredsite/react-native,doochik/react-native,jeffchienzabinet/react-native,dikaiosune/react-native,charlesvinette/react-native,lprhodes/react-native,skatpgusskat/react-native,imjerrybao/react-native,jaggs6/react-native,pandiaraj44/react-native,wesley1001/react-native,ultralame/react-native,cosmith/react-native,happypancake/react-native,exponent/react-native,satya164/react-native,exponent/react-native,forcedotcom/react-native,corbt/react-native,jhen0409/react-native,InterfaceInc/react-native,DannyvanderJagt/react-native,yamill/react-native,farazs/react-native,machard/react-native,gitim/react-native,shinate/react-native,myntra/react-native,salanki/react-native,skatpgusskat/react-native,compulim/react-native,tszajna0/react-native,iodine/react-native,hoastoolshop/react-native,MattFoley/react-native,browniefed/react-native,DerayGa/react-native,rickbeerendonk/react-native,xiayz/react-native,mrspeaker/react-native,yamill/react-native,lprhodes/react-native,imjerrybao/react-native,christopherdro/react-native,DanielMSchmidt/react-native,happypancake/react-native,tszajna0/react-native,dikaiosune/react-native,dubert/react-native,lprhodes/react-native,sghiassy/react-native,wenpkpk/react-native,farazs/react-native,sghiassy/react-native,cosmith/react-native,urvashi01/react-native,thotegowda/react-native,Purii/react-native,hoastoolshop/react-native,apprennet/react-native,exponentjs/react-native,imjerrybao/react-native,jasonnoahchoi/react-native,rebeccahughes/react-native,happypancake/react-native,MattFoley/react-native,mironiasty/react-native,naoufal/react-native,imDangerous/react-native,cosmith/react-native,wenpkpk/react-native,PlexChat/react-native,csatf/react-native,a2/react-native,CntChen/react-native,Ehesp/react-native,tszajna0/react-native,a2/react-native,ptmt/react-native-macos,gre/react-native,vjeux/react-native,iodine/react-native,BretJohnson/react-native,kesha-antonov/react-native,aljs/react-native,dubert/react-native,orenklein/react-native,chnfeeeeeef/react-native,luqin/react-native,tadeuzagallo/react-native,jeffchienzabinet/react-native,christopherdro/react-native,forcedotcom/react-native,janicduplessis/react-native,nickhudkins/react-native,hoangpham95/react-native,skatpgusskat/react-native,gilesvangruisen/react-native,CntChen/react-native,catalinmiron/react-native,iodine/react-native,BretJohnson/react-native,makadaw/react-native,Andreyco/react-native,philikon/react-native,ndejesus1227/react-native,machard/react-native,christopherdro/react-native,luqin/react-native,cosmith/react-native,Bhullnatik/react-native,gitim/react-native,shrutic/react-native,philikon/react-native,cpunion/react-native,Andreyco/react-native,Guardiannw/react-native,nathanajah/react-native,pandiaraj44/react-native,PlexChat/react-native,wesley1001/react-native,farazs/react-native,janicduplessis/react-native,Ehesp/react-native,shinate/react-native,jasonnoahchoi/react-native,aaron-goshine/react-native,gre/react-native,Swaagie/react-native,facebook/react-native,yamill/react-native,machard/react-native,orenklein/react-native,martinbigio/react-native,lelandrichardson/react-native,foghina/react-native,satya164/react-native,forcedotcom/react-native,christopherdro/react-native,negativetwelve/react-native,formatlos/react-native,wesley1001/react-native,hoangpham95/react-native,pandiaraj44/react-native,negativetwelve/react-native,tszajna0/react-native,peterp/react-native,urvashi01/react-native,shrutic/react-native,nathanajah/react-native,tarkus/react-native-appletv,javache/react-native,wesley1001/react-native,satya164/react-native,negativetwelve/react-native,CntChen/react-native,nathanajah/react-native,charlesvinette/react-native,facebook/react-native,Emilios1995/react-native,DanielMSchmidt/react-native,catalinmiron/react-native,exponentjs/react-native,Swaagie/react-native,DannyvanderJagt/react-native,Livyli/react-native,hammerandchisel/react-native,xiayz/react-native,nickhudkins/react-native,adamjmcgrath/react-native,tszajna0/react-native,ankitsinghania94/react-native,janicduplessis/react-native,callstack-io/react-native,MattFoley/react-native,salanki/react-native,kesha-antonov/react-native,foghina/react-native,csatf/react-native,cosmith/react-native,adamjmcgrath/react-native,dikaiosune/react-native,christopherdro/react-native,Purii/react-native,Bhullnatik/react-native,naoufal/react-native,rickbeerendonk/react-native,gitim/react-native,lelandrichardson/react-native,javache/react-native,lelandrichardson/react-native,gilesvangruisen/react-native,urvashi01/react-native,Swaagie/react-native,esauter5/react-native,javache/react-native,rickbeerendonk/react-native,DanielMSchmidt/react-native,facebook/react-native,nickhudkins/react-native,hoangpham95/react-native,DannyvanderJagt/react-native,aaron-goshine/react-native,Livyli/react-native,jhen0409/react-native,ptmt/react-native-macos,jeffchienzabinet/react-native,rickbeerendonk/react-native,martinbigio/react-native,naoufal/react-native,frantic/react-native,catalinmiron/react-native,hayeah/react-native,adamjmcgrath/react-native,Livyli/react-native,machard/react-native,kesha-antonov/react-native,ptmt/react-native-macos,clozr/react-native,tgoldenberg/react-native,rickbeerendonk/react-native,philikon/react-native,myntra/react-native,Bhullnatik/react-native,Bhullnatik/react-native,wenpkpk/react-native,thotegowda/react-native,brentvatne/react-native,happypancake/react-native,lprhodes/react-native,javache/react-native,peterp/react-native,jadbox/react-native,pglotov/react-native,skevy/react-native,callstack-io/react-native,vjeux/react-native,xiayz/react-native,salanki/react-native,shrutic123/react-native,tarkus/react-native-appletv,htc2u/react-native,jaggs6/react-native,mironiasty/react-native,pandiaraj44/react-native,skatpgusskat/react-native,jadbox/react-native,InterfaceInc/react-native,lprhodes/react-native,Tredsite/react-native,exponentjs/react-native,gilesvangruisen/react-native,ankitsinghania94/react-native,arbesfeld/react-native,callstack-io/react-native,formatlos/react-native,chnfeeeeeef/react-native,a2/react-native,nickhudkins/react-native,machard/react-native,doochik/react-native,Tredsite/react-native,miracle2k/react-native,miracle2k/react-native,makadaw/react-native,gitim/react-native,farazs/react-native,mrspeaker/react-native,ptomasroos/react-native,MattFoley/react-native,corbt/react-native,naoufal/react-native,compulim/react-native,myntra/react-native,tarkus/react-native-appletv,negativetwelve/react-native,Ehesp/react-native,facebook/react-native,dubert/react-native,frantic/react-native,dikaiosune/react-native,adamjmcgrath/react-native,vjeux/react-native,orenklein/react-native,Bhullnatik/react-native,facebook/react-native,CodeLinkIO/react-native,tarkus/react-native-appletv,MattFoley/react-native,Emilios1995/react-native,forcedotcom/react-native,dabit3/react-native,dikaiosune/react-native,orenklein/react-native,forcedotcom/react-native,DerayGa/react-native,cdlewis/react-native,arbesfeld/react-native,ptomasroos/react-native,hoastoolshop/react-native,Guardiannw/react-native,htc2u/react-native,Ehesp/react-native,vjeux/react-native,nathanajah/react-native,tadeuzagallo/react-native,cdlewis/react-native,wesley1001/react-native,alin23/react-native,imDangerous/react-native,cdlewis/react-native,arbesfeld/react-native,a2/react-native,pglotov/react-native,wenpkpk/react-native,jaggs6/react-native,satya164/react-native,gilesvangruisen/react-native,jhen0409/react-native,happypancake/react-native,janicduplessis/react-native,pglotov/react-native,DerayGa/react-native,formatlos/react-native,CodeLinkIO/react-native,makadaw/react-native,tadeuzagallo/react-native,adamjmcgrath/react-native,aljs/react-native,doochik/react-native,Livyli/react-native,peterp/react-native,lelandrichardson/react-native,hammerandchisel/react-native,brentvatne/react-native,ndejesus1227/react-native,PlexChat/react-native,InterfaceInc/react-native,skevy/react-native,luqin/react-native,jasonnoahchoi/react-native,callstack-io/react-native,mironiasty/react-native,brentvatne/react-native,cpunion/react-native,csatf/react-native,salanki/react-native,imjerrybao/react-native,martinbigio/react-native,pandiaraj44/react-native,tszajna0/react-native,apprennet/react-native,jadbox/react-native,naoufal/react-native,xiayz/react-native,eduardinni/react-native,csatf/react-native,salanki/react-native,farazs/react-native,nathanajah/react-native,CodeLinkIO/react-native,vjeux/react-native,hammerandchisel/react-native,callstack-io/react-native,brentvatne/react-native,cdlewis/react-native,lelandrichardson/react-native,dubert/react-native,foghina/react-native,DanielMSchmidt/react-native,pglotov/react-native,alin23/react-native,catalinmiron/react-native,aaron-goshine/react-native,hoangpham95/react-native,jeffchienzabinet/react-native,jeffchienzabinet/react-native,aljs/react-native,machard/react-native,frantic/react-native,DannyvanderJagt/react-native,browniefed/react-native,catalinmiron/react-native,shrutic/react-native,cdlewis/react-native,mrspeaker/react-native,skevy/react-native,clozr/react-native,urvashi01/react-native,wenpkpk/react-native,Maxwell2022/react-native,tgoldenberg/react-native,chnfeeeeeef/react-native,charlesvinette/react-native,exponentjs/react-native,miracle2k/react-native,Bhullnatik/react-native,chnfeeeeeef/react-native,jaggs6/react-native,exponentjs/react-native,myntra/react-native,farazs/react-native,jevakallio/react-native,cpunion/react-native,MattFoley/react-native,rickbeerendonk/react-native,satya164/react-native,ndejesus1227/react-native,Bhullnatik/react-native,wesley1001/react-native,PlexChat/react-native,makadaw/react-native,esauter5/react-native,Andreyco/react-native,tszajna0/react-native,htc2u/react-native,ankitsinghania94/react-native,jeffchienzabinet/react-native,tarkus/react-native-appletv,Maxwell2022/react-native,cpunion/react-native,mironiasty/react-native,Tredsite/react-native,martinbigio/react-native,cdlewis/react-native,charlesvinette/react-native,BretJohnson/react-native,javache/react-native,shrutic123/react-native,miracle2k/react-native,gitim/react-native,imDangerous/react-native,imjerrybao/react-native,clozr/react-native,doochik/react-native,chirag04/react-native,rickbeerendonk/react-native,luqin/react-native,Ehesp/react-native,dabit3/react-native,negativetwelve/react-native,luqin/react-native,cpunion/react-native,jhen0409/react-native,catalinmiron/react-native,clozr/react-native,sghiassy/react-native,salanki/react-native,nickhudkins/react-native,shrutic123/react-native,arthuralee/react-native,alin23/react-native,shinate/react-native,iodine/react-native,formatlos/react-native,tsjing/react-native,BretJohnson/react-native,DerayGa/react-native,CodeLinkIO/react-native,dabit3/react-native,xiayz/react-native,pandiaraj44/react-native,thotegowda/react-native,peterp/react-native,ptmt/react-native-macos,philikon/react-native,htc2u/react-native,Andreyco/react-native,InterfaceInc/react-native,Maxwell2022/react-native,arbesfeld/react-native,naoufal/react-native,CodeLinkIO/react-native,hoangpham95/react-native,miracle2k/react-native,imDangerous/react-native,jevakallio/react-native,CntChen/react-native,arbesfeld/react-native,Swaagie/react-native,Emilios1995/react-native,esauter5/react-native,lprhodes/react-native,mironiasty/react-native,makadaw/react-native,janicduplessis/react-native,kesha-antonov/react-native,compulim/react-native,aaron-goshine/react-native,apprennet/react-native,Maxwell2022/react-native,chirag04/react-native,apprennet/react-native,csatf/react-native,wenpkpk/react-native,nathanajah/react-native,skevy/react-native,mironiasty/react-native,BretJohnson/react-native,happypancake/react-native,tsjing/react-native,pandiaraj44/react-native,PlexChat/react-native,arthuralee/react-native,jevakallio/react-native,jasonnoahchoi/react-native,salanki/react-native,jeffchienzabinet/react-native,InterfaceInc/react-native,cdlewis/react-native,jhen0409/react-native,DannyvanderJagt/react-native,DanielMSchmidt/react-native,callstack-io/react-native,exponent/react-native,Emilios1995/react-native,nickhudkins/react-native,tgoldenberg/react-native,pglotov/react-native,wenpkpk/react-native,eduardinni/react-native,Andreyco/react-native,hayeah/react-native,browniefed/react-native,esauter5/react-native,pandiaraj44/react-native,ankitsinghania94/react-native,lprhodes/react-native,DanielMSchmidt/react-native,sghiassy/react-native,lelandrichardson/react-native,sghiassy/react-native,aaron-goshine/react-native,tsjing/react-native,shrutic123/react-native,pglotov/react-native,alin23/react-native,cdlewis/react-native,hoastoolshop/react-native,arbesfeld/react-native,shinate/react-native,jevakallio/react-native,brentvatne/react-native,lelandrichardson/react-native,hammerandchisel/react-native,Maxwell2022/react-native,Purii/react-native,orenklein/react-native,kesha-antonov/react-native,gitim/react-native
--- +++ @@ -0,0 +1,20 @@ +import contextlib +import os +import shutil +import sys +import tempfile +import zipfile + +# Helper that unpacks the contents of the res folder of an .aar file +# into given destination. + +@contextlib.contextmanager +def cleanup(path): + yield path + shutil.rmtree(path) + +if __name__ == '__main__': + with zipfile.ZipFile(sys.argv[1], 'r') as z: + with cleanup(tempfile.mkdtemp()) as temp_path: + z.extractall(temp_path, filter(lambda n: n.startswith('res/'), z.namelist())) + shutil.move(os.path.join(temp_path, 'res'), sys.argv[2])
359726a1085b00936e7e7ee3f876b09f72988921
dipy/utils/tests/test_tripwire.py
dipy/utils/tests/test_tripwire.py
""" Testing tripwire module. """ from ..tripwire import TripWire, is_tripwire, TripWireError from nose import SkipTest from nose.tools import (assert_true, assert_false, assert_raises, assert_equal, assert_not_equal) def test_is_tripwire(): assert_false(is_tripwire(object())) assert_true(is_tripwire(TripWire('some message'))) def test_tripwire(): # Test tripwire object silly_module_name = TripWire('We do not have silly_module_name') assert_raises(TripWireError, getattr, silly_module_name, 'do_silly_thing') # Check AttributeError can be checked too try: silly_module_name.__wrapped__ except TripWireError as err: assert_true(isinstance(err, AttributeError)) else: raise RuntimeError("No error raised, but expected")
Add a test for TripWire.
TST: Add a test for TripWire.
Python
bsd-3-clause
StongeEtienne/dipy,villalonreina/dipy,villalonreina/dipy,nilgoyyou/dipy,FrancoisRheaultUS/dipy,matthieudumont/dipy,StongeEtienne/dipy,nilgoyyou/dipy,matthieudumont/dipy,FrancoisRheaultUS/dipy
--- +++ @@ -0,0 +1,29 @@ +""" Testing tripwire module. +""" + +from ..tripwire import TripWire, is_tripwire, TripWireError + +from nose import SkipTest +from nose.tools import (assert_true, assert_false, assert_raises, + assert_equal, assert_not_equal) + + +def test_is_tripwire(): + assert_false(is_tripwire(object())) + assert_true(is_tripwire(TripWire('some message'))) + + +def test_tripwire(): + # Test tripwire object + silly_module_name = TripWire('We do not have silly_module_name') + assert_raises(TripWireError, + getattr, + silly_module_name, + 'do_silly_thing') + # Check AttributeError can be checked too + try: + silly_module_name.__wrapped__ + except TripWireError as err: + assert_true(isinstance(err, AttributeError)) + else: + raise RuntimeError("No error raised, but expected")
1103c498635480d516eeebbec615f8d13db51bb7
api/tests/test_topic_api.py
api/tests/test_topic_api.py
import pytz from datetime import datetime, timedelta from django.test import TestCase from api.factories import TopicFactory from rest_framework.test import APIRequestFactory, force_authenticate from api.factories import UserFactory from api.serializers import TopicSerializer from api.views.topics import TopicViewSet class TopicAPITest(TestCase): def setUp(self): self.factory = APIRequestFactory() self.active_view = TopicViewSet.as_view({'get': 'active'}) self.inactive_view = TopicViewSet.as_view({'get': 'inactive'}) self.user = UserFactory() self.active_topics_count = 5 self.inactive_topics_count = 5 TopicFactory.create_batch(self.active_topics_count, start=datetime.now(pytz.utc) - timedelta(days=3), end=datetime.now(pytz.utc) + timedelta(days=4)) TopicFactory.create_batch(self.inactive_topics_count, start=datetime.now(pytz.utc) + timedelta(weeks=1), end=datetime.now(pytz.utc) + timedelta(weeks=2)) def test_active_topics_are_returned_correctly(self): request = self.factory.get("api/topics/active") force_authenticate(request, user=self.user) response = self.active_view(request) topics = TopicSerializer(data=response.data, many=True) topics.is_valid(raise_exception=True) topics = topics.create(topics.validated_data) self.assertEqual(self.active_topics_count, len(topics)) for topic in topics: self.assertTrue(topic.is_active()) def test_inactive_topics_are_returned_correctly(self): request = self.factory.get("api/topics/inactive") force_authenticate(request, user=self.user) response = self.inactive_view(request) topics = TopicSerializer(data=response.data, many=True) topics.is_valid(raise_exception=True) topics = topics.create(topics.validated_data) self.assertEqual(self.inactive_topics_count, len(topics)) for topic in topics: self.assertFalse(topic.is_active())
Write tests for active and inactive list routes
Write tests for active and inactive list routes
Python
mit
frostblooded/kanq,frostblooded/kanq,frostblooded/kanq,frostblooded/kanq,frostblooded/kanq
--- +++ @@ -0,0 +1,51 @@ +import pytz +from datetime import datetime, timedelta +from django.test import TestCase + +from api.factories import TopicFactory +from rest_framework.test import APIRequestFactory, force_authenticate +from api.factories import UserFactory +from api.serializers import TopicSerializer +from api.views.topics import TopicViewSet + + +class TopicAPITest(TestCase): + def setUp(self): + self.factory = APIRequestFactory() + self.active_view = TopicViewSet.as_view({'get': 'active'}) + self.inactive_view = TopicViewSet.as_view({'get': 'inactive'}) + self.user = UserFactory() + + self.active_topics_count = 5 + self.inactive_topics_count = 5 + + TopicFactory.create_batch(self.active_topics_count, start=datetime.now(pytz.utc) - timedelta(days=3), + end=datetime.now(pytz.utc) + timedelta(days=4)) + TopicFactory.create_batch(self.inactive_topics_count, start=datetime.now(pytz.utc) + timedelta(weeks=1), + end=datetime.now(pytz.utc) + timedelta(weeks=2)) + + def test_active_topics_are_returned_correctly(self): + request = self.factory.get("api/topics/active") + force_authenticate(request, user=self.user) + response = self.active_view(request) + topics = TopicSerializer(data=response.data, many=True) + topics.is_valid(raise_exception=True) + topics = topics.create(topics.validated_data) + + self.assertEqual(self.active_topics_count, len(topics)) + + for topic in topics: + self.assertTrue(topic.is_active()) + + def test_inactive_topics_are_returned_correctly(self): + request = self.factory.get("api/topics/inactive") + force_authenticate(request, user=self.user) + response = self.inactive_view(request) + topics = TopicSerializer(data=response.data, many=True) + topics.is_valid(raise_exception=True) + topics = topics.create(topics.validated_data) + + self.assertEqual(self.inactive_topics_count, len(topics)) + + for topic in topics: + self.assertFalse(topic.is_active())
fb876d0ddc12fa89db5e3bec519ad4e93e914b8d
designate/tests/unit/api/test_version.py
designate/tests/unit/api/test_version.py
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from oslo_config import cfg from oslo_config import fixture as cfg_fixture import oslotest.base import webtest from designate.api import versions from designate.common import constants CONF = cfg.CONF class TestApiVersion(oslotest.base.BaseTestCase): def setUp(self): super(TestApiVersion, self).setUp() self.useFixture(cfg_fixture.Config(CONF)) def test_add_a_version(self): api_url = 'http://localhost/v2' results = [] versions._add_a_version( results, 'v2.1', api_url, constants.EXPERIMENTAL, '2022-08-10T00:00:00Z') self.assertEqual(1, len(results)) self.assertEqual('v2.1', results[0]['id']) self.assertEqual(constants.EXPERIMENTAL, results[0]['status']) self.assertEqual('2022-08-10T00:00:00Z', results[0]['updated']) self.assertEqual(2, len(results[0]['links'])) def test_get_versions(self): CONF.set_override('enable_host_header', False, 'service:api') CONF.set_override( 'api_base_uri', 'http://127.0.0.2:9001/', 'service:api' ) self.app = versions.factory({}) self.client = webtest.TestApp(self.app) response = self.client.get('/') self.assertEqual(200, response.status_int) self.assertEqual('application/json', response.content_type) self.assertEqual(2, len(response.json['versions'])) self.assertEqual( 'http://127.0.0.2:9001/v2', response.json['versions'][0]['links'][0]['href'] ) def test_get_versions_with_enable_host_header(self): CONF.set_override('enable_host_header', True, 'service:api') self.app = versions.factory({}) self.client = webtest.TestApp(self.app) response = self.client.get('/') self.assertEqual(200, response.status_int) self.assertEqual('application/json', response.content_type) self.assertEqual(2, len(response.json['versions'])) self.assertEqual( 'http://localhost/v2', response.json['versions'][0]['links'][0]['href'] )
Add basic api version test coverage
Add basic api version test coverage Change-Id: Idb04e81ce5954ca8b5e387dc7c0776fdf7d08779
Python
apache-2.0
openstack/designate,openstack/designate,openstack/designate
--- +++ @@ -0,0 +1,78 @@ +# Licensed under the Apache License, Version 2.0 (the "License"); you may +# not use this file except in compliance with the License. You may obtain +# a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT +# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the +# License for the specific language governing permissions and limitations +# under the License. + +from oslo_config import cfg +from oslo_config import fixture as cfg_fixture +import oslotest.base +import webtest + +from designate.api import versions +from designate.common import constants + + +CONF = cfg.CONF + + +class TestApiVersion(oslotest.base.BaseTestCase): + + def setUp(self): + super(TestApiVersion, self).setUp() + self.useFixture(cfg_fixture.Config(CONF)) + + def test_add_a_version(self): + api_url = 'http://localhost/v2' + results = [] + + versions._add_a_version( + results, 'v2.1', api_url, constants.EXPERIMENTAL, + '2022-08-10T00:00:00Z') + + self.assertEqual(1, len(results)) + self.assertEqual('v2.1', results[0]['id']) + self.assertEqual(constants.EXPERIMENTAL, results[0]['status']) + self.assertEqual('2022-08-10T00:00:00Z', results[0]['updated']) + self.assertEqual(2, len(results[0]['links'])) + + def test_get_versions(self): + CONF.set_override('enable_host_header', False, 'service:api') + CONF.set_override( + 'api_base_uri', 'http://127.0.0.2:9001/', 'service:api' + ) + + self.app = versions.factory({}) + self.client = webtest.TestApp(self.app) + + response = self.client.get('/') + self.assertEqual(200, response.status_int) + self.assertEqual('application/json', response.content_type) + + self.assertEqual(2, len(response.json['versions'])) + self.assertEqual( + 'http://127.0.0.2:9001/v2', + response.json['versions'][0]['links'][0]['href'] + ) + + def test_get_versions_with_enable_host_header(self): + CONF.set_override('enable_host_header', True, 'service:api') + + self.app = versions.factory({}) + self.client = webtest.TestApp(self.app) + + response = self.client.get('/') + self.assertEqual(200, response.status_int) + self.assertEqual('application/json', response.content_type) + + self.assertEqual(2, len(response.json['versions'])) + self.assertEqual( + 'http://localhost/v2', + response.json['versions'][0]['links'][0]['href'] + )
964895852bb9acbe866c1a7bc7ba98b972100fa8
readthedocs/builds/migrations/0014_migrate-doctype-from-project-to-version.py
readthedocs/builds/migrations/0014_migrate-doctype-from-project-to-version.py
# -*- coding: utf-8 -*- # Generated by Django 1.11.27 on 2020-01-14 17:40 from django.db import migrations from django.db.models import OuterRef, Subquery def forwards_func(apps, schema_editor): """Migrate ``Project.documentation_type`` to ``Version.documentation_type``.""" Version = apps.get_model('builds', 'Version') Project = apps.get_model('projects', 'Project') Version.objects.all().update( documentation_type=Subquery( Project.objects .filter(pk=OuterRef('project')) .values('documentation_type')[:1] ), ) class Migration(migrations.Migration): dependencies = [ ('builds', '0013_version_documentation_type'), ] operations = [ migrations.RunPython(forwards_func), ]
Migrate documentation_type from Project to Version
Migrate documentation_type from Project to Version
Python
mit
rtfd/readthedocs.org,rtfd/readthedocs.org,rtfd/readthedocs.org,rtfd/readthedocs.org
--- +++ @@ -0,0 +1,30 @@ +# -*- coding: utf-8 -*- +# Generated by Django 1.11.27 on 2020-01-14 17:40 + +from django.db import migrations +from django.db.models import OuterRef, Subquery + + +def forwards_func(apps, schema_editor): + """Migrate ``Project.documentation_type`` to ``Version.documentation_type``.""" + Version = apps.get_model('builds', 'Version') + Project = apps.get_model('projects', 'Project') + + Version.objects.all().update( + documentation_type=Subquery( + Project.objects + .filter(pk=OuterRef('project')) + .values('documentation_type')[:1] + ), + ) + + +class Migration(migrations.Migration): + + dependencies = [ + ('builds', '0013_version_documentation_type'), + ] + + operations = [ + migrations.RunPython(forwards_func), + ]
adca0c2ff48b716e9ac3ef706d66d86943fb29b7
cpm_data/migrations/0015_add_jury_members_to_seasons_12_14.py
cpm_data/migrations/0015_add_jury_members_to_seasons_12_14.py
# -*- coding: utf-8 -*- from __future__ import unicode_literals from functools import partial from django.db import migrations def add_to_season(apps, schema_editor, season_name, jury_names): JuryMember = apps.get_model('cpm_data.JuryMember') Season = apps.get_model('cpm_data.Season') SeasonJuryMember = apps.get_model('cpm_data.SeasonRelatedJuryMember') season = Season.objects.get(name_en=season_name) for name in jury_names: try: jury_member = JuryMember.objects.get(name_en=name) except JuryMember.DoesNotExist: raise SeasonJuryMember.objects.get_or_create(season=season, jury_member=jury_member) def remove_from_season(apps, schema_editor, season_name, jury_names): Season = apps.get_model('cpm_data.Season') SeasonJuryMember = apps.get_model('cpm_data.SeasonRelatedJuryMember') season = Season.objects.get(name_en=season_name) season_jury_members = SeasonJuryMember.objects.filter( season=season, jury_member__name_en__in=jury_names ) season_jury_members.delete() class Migration(migrations.Migration): dependencies = [ ('cpm_data', '0014_remove_jury_duplicates'), ] operations = [ migrations.RunPython( partial(add_to_season, season_name=season_name, jury_names=jury_names), partial(remove_from_season, season_name=season_name, jury_names=jury_names), ) for season_name, jury_names in [ ('2012', ['Agricola de Cologne', 'Victor Aslyuk', 'Maciej Gil', 'Pavel Ivanov', 'Yuri Igrusha', 'Andrew Kureichik', 'Sergey Krasikov', 'Bohdana Smirnova', 'Cory McAbee']), ('2013', ['Volha Dashuk', 'Irina Demyanova', 'Anton Sidorenko', 'Pavel Ivanov', 'Yuri Igrusha', 'Sergey Krasikov', 'Bohdana Smirnova', 'Jon Rubin', 'Maciej Gil', 'Pierre-Luc Vaillancourt', 'Karsten Weber', 'Lee Sang-woo', 'Cory McAbee']), ('2014', ['Yuri Igrusha']), # more added in previous migrations ] ]
Add jury to 2012-2014 - take the data from results pages
Add jury to 2012-2014 - take the data from results pages
Python
unlicense
kinaklub/next.filmfest.by,kinaklub/next.filmfest.by,kinaklub/next.filmfest.by,kinaklub/next.filmfest.by
--- +++ @@ -0,0 +1,63 @@ +# -*- coding: utf-8 -*- +from __future__ import unicode_literals + +from functools import partial + +from django.db import migrations + + +def add_to_season(apps, schema_editor, season_name, jury_names): + JuryMember = apps.get_model('cpm_data.JuryMember') + Season = apps.get_model('cpm_data.Season') + SeasonJuryMember = apps.get_model('cpm_data.SeasonRelatedJuryMember') + + season = Season.objects.get(name_en=season_name) + + for name in jury_names: + try: + jury_member = JuryMember.objects.get(name_en=name) + except JuryMember.DoesNotExist: + raise + + SeasonJuryMember.objects.get_or_create(season=season, + jury_member=jury_member) + + +def remove_from_season(apps, schema_editor, season_name, jury_names): + Season = apps.get_model('cpm_data.Season') + SeasonJuryMember = apps.get_model('cpm_data.SeasonRelatedJuryMember') + + season = Season.objects.get(name_en=season_name) + season_jury_members = SeasonJuryMember.objects.filter( + season=season, jury_member__name_en__in=jury_names + ) + + season_jury_members.delete() + + +class Migration(migrations.Migration): + + dependencies = [ + ('cpm_data', '0014_remove_jury_duplicates'), + ] + + operations = [ + migrations.RunPython( + partial(add_to_season, + season_name=season_name, + jury_names=jury_names), + partial(remove_from_season, + season_name=season_name, + jury_names=jury_names), + ) for season_name, jury_names in [ + ('2012', ['Agricola de Cologne', 'Victor Aslyuk', 'Maciej Gil', + 'Pavel Ivanov', 'Yuri Igrusha', 'Andrew Kureichik', + 'Sergey Krasikov', 'Bohdana Smirnova', 'Cory McAbee']), + ('2013', ['Volha Dashuk', 'Irina Demyanova', 'Anton Sidorenko', + 'Pavel Ivanov', 'Yuri Igrusha', 'Sergey Krasikov', + 'Bohdana Smirnova', 'Jon Rubin', 'Maciej Gil', + 'Pierre-Luc Vaillancourt', 'Karsten Weber', + 'Lee Sang-woo', 'Cory McAbee']), + ('2014', ['Yuri Igrusha']), # more added in previous migrations + ] + ]
48a37b00bb73ea0398644e0eb22c50c34919b325
examples/non_terminating_checks_test.py
examples/non_terminating_checks_test.py
from seleniumbase import BaseCase class MyTestClass(BaseCase): def test_non_terminating_checks(self): self.open('http://xkcd.com/993/') self.wait_for_element('#comic') self.check_assert_element('img[alt="Brand Identity"]') self.check_assert_element('img[alt="Rocket Ship"]') # Will Fail self.check_assert_element('#comicmap') self.check_assert_text('Fake Item', '#middleContainer') # Will Fail self.check_assert_text('Random', '#middleContainer') self.check_assert_element('a[name="Super Fake !!!"]') # Will Fail self.process_checks()
Add example test for non-terminating checks
Add example test for non-terminating checks
Python
mit
mdmintz/SeleniumBase,possoumous/Watchers,possoumous/Watchers,ktp420/SeleniumBase,mdmintz/SeleniumBase,seleniumbase/SeleniumBase,seleniumbase/SeleniumBase,ktp420/SeleniumBase,ktp420/SeleniumBase,possoumous/Watchers,mdmintz/seleniumspot,seleniumbase/SeleniumBase,seleniumbase/SeleniumBase,mdmintz/seleniumspot,mdmintz/SeleniumBase,ktp420/SeleniumBase,possoumous/Watchers,mdmintz/SeleniumBase
--- +++ @@ -0,0 +1,15 @@ +from seleniumbase import BaseCase + + +class MyTestClass(BaseCase): + + def test_non_terminating_checks(self): + self.open('http://xkcd.com/993/') + self.wait_for_element('#comic') + self.check_assert_element('img[alt="Brand Identity"]') + self.check_assert_element('img[alt="Rocket Ship"]') # Will Fail + self.check_assert_element('#comicmap') + self.check_assert_text('Fake Item', '#middleContainer') # Will Fail + self.check_assert_text('Random', '#middleContainer') + self.check_assert_element('a[name="Super Fake !!!"]') # Will Fail + self.process_checks()
6e02077b0aab178a2d8100a26811a485a74852f2
flexget/plugins/urlrewrite_shortened.py
flexget/plugins/urlrewrite_shortened.py
from __future__ import unicode_literals, division, absolute_import import logging from urlparse import urlparse from flexget import plugin from flexget.utils import requests from flexget.event import event log = logging.getLogger('shortened') class UrlRewriteShortened(object): """Shortened url rewriter.""" def url_rewritable(self, task, entry): return urlparse(entry['url']).netloc in ['bit.ly', 't.co'] def url_rewrite(self, task, entry): request = task.requests.head(entry['url'], allow_redirects=True) entry['url'] = request.url @event('plugin.register') def register_plugin(): plugin.register(UrlRewriteShortened, 'shortened', groups=['urlrewriter'], api_ver=2)
Add url rewriter for shortened urls
Add url rewriter for shortened urls
Python
mit
ratoaq2/Flexget,gazpachoking/Flexget,qvazzler/Flexget,Flexget/Flexget,X-dark/Flexget,poulpito/Flexget,LynxyssCZ/Flexget,offbyone/Flexget,poulpito/Flexget,X-dark/Flexget,sean797/Flexget,malkavi/Flexget,tsnoam/Flexget,JorisDeRieck/Flexget,Pretagonist/Flexget,jawilson/Flexget,OmgOhnoes/Flexget,Pretagonist/Flexget,qvazzler/Flexget,grrr2/Flexget,thalamus/Flexget,jacobmetrick/Flexget,oxc/Flexget,voriux/Flexget,Pretagonist/Flexget,tarzasai/Flexget,OmgOhnoes/Flexget,patsissons/Flexget,LynxyssCZ/Flexget,Flexget/Flexget,cvium/Flexget,voriux/Flexget,thalamus/Flexget,jawilson/Flexget,ZefQ/Flexget,offbyone/Flexget,ianstalk/Flexget,crawln45/Flexget,OmgOhnoes/Flexget,antivirtel/Flexget,ibrahimkarahan/Flexget,sean797/Flexget,drwyrm/Flexget,drwyrm/Flexget,qk4l/Flexget,v17al/Flexget,xfouloux/Flexget,patsissons/Flexget,tvcsantos/Flexget,vfrc2/Flexget,tsnoam/Flexget,antivirtel/Flexget,dsemi/Flexget,Flexget/Flexget,ratoaq2/Flexget,tvcsantos/Flexget,poulpito/Flexget,grrr2/Flexget,spencerjanssen/Flexget,oxc/Flexget,crawln45/Flexget,Danfocus/Flexget,qk4l/Flexget,tobinjt/Flexget,malkavi/Flexget,crawln45/Flexget,oxc/Flexget,lildadou/Flexget,vfrc2/Flexget,jacobmetrick/Flexget,xfouloux/Flexget,xfouloux/Flexget,tobinjt/Flexget,ratoaq2/Flexget,tobinjt/Flexget,ianstalk/Flexget,gazpachoking/Flexget,lildadou/Flexget,JorisDeRieck/Flexget,Danfocus/Flexget,qk4l/Flexget,jawilson/Flexget,drwyrm/Flexget,tsnoam/Flexget,X-dark/Flexget,vfrc2/Flexget,tarzasai/Flexget,LynxyssCZ/Flexget,ZefQ/Flexget,malkavi/Flexget,v17al/Flexget,Danfocus/Flexget,LynxyssCZ/Flexget,antivirtel/Flexget,grrr2/Flexget,dsemi/Flexget,v17al/Flexget,crawln45/Flexget,jawilson/Flexget,ibrahimkarahan/Flexget,spencerjanssen/Flexget,spencerjanssen/Flexget,cvium/Flexget,patsissons/Flexget,Flexget/Flexget,cvium/Flexget,tarzasai/Flexget,ianstalk/Flexget,ZefQ/Flexget,sean797/Flexget,JorisDeRieck/Flexget,thalamus/Flexget,malkavi/Flexget,JorisDeRieck/Flexget,Danfocus/Flexget,ibrahimkarahan/Flexget,qvazzler/Flexget,lildadou/Flexget,camon/Flexget,tobinjt/Flexget,dsemi/Flexget,camon/Flexget,offbyone/Flexget,jacobmetrick/Flexget
--- +++ @@ -0,0 +1,24 @@ +from __future__ import unicode_literals, division, absolute_import +import logging +from urlparse import urlparse + +from flexget import plugin +from flexget.utils import requests +from flexget.event import event + +log = logging.getLogger('shortened') + + +class UrlRewriteShortened(object): + """Shortened url rewriter.""" + + def url_rewritable(self, task, entry): + return urlparse(entry['url']).netloc in ['bit.ly', 't.co'] + + def url_rewrite(self, task, entry): + request = task.requests.head(entry['url'], allow_redirects=True) + entry['url'] = request.url + +@event('plugin.register') +def register_plugin(): + plugin.register(UrlRewriteShortened, 'shortened', groups=['urlrewriter'], api_ver=2)
2ffcae174d7ff10aa816b8fb839c9c0ae6fb3689
bidwire/alembic/versions/a1b42c9006a7_absolute_massgov_eopss_url.py
bidwire/alembic/versions/a1b42c9006a7_absolute_massgov_eopss_url.py
"""absolute_massgov_eopss_url Revision ID: a1b42c9006a7 Revises: 9b30b0fe231a Create Date: 2017-06-26 00:02:45.998655 """ from alembic import op import sqlalchemy as sa from sqlalchemy.orm.session import Session import os import sys sys.path.append(os.path.dirname(os.path.dirname(__file__))) from document import Document from urllib import parse # revision identifiers, used by Alembic. revision = 'a1b42c9006a7' down_revision = '9b30b0fe231a' branch_labels = None depends_on = None def upgrade(): def ensure_absolute(url): root_url = "https://www.mass.gov/" if not url.startswith(root_url): return parse.urljoin(root_url, url) return url # Attach to the migration's session session = Session(bind=op.get_bind()) docs = session.query(Document).filter( Document.site == Document.Site.MASSGOV_EOPSS.name).all() for doc in docs: doc.url = ensure_absolute(doc.url) session.add_all(docs) session.commit() def downgrade(): # Do nothing for the rollback. pass
Add database migration to make all Massgov Eopss URLs absolute.
Add database migration to make all Massgov Eopss URLs absolute.
Python
mit
RagtagOpen/bidwire,RagtagOpen/bidwire,RagtagOpen/bidwire
--- +++ @@ -0,0 +1,44 @@ +"""absolute_massgov_eopss_url + +Revision ID: a1b42c9006a7 +Revises: 9b30b0fe231a +Create Date: 2017-06-26 00:02:45.998655 + +""" +from alembic import op +import sqlalchemy as sa +from sqlalchemy.orm.session import Session + +import os +import sys +sys.path.append(os.path.dirname(os.path.dirname(__file__))) +from document import Document +from urllib import parse + +# revision identifiers, used by Alembic. +revision = 'a1b42c9006a7' +down_revision = '9b30b0fe231a' +branch_labels = None +depends_on = None + + +def upgrade(): + def ensure_absolute(url): + root_url = "https://www.mass.gov/" + if not url.startswith(root_url): + return parse.urljoin(root_url, url) + return url + + # Attach to the migration's session + session = Session(bind=op.get_bind()) + docs = session.query(Document).filter( + Document.site == Document.Site.MASSGOV_EOPSS.name).all() + for doc in docs: + doc.url = ensure_absolute(doc.url) + session.add_all(docs) + session.commit() + + +def downgrade(): + # Do nothing for the rollback. + pass
cfca3ecf6d36611ff0ffe3537f01c2711e94f87e
kpub/tests/test_counts.py
kpub/tests/test_counts.py
import kpub def test_annual_count(): # Does the cumulative count match the annual count? db = kpub.PublicationDB() annual = db.get_annual_publication_count() cumul = db.get_annual_publication_count_cumulative() assert annual['k2'][2010] == 0 # K2 didn't exist in 2010 # The first K2 papers started appearing in 2014; the cumulative counts should reflect that: assert (annual['k2'][2014] + annual['k2'][2015]) == cumul['k2'][2015] assert (annual['k2'][2014] + annual['k2'][2015] + annual['k2'][2016]) == cumul['k2'][2016]
Add unit test for get_annual_publication_count
Add unit test for get_annual_publication_count
Python
mit
KeplerGO/kpub
--- +++ @@ -0,0 +1,12 @@ +import kpub + + +def test_annual_count(): + # Does the cumulative count match the annual count? + db = kpub.PublicationDB() + annual = db.get_annual_publication_count() + cumul = db.get_annual_publication_count_cumulative() + assert annual['k2'][2010] == 0 # K2 didn't exist in 2010 + # The first K2 papers started appearing in 2014; the cumulative counts should reflect that: + assert (annual['k2'][2014] + annual['k2'][2015]) == cumul['k2'][2015] + assert (annual['k2'][2014] + annual['k2'][2015] + annual['k2'][2016]) == cumul['k2'][2016]
d0b98a062eaca03dba53f610bb4b326d769464e9
foundation/organisation/migrations/0008_auto_20160707_0752.py
foundation/organisation/migrations/0008_auto_20160707_0752.py
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('organisation', '0007_add_old_project_bool_to_project_model'), ] operations = [ migrations.AlterField( model_name='networkgroupmembership', name='order', field=models.IntegerField(help_text=b'The lower the number the higher on the page this Person will be shown.', null=True, blank=True), ), ]
Add updated help text for order in networkgroup
Add updated help text for order in networkgroup
Python
mit
okfn/website,okfn/foundation,okfn/website,okfn/website,okfn/foundation,okfn/foundation,okfn/foundation,okfn/website
--- +++ @@ -0,0 +1,19 @@ +# -*- coding: utf-8 -*- +from __future__ import unicode_literals + +from django.db import migrations, models + + +class Migration(migrations.Migration): + + dependencies = [ + ('organisation', '0007_add_old_project_bool_to_project_model'), + ] + + operations = [ + migrations.AlterField( + model_name='networkgroupmembership', + name='order', + field=models.IntegerField(help_text=b'The lower the number the higher on the page this Person will be shown.', null=True, blank=True), + ), + ]
3075a2dfcd51b932fd142e51e6bdc1694ab4e750
gitlabform/gitlabform/test/test_tags.py
gitlabform/gitlabform/test/test_tags.py
from gitlabform.gitlabform.test import ( run_gitlabform, ) class TestArchiveProject: def test__archive_project(self, gitlab, group, project): group_and_project = f"{group}/{project}" config = f""" projects_and_groups: {group_and_project}: project: archive: true """ run_gitlabform(config, group_and_project) project = gitlab.get_project(group_and_project) assert project["archived"] is True def test__unarchive_project(self, gitlab, group, project): group_and_project = f"{group}/{project}" archive_project = f""" projects_and_groups: {group_and_project}: project: archive: true """ unarchive_project = f""" projects_and_groups: {group_and_project}: project: archive: false """ run_gitlabform(archive_project, group_and_project) project = gitlab.get_project(group_and_project) assert project["archived"] is True run_gitlabform(unarchive_project, group_and_project) project = gitlab.get_project(group_and_project) assert project["archived"] is False def test__dont_edit_archived_project(self, gitlab, group, project): group_and_project = f"{group}/{project}" archive_project = f""" projects_and_groups: {group_and_project}: project: archive: true """ run_gitlabform(archive_project, group_and_project) project = gitlab.get_project(group_and_project) assert project["archived"] is True edit_archived_project = f""" # the project has to be configured as archived # for other configs for it to be ignored projects_and_groups: {group_and_project}: project: archive: true {group}/*: files: README.md: overwrite: true branches: - main content: | Some other content that the default one """ run_gitlabform(edit_archived_project, group_and_project) # the fact that we are not getting an exception because of trying to edit # an archived project means that the test is passing
Add tests for protected tags
Add tests for protected tags
Python
mit
egnyte/gitlabform,egnyte/gitlabform
--- +++ @@ -0,0 +1,81 @@ +from gitlabform.gitlabform.test import ( + run_gitlabform, +) + + +class TestArchiveProject: + def test__archive_project(self, gitlab, group, project): + group_and_project = f"{group}/{project}" + + config = f""" + projects_and_groups: + {group_and_project}: + project: + archive: true + """ + + run_gitlabform(config, group_and_project) + project = gitlab.get_project(group_and_project) + assert project["archived"] is True + + def test__unarchive_project(self, gitlab, group, project): + group_and_project = f"{group}/{project}" + + archive_project = f""" + projects_and_groups: + {group_and_project}: + project: + archive: true + """ + + unarchive_project = f""" + projects_and_groups: + {group_and_project}: + project: + archive: false + """ + + run_gitlabform(archive_project, group_and_project) + project = gitlab.get_project(group_and_project) + assert project["archived"] is True + + run_gitlabform(unarchive_project, group_and_project) + project = gitlab.get_project(group_and_project) + assert project["archived"] is False + + def test__dont_edit_archived_project(self, gitlab, group, project): + group_and_project = f"{group}/{project}" + + archive_project = f""" + projects_and_groups: + {group_and_project}: + project: + archive: true + """ + + run_gitlabform(archive_project, group_and_project) + project = gitlab.get_project(group_and_project) + assert project["archived"] is True + + edit_archived_project = f""" + # the project has to be configured as archived + # for other configs for it to be ignored + projects_and_groups: + {group_and_project}: + project: + archive: true + + {group}/*: + files: + README.md: + overwrite: true + branches: + - main + content: | + Some other content that the default one + """ + + run_gitlabform(edit_archived_project, group_and_project) + + # the fact that we are not getting an exception because of trying to edit + # an archived project means that the test is passing
1ac9175d661cc3d42416f3f55c5c8212467c612c
make_plots.py
make_plots.py
"""Make plots of the results of Dakotathon experiments.""" import numpy as np import matplotlib.pyplot as mpl def read_dat_header(dat_file): try: with open(dat_file, 'r') as fp: names = fp.readline().split() except IOError: pass else: return names def read_dat_file(dat_file): names = read_dat_header(dat_file) rnames = range(len(names)) rnames.pop(names.index('interface')) return np.loadtxt(dat_file, skiprows=1, unpack=True, usecols=rnames) def make_stacked_surface_plot(): pass def make_pdf_and_cdf_plot(): pass if __name__ == '__main__': make_stacked_surface_plot() make_pdf_and_cdf_plot()
Read Dakota tabular graphics output file
Read Dakota tabular graphics output file
Python
mit
mdpiper/AGU-2016
--- +++ @@ -0,0 +1,35 @@ +"""Make plots of the results of Dakotathon experiments.""" + +import numpy as np +import matplotlib.pyplot as mpl + + + +def read_dat_header(dat_file): + try: + with open(dat_file, 'r') as fp: + names = fp.readline().split() + except IOError: + pass + else: + return names + + +def read_dat_file(dat_file): + names = read_dat_header(dat_file) + rnames = range(len(names)) + rnames.pop(names.index('interface')) + return np.loadtxt(dat_file, skiprows=1, unpack=True, usecols=rnames) + + +def make_stacked_surface_plot(): + pass + + +def make_pdf_and_cdf_plot(): + pass + + +if __name__ == '__main__': + make_stacked_surface_plot() + make_pdf_and_cdf_plot()
38978ce807966a1367ae2ad95dc81f87a08c2ca3
pbm2sh.py
pbm2sh.py
import sys import re with open(sys.argv[1]) as f: magic = f.readline().strip() if magic != "P1": print "Expected file to start with P1" sys.exit(1) f.readline() # Ignore comment line w, h = [int(n) for n in f.readline().strip().split(" ")] pixels = re.sub(r"[^01]", "", f.read()) lines = [pixels[i:i+w] for i in xrange(0, len(pixels), w)] if len(lines) > 64: print "Max 64 lines" sys.exit(1) lines.extend(["0"*w] * (64 % len(lines))) # Pad out to 64 lines cols = w if cols > 128: print "Max 128 columns" sys.exit(1) for i in range(cols): for byteoff in range(8): bits = [] for bitoff in range(8): row = lines[byteoff * 8 + bitoff] bits.append(row[i]) print "write_data", " ".join(reversed(bits)) print ""
Add script to generate script from pbm files
Add script to generate script from pbm files
Python
mit
pib/gpio_ssd1306,pib/gpio_ssd1306
--- +++ @@ -0,0 +1,32 @@ +import sys +import re + +with open(sys.argv[1]) as f: + magic = f.readline().strip() + if magic != "P1": + print "Expected file to start with P1" + sys.exit(1) + + f.readline() # Ignore comment line + w, h = [int(n) for n in f.readline().strip().split(" ")] + pixels = re.sub(r"[^01]", "", f.read()) + lines = [pixels[i:i+w] for i in xrange(0, len(pixels), w)] + +if len(lines) > 64: + print "Max 64 lines" + sys.exit(1) + +lines.extend(["0"*w] * (64 % len(lines))) # Pad out to 64 lines +cols = w +if cols > 128: + print "Max 128 columns" + sys.exit(1) + +for i in range(cols): + for byteoff in range(8): + bits = [] + for bitoff in range(8): + row = lines[byteoff * 8 + bitoff] + bits.append(row[i]) + print "write_data", " ".join(reversed(bits)) + print ""
5de36454db293c03039bec3c8c628995f07825b0
src/txkube/_compat.py
src/txkube/_compat.py
# Copyright Least Authority Enterprises. # See LICENSE for details. """ Helpers for Python 2/3 compatibility. """ from json import dumps from twisted.python.compat import unicode def dumps_bytes(obj): """ Serialize ``obj`` to JSON formatted ``bytes``. """ b = dumps(obj) if isinstance(b, unicode): b = b.encode("ascii") return b
Add a dumps_bytes() helper method to handle Python 2/3 compatibility.
Add a dumps_bytes() helper method to handle Python 2/3 compatibility.
Python
mit
LeastAuthority/txkube
--- +++ @@ -0,0 +1,19 @@ +# Copyright Least Authority Enterprises. +# See LICENSE for details. + +""" +Helpers for Python 2/3 compatibility. +""" + +from json import dumps + +from twisted.python.compat import unicode + +def dumps_bytes(obj): + """ + Serialize ``obj`` to JSON formatted ``bytes``. + """ + b = dumps(obj) + if isinstance(b, unicode): + b = b.encode("ascii") + return b
d7f184dd7c41bb3cacba5f77c81ae961b3a12760
subsample_bam_file.py
subsample_bam_file.py
#!/usr/bin/env python """ This script subsamples the alignments of a BAM file. For this a likelihood (0.0 < p(keep) < 1.0) of keeping all alignments of a read has to be provided. All alignments of a read are treated the same (i.e. are discarded or kept). """ import argparse import random import sys import pysam __description__ = "Subsample BAM file entries" __author__ = "Konrad Foerstner <konrad@foerstner.org>" __copyright__ = "2013 by Konrad Foerstner <konrad@foerstner.org>" __license__ = "ISC license" __email__ = "konrad@foerstner.org" __version__ = "0.1" parser = argparse.ArgumentParser() parser.add_argument("input_bam") parser.add_argument("output_bam") parser.add_argument("keeping_likelihood", type=float) args = parser.parse_args() input_bam = pysam.Samfile(args.input_bam, "rb") output_bam = pysam.Samfile( args.output_bam, "wb", referencenames=input_bam.references, referencelengths=input_bam.lengths, header=input_bam.header, text=input_bam.text) prev_query = None prev_keep = None for alignment in input_bam: # This is for reads that multiple alignments. If there previous # alignment comes from the same read treat the current one the # same way (keep or discard). if alignment.qname == prev_query: if prev_keep is True: output_bam.write(alignment) continue else: continue if random.random() <= args.keeping_likelihood: output_bam.write(alignment) prev_keep = True else: prev_keep = False prev_query = alignment.qname
Add script to subsample bam file entries
Add script to subsample bam file entries
Python
isc
konrad/kuf_bio_scripts
--- +++ @@ -0,0 +1,52 @@ +#!/usr/bin/env python +""" + +This script subsamples the alignments of a BAM file. For this a +likelihood (0.0 < p(keep) < 1.0) of keeping all alignments of a read +has to be provided. All alignments of a read are treated the same +(i.e. are discarded or kept). + +""" + +import argparse +import random +import sys +import pysam + +__description__ = "Subsample BAM file entries" +__author__ = "Konrad Foerstner <konrad@foerstner.org>" +__copyright__ = "2013 by Konrad Foerstner <konrad@foerstner.org>" +__license__ = "ISC license" +__email__ = "konrad@foerstner.org" +__version__ = "0.1" + +parser = argparse.ArgumentParser() +parser.add_argument("input_bam") +parser.add_argument("output_bam") +parser.add_argument("keeping_likelihood", type=float) +args = parser.parse_args() + +input_bam = pysam.Samfile(args.input_bam, "rb") +output_bam = pysam.Samfile( + args.output_bam, "wb", referencenames=input_bam.references, + referencelengths=input_bam.lengths, header=input_bam.header, + text=input_bam.text) + +prev_query = None +prev_keep = None +for alignment in input_bam: + # This is for reads that multiple alignments. If there previous + # alignment comes from the same read treat the current one the + # same way (keep or discard). + if alignment.qname == prev_query: + if prev_keep is True: + output_bam.write(alignment) + continue + else: + continue + if random.random() <= args.keeping_likelihood: + output_bam.write(alignment) + prev_keep = True + else: + prev_keep = False + prev_query = alignment.qname
67120e72883d4a5fd86dca2fba26599e65e7ea39
every_election/apps/elections/management/commands/add_referendum.py
every_election/apps/elections/management/commands/add_referendum.py
from django.core.management import BaseCommand from elections.models import ( Election, ElectionType, ModerationHistory, ModerationStatuses, ) from organisations.models.organisations import Organisation class Command(BaseCommand): help = """ Adds an election with an election type of referendum Example usage: python manage.py add_referendum --date 2021-10-07 \ --council croydon \ --election-title "Governance referendum" \ --official-identifier CRY \ --org-type local-authority """ def add_arguments(self, parser): parser.add_argument( "-d", "--date", action="store", dest="date", help="The date the referendum is taking place", required=True, ) parser.add_argument( "-c", "--council", action="store", dest="council", help="The council area the referendum is taking place", required=True, ) parser.add_argument( "-t", "--election-title", action="store", dest="election_title", help="The election title to be used", required=True, ) parser.add_argument( "--official-identifier", action="store", dest="official_identifier", help="The official identifier for the related Organisation. Election will cover the whole organisation unless a division ID is passed", required=True, ) parser.add_argument( "--org-type", action="store", dest="org_type", help="The council area the referendum is taking place", required=True, ) parser.add_argument( "--division-id", action="store", dest="division_id", help="ID of a OrganisationDivision if applicable", required=False, ) def handle(self, *args, **options): election_date = options["date"] election_id = f"ref.{options['council']}.{election_date}" referendum_type = ElectionType.objects.get(election_type="ref") ref_election, created = Election.private_objects.update_or_create( election_id=election_id, defaults={ "election_type": referendum_type, "poll_open_date": election_date, "election_title": options["election_title"], "current": True, "division_id": options["division_id"], }, ) self.stdout.write(f"{'Created' if created else 'Updated'} {election_id}") ModerationHistory.objects.get_or_create( status_id=ModerationStatuses.approved.value, election=ref_election, ) org = Organisation.objects.get_by_date( date=election_date, official_identifier=options["official_identifier"], organisation_type=options["org_type"], ) ref_election.organisation = org ref_election.organisation_geography = org.geographies.latest() if ref_election.division: ref_election.division_geography = ref_election.division.geography ref_election.save()
Add management command to add a refendeum election
Add management command to add a refendeum election - Added to allow us to add Croydon referendum initially, with the possibility of being able to use again in the future as required
Python
bsd-3-clause
DemocracyClub/EveryElection,DemocracyClub/EveryElection,DemocracyClub/EveryElection
--- +++ @@ -0,0 +1,103 @@ +from django.core.management import BaseCommand + +from elections.models import ( + Election, + ElectionType, + ModerationHistory, + ModerationStatuses, +) +from organisations.models.organisations import Organisation + + +class Command(BaseCommand): + help = """ + Adds an election with an election type of referendum + Example usage: + python manage.py add_referendum --date 2021-10-07 \ + --council croydon \ + --election-title "Governance referendum" \ + --official-identifier CRY \ + --org-type local-authority + """ + + def add_arguments(self, parser): + parser.add_argument( + "-d", + "--date", + action="store", + dest="date", + help="The date the referendum is taking place", + required=True, + ) + parser.add_argument( + "-c", + "--council", + action="store", + dest="council", + help="The council area the referendum is taking place", + required=True, + ) + parser.add_argument( + "-t", + "--election-title", + action="store", + dest="election_title", + help="The election title to be used", + required=True, + ) + parser.add_argument( + "--official-identifier", + action="store", + dest="official_identifier", + help="The official identifier for the related Organisation. Election will cover the whole organisation unless a division ID is passed", + required=True, + ) + parser.add_argument( + "--org-type", + action="store", + dest="org_type", + help="The council area the referendum is taking place", + required=True, + ) + parser.add_argument( + "--division-id", + action="store", + dest="division_id", + help="ID of a OrganisationDivision if applicable", + required=False, + ) + + def handle(self, *args, **options): + election_date = options["date"] + election_id = f"ref.{options['council']}.{election_date}" + + referendum_type = ElectionType.objects.get(election_type="ref") + ref_election, created = Election.private_objects.update_or_create( + election_id=election_id, + defaults={ + "election_type": referendum_type, + "poll_open_date": election_date, + "election_title": options["election_title"], + "current": True, + "division_id": options["division_id"], + }, + ) + + self.stdout.write(f"{'Created' if created else 'Updated'} {election_id}") + + ModerationHistory.objects.get_or_create( + status_id=ModerationStatuses.approved.value, + election=ref_election, + ) + + org = Organisation.objects.get_by_date( + date=election_date, + official_identifier=options["official_identifier"], + organisation_type=options["org_type"], + ) + ref_election.organisation = org + ref_election.organisation_geography = org.geographies.latest() + if ref_election.division: + ref_election.division_geography = ref_election.division.geography + + ref_election.save()
c008d60fda4f3e3d58d123ab4a972016e669a7dc
tests/test_summary_downloader.py
tests/test_summary_downloader.py
#!/usr/bin/env python # -*- coding: utf-8 -*- import os import itertools import tempfile from zipfile import ZipFile from utils.summary_downloader import SummaryDownloader def test_download_unzipped(): date = "Oct 24, 2016" tgt_dir = tempfile.mkdtemp(prefix='sdl_test_') prefixes = ["ES", "FC", "GS", "PL", "RO", "SS", "TH", "TV"] game_ids = ["020081", "020082"] # setting up list of all HTML report files that should be downloaded for # specified date files = ["".join(c) + ".HTM" for c in list( itertools.product(prefixes, game_ids))] # adding JSON game report files files.extend(["".join((gid, ".json")) for gid in game_ids]) # adding shootout report for one of the games files.append("SO020082.HTM") sdl = SummaryDownloader(tgt_dir, date, zip_summaries=False) sdl.run() tgt_dir = sdl.get_tgt_dir() assert sorted(os.listdir(tgt_dir)) == sorted(files) def test_download_zipped(): date = "Oct 24, 2016" tgt_dir = tempfile.mkdtemp(prefix='sdl_test_') prefixes = ["ES", "FC", "GS", "PL", "RO", "SS", "TH", "TV"] game_ids = ["020081", "020082"] # setting up list of all HTML report files that should be downloaded for # specified date files = ["".join(c) + ".HTM" for c in list( itertools.product(prefixes, game_ids))] # adding JSON game report files files.extend(["".join((gid, ".json")) for gid in game_ids]) # adding shootout report for one of the games files.append("SO020082.HTM") sdl = SummaryDownloader(tgt_dir, date) sdl.run() zip_path = sdl.get_zip_path() zip = ZipFile(zip_path) assert sorted(zip.namelist()) == sorted(files)
Add tests for download of game summaries
Add tests for download of game summaries
Python
mit
leaffan/pynhldb
--- +++ @@ -0,0 +1,57 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- + +import os +import itertools +import tempfile +from zipfile import ZipFile + +from utils.summary_downloader import SummaryDownloader + + +def test_download_unzipped(): + + date = "Oct 24, 2016" + tgt_dir = tempfile.mkdtemp(prefix='sdl_test_') + prefixes = ["ES", "FC", "GS", "PL", "RO", "SS", "TH", "TV"] + game_ids = ["020081", "020082"] + + # setting up list of all HTML report files that should be downloaded for + # specified date + files = ["".join(c) + ".HTM" for c in list( + itertools.product(prefixes, game_ids))] + # adding JSON game report files + files.extend(["".join((gid, ".json")) for gid in game_ids]) + # adding shootout report for one of the games + files.append("SO020082.HTM") + + sdl = SummaryDownloader(tgt_dir, date, zip_summaries=False) + sdl.run() + tgt_dir = sdl.get_tgt_dir() + + assert sorted(os.listdir(tgt_dir)) == sorted(files) + + +def test_download_zipped(): + + date = "Oct 24, 2016" + tgt_dir = tempfile.mkdtemp(prefix='sdl_test_') + prefixes = ["ES", "FC", "GS", "PL", "RO", "SS", "TH", "TV"] + game_ids = ["020081", "020082"] + + # setting up list of all HTML report files that should be downloaded for + # specified date + files = ["".join(c) + ".HTM" for c in list( + itertools.product(prefixes, game_ids))] + # adding JSON game report files + files.extend(["".join((gid, ".json")) for gid in game_ids]) + # adding shootout report for one of the games + files.append("SO020082.HTM") + + sdl = SummaryDownloader(tgt_dir, date) + sdl.run() + zip_path = sdl.get_zip_path() + + zip = ZipFile(zip_path) + + assert sorted(zip.namelist()) == sorted(files)
02a8a627bfffb1b0cda4455a2c639e92ef1d7a46
lib/portbuild/qthreads.py
lib/portbuild/qthreads.py
#! /usr/bin/env python import threading import time from pika.adapters import BlockingConnection class QueueThread(threading.Thread): def __init__(self): threading.Thread.__init__(self) self.name = self.__class__.__name__ self.connection = BlockingConnection() self._stop = threading.Event() self.setDaemon(True) def dprint(self, msg): print "%s: %s" % (self.name, msg) class QueueConsumerThread(QueueThread): def __init__(self): QueueThread.__init__(self) def run(self): self.setup() self.channel.start_consuming() def stop(self): self.channel.stop_consuming() self._stop.set() class QueueProducerThread(QueueThread): def __init__(self, freq=5): QueueThread.__init__(self) self.freq = freq def run(self): self.setup() # Need to find something better to avoid active loops. # Using signal.pause() when produce() is a no-op doesn't work. while True: self.produce() time.sleep(self.freq) def stop(self): self.connection.close() self._stop.set() # vim: tabstop=2 shiftwidth=2 softtabstop=2 expandtab
Add queue thread convenience classes.
Add queue thread convenience classes.
Python
bsd-2-clause
flz/portbuild-ng
--- +++ @@ -0,0 +1,48 @@ +#! /usr/bin/env python + +import threading +import time + +from pika.adapters import BlockingConnection + +class QueueThread(threading.Thread): + def __init__(self): + threading.Thread.__init__(self) + self.name = self.__class__.__name__ + self.connection = BlockingConnection() + self._stop = threading.Event() + self.setDaemon(True) + + def dprint(self, msg): + print "%s: %s" % (self.name, msg) + +class QueueConsumerThread(QueueThread): + def __init__(self): + QueueThread.__init__(self) + + def run(self): + self.setup() + self.channel.start_consuming() + + def stop(self): + self.channel.stop_consuming() + self._stop.set() + +class QueueProducerThread(QueueThread): + def __init__(self, freq=5): + QueueThread.__init__(self) + self.freq = freq + + def run(self): + self.setup() + # Need to find something better to avoid active loops. + # Using signal.pause() when produce() is a no-op doesn't work. + while True: + self.produce() + time.sleep(self.freq) + + def stop(self): + self.connection.close() + self._stop.set() + +# vim: tabstop=2 shiftwidth=2 softtabstop=2 expandtab
71b308f123aff1b4498fe14188380a6352d34d49
numba/tests/builtins/test_builtin_pow.py
numba/tests/builtins/test_builtin_pow.py
# adapted from cython/tests/run/builtin_pow.pyx """ >>> pow3(2,3,5) 3 >>> pow3(3,3,5) 2 >>> pow3_const() 3 >>> pow2(2,3) 8 >>> pow2(3,3) 27 >>> pow2_const() 8 """ @autojit(backend='ast') def pow3(a,b,c): return pow(a,b,c) @autojit(backend='ast') def pow3_const(): return pow(2,3,5) @autojit(backend='ast') def pow2(a,b): return pow(a,b) @autojit(backend='ast') def pow2_const(): return pow(2,3) if __name__ == '__main__': import doctest doctest.testmod()
Add testcase for pow builtin
Add testcase for pow builtin
Python
bsd-2-clause
stefanseefeld/numba,GaZ3ll3/numba,sklam/numba,gdementen/numba,numba/numba,gmarkall/numba,cpcloud/numba,gmarkall/numba,pitrou/numba,jriehl/numba,sklam/numba,pombredanne/numba,stonebig/numba,pitrou/numba,gdementen/numba,shiquanwang/numba,cpcloud/numba,stefanseefeld/numba,stefanseefeld/numba,pitrou/numba,stuartarchibald/numba,gdementen/numba,gdementen/numba,seibert/numba,seibert/numba,GaZ3ll3/numba,stonebig/numba,stuartarchibald/numba,seibert/numba,stuartarchibald/numba,sklam/numba,pitrou/numba,stonebig/numba,cpcloud/numba,numba/numba,seibert/numba,IntelLabs/numba,stonebig/numba,jriehl/numba,jriehl/numba,pombredanne/numba,gmarkall/numba,sklam/numba,IntelLabs/numba,GaZ3ll3/numba,gdementen/numba,sklam/numba,cpcloud/numba,IntelLabs/numba,seibert/numba,gmarkall/numba,gmarkall/numba,ssarangi/numba,stefanseefeld/numba,GaZ3ll3/numba,shiquanwang/numba,numba/numba,pombredanne/numba,pombredanne/numba,shiquanwang/numba,stefanseefeld/numba,numba/numba,ssarangi/numba,cpcloud/numba,pombredanne/numba,ssarangi/numba,IntelLabs/numba,ssarangi/numba,jriehl/numba,IntelLabs/numba,numba/numba,ssarangi/numba,jriehl/numba,stuartarchibald/numba,pitrou/numba,GaZ3ll3/numba,stuartarchibald/numba,stonebig/numba
--- +++ @@ -0,0 +1,39 @@ +# adapted from cython/tests/run/builtin_pow.pyx + +""" +>>> pow3(2,3,5) +3 +>>> pow3(3,3,5) +2 + +>>> pow3_const() +3 + +>>> pow2(2,3) +8 +>>> pow2(3,3) +27 + +>>> pow2_const() +8 +""" + +@autojit(backend='ast') +def pow3(a,b,c): + return pow(a,b,c) + +@autojit(backend='ast') +def pow3_const(): + return pow(2,3,5) + +@autojit(backend='ast') +def pow2(a,b): + return pow(a,b) + +@autojit(backend='ast') +def pow2_const(): + return pow(2,3) + +if __name__ == '__main__': + import doctest + doctest.testmod()
cf5a5594ebecc03c1087f09575dba1a480b575f3
tests/test_authjob.py
tests/test_authjob.py
from disco.test import DiscoJobTestFixture, DiscoTestCase from disco.ddfs import DDFS from disco.util import ddfs_name from cStringIO import StringIO class AuthJobTestCase(DiscoJobTestFixture, DiscoTestCase): input = [] @staticmethod def map(e, params): return [(e.strip(), '')] @property def answers(self): return [('blobdata', '')] def setUp(self): tag = 'disco:test:authjob' self.ddfs = DDFS(self.disco_master_url) pushed = self.ddfs.push(tag, [(StringIO('blobdata'), 'blob')]) self.ddfs.setattr(tag, 'ddfs:read-token', 'r') self.input = ['tag://u:r@/' + tag] super(AuthJobTestCase, self).setUp() def tearDown(self): super(AuthJobTestCase, self).tearDown() self.ddfs.delete('disco:test:authjob')
Add a test case for auth protected inputs.
Add a test case for auth protected inputs.
Python
bsd-3-clause
seabirdzh/disco,pombredanne/disco,ktkt2009/disco,ErikDubbelboer/disco,pooya/disco,pombredanne/disco,pooya/disco,oldmantaiter/disco,pooya/disco,seabirdzh/disco,beni55/disco,ktkt2009/disco,ErikDubbelboer/disco,oldmantaiter/disco,discoproject/disco,simudream/disco,mwilliams3/disco,pavlobaron/disco_playground,ktkt2009/disco,scrapinghub/disco,beni55/disco,pombredanne/disco,oldmantaiter/disco,discoproject/disco,ktkt2009/disco,mwilliams3/disco,simudream/disco,mwilliams3/disco,ErikDubbelboer/disco,pombredanne/disco,simudream/disco,pooya/disco,mwilliams3/disco,oldmantaiter/disco,seabirdzh/disco,ktkt2009/disco,seabirdzh/disco,pavlobaron/disco_playground,beni55/disco,mozilla/disco,ErikDubbelboer/disco,oldmantaiter/disco,discoproject/disco,mozilla/disco,pavlobaron/disco_playground,mwilliams3/disco,scrapinghub/disco,pavlobaron/disco_playground,discoproject/disco,beni55/disco,seabirdzh/disco,simudream/disco,discoproject/disco,simudream/disco,mozilla/disco,scrapinghub/disco,pombredanne/disco,scrapinghub/disco,mozilla/disco,beni55/disco,ErikDubbelboer/disco
--- +++ @@ -0,0 +1,29 @@ +from disco.test import DiscoJobTestFixture, DiscoTestCase + +from disco.ddfs import DDFS +from disco.util import ddfs_name + +from cStringIO import StringIO + +class AuthJobTestCase(DiscoJobTestFixture, DiscoTestCase): + input = [] + + @staticmethod + def map(e, params): + return [(e.strip(), '')] + + @property + def answers(self): + return [('blobdata', '')] + + def setUp(self): + tag = 'disco:test:authjob' + self.ddfs = DDFS(self.disco_master_url) + pushed = self.ddfs.push(tag, [(StringIO('blobdata'), 'blob')]) + self.ddfs.setattr(tag, 'ddfs:read-token', 'r') + self.input = ['tag://u:r@/' + tag] + super(AuthJobTestCase, self).setUp() + + def tearDown(self): + super(AuthJobTestCase, self).tearDown() + self.ddfs.delete('disco:test:authjob')
e6a57c43a693f71069778ddca3a28d5c73b75830
nova/scheduler/multi.py
nova/scheduler/multi.py
# vim: tabstop=4 shiftwidth=4 softtabstop=4 # Copyright (c) 2010 Openstack, LLC. # Copyright 2010 United States Government as represented by the # Administrator of the National Aeronautics and Space Administration. # All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. """ Scheduler that allows routing some calls to one driver and others to another. """ from nova import flags from nova import utils from nova.scheduler import driver FLAGS = flags.FLAGS flags.DEFINE_string('compute_scheduler_driver', 'nova.scheduler.chance.ChanceScheduler', 'Driver to use for scheduling compute calls') flags.DEFINE_string('volume_scheduler_driver', 'nova.scheduler.chance.ChanceScheduler', 'Driver to use for scheduling volume calls') # A mapping of methods to topics so we can figure out which driver to use. _METHOD_MAP = {'run_instance': 'compute', 'start_instance': 'compute', 'create_volume': 'volume'} class MultiScheduler(driver.Scheduler): """A scheduler that holds multiple sub-schedulers. This exists to allow flag-driven composibility of schedulers, allowing third parties to integrate custom schedulers more easily. """ def __init__(self): super(MultiScheduler, self).__init__() compute_driver = utils.import_object(FLAGS.compute_scheduler_driver) volume_driver = utils.import_object(FLAGS.volume_scheduler_driver) self.drivers = {'compute': compute_driver, 'volume': volume_driver} def __getattr__(self, key): if not key.startswith('schedule_'): raise AttributeError(key) method = key[len('schedule_'):] if method not in _METHOD_MAP: raise AttributeError(key) return getattr(self.drivers[_METHOD_MAP[method]], key) def set_zone_manager(self, zone_manager): for k, v in self.drivers.iteritems(): v.set_zone_manager(zone_manager) def schedule(self, context, topic, *_args, **_kwargs): return self.drivers[topic].schedule(context, topic, *_args, **_kwargs)
Allow different schedulers for compute and volume.
Allow different schedulers for compute and volume.
Python
apache-2.0
n0ano/ganttclient
--- +++ @@ -0,0 +1,73 @@ +# vim: tabstop=4 shiftwidth=4 softtabstop=4 + +# Copyright (c) 2010 Openstack, LLC. +# Copyright 2010 United States Government as represented by the +# Administrator of the National Aeronautics and Space Administration. +# All Rights Reserved. +# +# Licensed under the Apache License, Version 2.0 (the "License"); you may +# not use this file except in compliance with the License. You may obtain +# a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, WITHOUT +# WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the +# License for the specific language governing permissions and limitations +# under the License. + +""" +Scheduler that allows routing some calls to one driver and others to another. +""" + +from nova import flags +from nova import utils +from nova.scheduler import driver + + +FLAGS = flags.FLAGS +flags.DEFINE_string('compute_scheduler_driver', + 'nova.scheduler.chance.ChanceScheduler', + 'Driver to use for scheduling compute calls') +flags.DEFINE_string('volume_scheduler_driver', + 'nova.scheduler.chance.ChanceScheduler', + 'Driver to use for scheduling volume calls') + + +# A mapping of methods to topics so we can figure out which driver to use. +_METHOD_MAP = {'run_instance': 'compute', + 'start_instance': 'compute', + 'create_volume': 'volume'} + + +class MultiScheduler(driver.Scheduler): + """A scheduler that holds multiple sub-schedulers. + + This exists to allow flag-driven composibility of schedulers, allowing + third parties to integrate custom schedulers more easily. + + """ + + def __init__(self): + super(MultiScheduler, self).__init__() + compute_driver = utils.import_object(FLAGS.compute_scheduler_driver) + volume_driver = utils.import_object(FLAGS.volume_scheduler_driver) + + self.drivers = {'compute': compute_driver, + 'volume': volume_driver} + + def __getattr__(self, key): + if not key.startswith('schedule_'): + raise AttributeError(key) + method = key[len('schedule_'):] + if method not in _METHOD_MAP: + raise AttributeError(key) + return getattr(self.drivers[_METHOD_MAP[method]], key) + + def set_zone_manager(self, zone_manager): + for k, v in self.drivers.iteritems(): + v.set_zone_manager(zone_manager) + + def schedule(self, context, topic, *_args, **_kwargs): + return self.drivers[topic].schedule(context, topic, *_args, **_kwargs)
a32477899c7d21a82382f4c274cac129074b3e32
pombola/kenya/management/commands/kenya_import_2017_photos.py
pombola/kenya/management/commands/kenya_import_2017_photos.py
from __future__ import print_function import csv from os.path import abspath, dirname, exists, join import shutil from django.conf import settings from django.core.files.storage import FileSystemStorage from django.core.management.base import BaseCommand from images.models import Image from pombola.core.models import Person from pombola.core.utils import mkdir_p from PIL import Image as PillowImage import requests results_directory = abspath(join( dirname(__file__), '..', '..', 'election_data_2017', 'results' )) class Command(BaseCommand): help = 'Import photos for elected representatives from the 2017 election' def handle_person(self, row): person = Person.objects.get( identifiers__scheme='ynr-ke', identifiers__identifier=row['id']) image_url = row['image_url'] if not image_url: return # If we haven't already downloaded this file, download it: image_filename = join(self.cache_directory, str(row['id'])) if not exists(image_filename): r = requests.get(image_url, stream=True) r.raise_for_status() with open(image_filename, 'wb') as f: shutil.copyfileobj(r.raw, f) # Check this is a file type we can cope with: try: pillow_image = PillowImage.open(image_filename) except IOError as e: if 'cannot identify image file' in e.args[0]: print("Ignoring a non-image file {0}".format(image_filename)) return None raise if pillow_image.format not in ('PNG', 'JPEG'): raise Exception("Found an unsupported image format: {0}".format(pillow_image.format)) extension = {'PNG': 'png', 'JPEG': 'jpg'}[pillow_image.format] storage = FileSystemStorage() desired_storage_path = join( 'images', 'kenya-ynr', '{}.{}'.format(row['id'], extension)) with open(image_filename, 'rb') as f: storage_filename = storage.save(desired_storage_path, f) Image.objects.create( content_object=person, source='http://kenya.ynr.mysociety.org/person/{0}'.format(row['id']), is_primary=(not person.images.exists()), image=storage_filename, ) print("Created image for:", person) def handle(self, **options): self.cache_directory = join(results_directory, '.downloaded-images') mkdir_p(self.cache_directory) for filename in ('na.csv', 'senate.csv', 'wo.csv'): full_filename = join(results_directory, filename) with open(full_filename) as f: reader = csv.DictReader(f) for row in reader: self.handle_person(row)
Add a script to import elected candidates' photos from YNR
Add a script to import elected candidates' photos from YNR This script uses the YNR IDs that the kenya_import_2017_election_results command added to import the photos from YNR and add them to the appropriate person in Mzalendo. If there were existing photos of the person, the YNR photo is not set as primary, otherwise it would be.
Python
agpl-3.0
mysociety/pombola,mysociety/pombola,mysociety/pombola,mysociety/pombola,mysociety/pombola,mysociety/pombola
--- +++ @@ -0,0 +1,68 @@ +from __future__ import print_function + +import csv +from os.path import abspath, dirname, exists, join +import shutil + +from django.conf import settings +from django.core.files.storage import FileSystemStorage +from django.core.management.base import BaseCommand +from images.models import Image +from pombola.core.models import Person +from pombola.core.utils import mkdir_p +from PIL import Image as PillowImage +import requests + +results_directory = abspath(join( + dirname(__file__), '..', '..', 'election_data_2017', 'results' +)) + +class Command(BaseCommand): + help = 'Import photos for elected representatives from the 2017 election' + + def handle_person(self, row): + person = Person.objects.get( + identifiers__scheme='ynr-ke', identifiers__identifier=row['id']) + image_url = row['image_url'] + if not image_url: + return + # If we haven't already downloaded this file, download it: + image_filename = join(self.cache_directory, str(row['id'])) + if not exists(image_filename): + r = requests.get(image_url, stream=True) + r.raise_for_status() + with open(image_filename, 'wb') as f: + shutil.copyfileobj(r.raw, f) + # Check this is a file type we can cope with: + try: + pillow_image = PillowImage.open(image_filename) + except IOError as e: + if 'cannot identify image file' in e.args[0]: + print("Ignoring a non-image file {0}".format(image_filename)) + return None + raise + if pillow_image.format not in ('PNG', 'JPEG'): + raise Exception("Found an unsupported image format: {0}".format(pillow_image.format)) + extension = {'PNG': 'png', 'JPEG': 'jpg'}[pillow_image.format] + storage = FileSystemStorage() + desired_storage_path = join( + 'images', 'kenya-ynr', '{}.{}'.format(row['id'], extension)) + with open(image_filename, 'rb') as f: + storage_filename = storage.save(desired_storage_path, f) + Image.objects.create( + content_object=person, + source='http://kenya.ynr.mysociety.org/person/{0}'.format(row['id']), + is_primary=(not person.images.exists()), + image=storage_filename, + ) + print("Created image for:", person) + + def handle(self, **options): + self.cache_directory = join(results_directory, '.downloaded-images') + mkdir_p(self.cache_directory) + for filename in ('na.csv', 'senate.csv', 'wo.csv'): + full_filename = join(results_directory, filename) + with open(full_filename) as f: + reader = csv.DictReader(f) + for row in reader: + self.handle_person(row)
0aef588b92adc6ccc175a6b6d34784ff4d7e290d
coprocess/bindings/python/sample_server.py
coprocess/bindings/python/sample_server.py
import coprocess_object_pb2 import grpc, time _ONE_DAY_IN_SECONDS = 60 * 60 * 24 from concurrent import futures def MyPreMiddleware(coprocess_object): coprocess_object.request.set_headers["myheader"] = "myvalue" return coprocess_object def MyPostMiddleware(coprocess_object): coprocess_object.request.set_headers["anotherheader"] = "anothervalue" return coprocess_object class MyDispatcher(coprocess_object_pb2.DispatcherServicer): def Dispatch(self, coprocess_object, context): if coprocess_object.hook_name == "MyPreMiddleware": coprocess_object = MyPreMiddleware(coprocess_object) if coprocess_object.hook_name == "MyPostMiddleware": coprocess_object = MyPostMiddleware(coprocess_object) return coprocess_object def serve(): server = grpc.server(futures.ThreadPoolExecutor(max_workers=10)) coprocess_object_pb2.add_DispatcherServicer_to_server( MyDispatcher(), server) server.add_insecure_port('[::]:5555') server.start() try: while True: time.sleep(_ONE_DAY_IN_SECONDS) except KeyboardInterrupt: server.stop(0) if __name__ == '__main__': serve()
import coprocess_object_pb2 import grpc, time, json _ONE_DAY_IN_SECONDS = 60 * 60 * 24 from concurrent import futures def MyPreMiddleware(coprocess_object): coprocess_object.request.set_headers["myheader"] = "myvalue" return coprocess_object def MyPostMiddleware(coprocess_object): coprocess_object.request.set_headers["anotherheader"] = "anothervalue" return coprocess_object class MyDispatcher(coprocess_object_pb2.DispatcherServicer): def Dispatch(self, coprocess_object, context): if coprocess_object.hook_name == "MyPreMiddleware": coprocess_object = MyPreMiddleware(coprocess_object) if coprocess_object.hook_name == "MyPostMiddleware": coprocess_object = MyPostMiddleware(coprocess_object) return coprocess_object def DispatchEvent(self, event_wrapper, context): event = json.loads(event_wrapper.payload) print("DispatchEvent:", event) return coprocess_object_pb2.EventReply() def serve(): server = grpc.server(futures.ThreadPoolExecutor(max_workers=10)) coprocess_object_pb2.add_DispatcherServicer_to_server( MyDispatcher(), server) server.add_insecure_port('[::]:5555') server.start() try: while True: time.sleep(_ONE_DAY_IN_SECONDS) except KeyboardInterrupt: server.stop(0) if __name__ == '__main__': serve()
Adjust gRPC/Python sample to handle events.
Adjust gRPC/Python sample to handle events.
Python
mpl-2.0
mvdan/tyk,nebolsin/tyk,mvdan/tyk,nebolsin/tyk,nebolsin/tyk,mvdan/tyk,nebolsin/tyk,nebolsin/tyk,mvdan/tyk,lonelycode/tyk,lonelycode/tyk,nebolsin/tyk,nebolsin/tyk,lonelycode/tyk,mvdan/tyk,mvdan/tyk,nebolsin/tyk,mvdan/tyk,mvdan/tyk
--- +++ @@ -1,6 +1,6 @@ import coprocess_object_pb2 -import grpc, time +import grpc, time, json _ONE_DAY_IN_SECONDS = 60 * 60 * 24 @@ -16,7 +16,6 @@ class MyDispatcher(coprocess_object_pb2.DispatcherServicer): def Dispatch(self, coprocess_object, context): - if coprocess_object.hook_name == "MyPreMiddleware": coprocess_object = MyPreMiddleware(coprocess_object) @@ -24,6 +23,11 @@ coprocess_object = MyPostMiddleware(coprocess_object) return coprocess_object + + def DispatchEvent(self, event_wrapper, context): + event = json.loads(event_wrapper.payload) + print("DispatchEvent:", event) + return coprocess_object_pb2.EventReply() def serve():
3e45602583a7760a5fb7b5beb47405b4dddd0f63
run_tests.py
run_tests.py
#!/usr/bin/python import optparse import sys # Install the Python unittest2 package before you run this script. import unittest2 USAGE = """%prog SDK_PATH Run unit tests for App Engine apps. The SDK Path is probably /usr/local/google_appengine on Mac OS SDK_PATH Path to the SDK installation""" def main(sdk_path, test_pattern): sys.path.insert(0, sdk_path) import dev_appserver dev_appserver.fix_sys_path() suite = unittest2.loader.TestLoader().discover("tests", test_pattern) unittest2.TextTestRunner(verbosity=2).run(suite) if __name__ == '__main__': parser = optparse.OptionParser(USAGE) options, args = parser.parse_args() if len(args) < 1: print 'Warning: Trying default SDK path.' sdk_path = "/usr/local/google_appengine" else: sdk_path = args[0] test_pattern = "test*.py" if len(args) > 1: test_pattern = args[1] main(sdk_path, test_pattern)
#!/usr/bin/python import optparse import sys # Install the Python unittest2 package before you run this script. import unittest2 USAGE = """%prog SDK_PATH Run unit tests for App Engine apps. The SDK Path is probably /usr/local/google_appengine on Mac OS SDK_PATH Path to the SDK installation""" def main(sdk_path, test_pattern): sys.path.insert(0, sdk_path) import dev_appserver dev_appserver.fix_sys_path() suite = unittest2.loader.TestLoader().discover("tests", test_pattern) tests = unittest2.TextTestRunner(verbosity=2).run(suite) if tests.wasSuccessful() == True: sys.exit(0) else: sys.exit(1) if __name__ == '__main__': parser = optparse.OptionParser(USAGE) options, args = parser.parse_args() if len(args) < 1: print 'Warning: Trying default SDK path.' sdk_path = "/usr/local/google_appengine" else: sdk_path = args[0] test_pattern = "test*.py" if len(args) > 1: test_pattern = args[1] main(sdk_path, test_pattern)
Exit with code 1 if tests fail.
Exit with code 1 if tests fail. Fixes #621 and Travis.
Python
mit
josephbisch/the-blue-alliance,1fish2/the-blue-alliance,nwalters512/the-blue-alliance,verycumbersome/the-blue-alliance,phil-lopreiato/the-blue-alliance,fangeugene/the-blue-alliance,the-blue-alliance/the-blue-alliance,1fish2/the-blue-alliance,bdaroz/the-blue-alliance,1fish2/the-blue-alliance,jaredhasenklein/the-blue-alliance,verycumbersome/the-blue-alliance,phil-lopreiato/the-blue-alliance,jaredhasenklein/the-blue-alliance,synth3tk/the-blue-alliance,jaredhasenklein/the-blue-alliance,josephbisch/the-blue-alliance,nwalters512/the-blue-alliance,phil-lopreiato/the-blue-alliance,phil-lopreiato/the-blue-alliance,bdaroz/the-blue-alliance,bdaroz/the-blue-alliance,tsteward/the-blue-alliance,fangeugene/the-blue-alliance,verycumbersome/the-blue-alliance,1fish2/the-blue-alliance,jaredhasenklein/the-blue-alliance,bvisness/the-blue-alliance,bdaroz/the-blue-alliance,bdaroz/the-blue-alliance,josephbisch/the-blue-alliance,bvisness/the-blue-alliance,tsteward/the-blue-alliance,synth3tk/the-blue-alliance,nwalters512/the-blue-alliance,synth3tk/the-blue-alliance,josephbisch/the-blue-alliance,the-blue-alliance/the-blue-alliance,nwalters512/the-blue-alliance,nwalters512/the-blue-alliance,verycumbersome/the-blue-alliance,josephbisch/the-blue-alliance,tsteward/the-blue-alliance,synth3tk/the-blue-alliance,the-blue-alliance/the-blue-alliance,fangeugene/the-blue-alliance,tsteward/the-blue-alliance,synth3tk/the-blue-alliance,tsteward/the-blue-alliance,bvisness/the-blue-alliance,fangeugene/the-blue-alliance,the-blue-alliance/the-blue-alliance,phil-lopreiato/the-blue-alliance,synth3tk/the-blue-alliance,1fish2/the-blue-alliance,nwalters512/the-blue-alliance,bvisness/the-blue-alliance,bdaroz/the-blue-alliance,the-blue-alliance/the-blue-alliance,fangeugene/the-blue-alliance,verycumbersome/the-blue-alliance,jaredhasenklein/the-blue-alliance,verycumbersome/the-blue-alliance,1fish2/the-blue-alliance,tsteward/the-blue-alliance,fangeugene/the-blue-alliance,josephbisch/the-blue-alliance,phil-lopreiato/the-blue-alliance,jaredhasenklein/the-blue-alliance,bvisness/the-blue-alliance,bvisness/the-blue-alliance,the-blue-alliance/the-blue-alliance
--- +++ @@ -15,9 +15,14 @@ sys.path.insert(0, sdk_path) import dev_appserver dev_appserver.fix_sys_path() - + suite = unittest2.loader.TestLoader().discover("tests", test_pattern) - unittest2.TextTestRunner(verbosity=2).run(suite) + tests = unittest2.TextTestRunner(verbosity=2).run(suite) + + if tests.wasSuccessful() == True: + sys.exit(0) + else: + sys.exit(1) if __name__ == '__main__': @@ -28,9 +33,9 @@ sdk_path = "/usr/local/google_appengine" else: sdk_path = args[0] - + test_pattern = "test*.py" if len(args) > 1: test_pattern = args[1] - + main(sdk_path, test_pattern)
c881b3176b122cdefe406d02e640600e6d4f4727
migrations/versions/ec3035d61f8b_.py
migrations/versions/ec3035d61f8b_.py
"""Add status column to tickets. Revision ID: ec3035d61f8b Revises: 0397e48c5db8 Create Date: 2020-08-30 21:04:30.936267 """ from alembic import op import sqlalchemy as sa # revision identifiers, used by Alembic. revision = 'ec3035d61f8b' down_revision = '0397e48c5db8' branch_labels = None depends_on = None def upgrade(): try: op.add_column('serials', sa.Column('status', sa.String(length=10), nullable=True)) except Exception: pass def downgrade(): with op.batch_alter_table('serials') as batch: batch.drop_column('status')
Add ticket's status column migration.
Add ticket's status column migration.
Python
mpl-2.0
mrf345/FQM,mrf345/FQM,mrf345/FQM,mrf345/FQM
--- +++ @@ -0,0 +1,28 @@ +"""Add status column to tickets. + +Revision ID: ec3035d61f8b +Revises: 0397e48c5db8 +Create Date: 2020-08-30 21:04:30.936267 + +""" +from alembic import op +import sqlalchemy as sa + + +# revision identifiers, used by Alembic. +revision = 'ec3035d61f8b' +down_revision = '0397e48c5db8' +branch_labels = None +depends_on = None + + +def upgrade(): + try: + op.add_column('serials', sa.Column('status', sa.String(length=10), nullable=True)) + except Exception: + pass + + +def downgrade(): + with op.batch_alter_table('serials') as batch: + batch.drop_column('status')
9d501aaa9d443e450224388854f053683ca2baae
permute/tests/test_npc.py
permute/tests/test_npc.py
from __future__ import division, print_function, absolute_import from nose.plugins.attrib import attr from nose.tools import assert_raises, raises import numpy as np from scipy.stats import norm from ..npc import (fisher, liptak, tippett) def test_fisher(): pvalues = np.linspace(0.05, 0.9, num=5) np.testing.assert_almost_equal(fisher(pvalues), 11.11546, 5) np.testing.assert_equal(fisher(1), -0.0) np.testing.assert_array_less(fisher(10), 0) def test_liptak(): pvalues = np.linspace(0.05, 0.9, num=5) np.testing.assert_almost_equal(liptak(pvalues), 0.5728894, 5) np.testing.assert_equal(liptak(1), norm.ppf(0)) np.testing.assert_equal(liptak(10), np.nan) def test_tippett(): pvalues = np.linspace(0.05, 0.9, num=5) np.testing.assert_almost_equal(tippett(pvalues), 0.95, 5) np.testing.assert_equal(tippett(1), 0) np.testing.assert_equal(tippett(10), -9)
Add tests for combining functions
TST: Add tests for combining functions
Python
bsd-2-clause
qqqube/permute,statlab/permute,kellieotto/permute,jarrodmillman/permute,kellieotto/permute
--- +++ @@ -0,0 +1,32 @@ +from __future__ import division, print_function, absolute_import + +from nose.plugins.attrib import attr +from nose.tools import assert_raises, raises + +import numpy as np +from scipy.stats import norm + +from ..npc import (fisher, + liptak, + tippett) + +def test_fisher(): + pvalues = np.linspace(0.05, 0.9, num=5) + np.testing.assert_almost_equal(fisher(pvalues), 11.11546, 5) + np.testing.assert_equal(fisher(1), -0.0) + np.testing.assert_array_less(fisher(10), 0) + + +def test_liptak(): + pvalues = np.linspace(0.05, 0.9, num=5) + np.testing.assert_almost_equal(liptak(pvalues), 0.5728894, 5) + np.testing.assert_equal(liptak(1), norm.ppf(0)) + np.testing.assert_equal(liptak(10), np.nan) + + +def test_tippett(): + pvalues = np.linspace(0.05, 0.9, num=5) + np.testing.assert_almost_equal(tippett(pvalues), 0.95, 5) + np.testing.assert_equal(tippett(1), 0) + np.testing.assert_equal(tippett(10), -9) +
827802f111fbe5c2f9a50b74a46afdef5eae2a2d
salt/pillar/nacl.py
salt/pillar/nacl.py
# -*- coding: utf-8 -*- ''' Decrypt pillar data through the builtin NACL renderer In most cases, you'll want to make this the last external pillar used. For example, to pair with the builtin stack pillar you could do something like this: .. code:: yaml nacl.config: keyfile: /root/.nacl ext_pillar: - stack: /path/to/stack.cfg - nacl: {} Set ``nacl.config`` in your config. ''' from __future__ import absolute_import import salt def ext_pillar(minion_id, pillar, *args, **kwargs): render_function = salt.loader.render(__opts__, __salt__).get("nacl") return render_function(pillar)
Decrypt NACL passwords on ext_pillar
Decrypt NACL passwords on ext_pillar
Python
apache-2.0
saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt
--- +++ @@ -0,0 +1,29 @@ +# -*- coding: utf-8 -*- + +''' +Decrypt pillar data through the builtin NACL renderer + +In most cases, you'll want to make this the last external pillar used. For +example, to pair with the builtin stack pillar you could do something like +this: + +.. code:: yaml + + nacl.config: + keyfile: /root/.nacl + + ext_pillar: + - stack: /path/to/stack.cfg + - nacl: {} + +Set ``nacl.config`` in your config. + +''' + +from __future__ import absolute_import +import salt + + +def ext_pillar(minion_id, pillar, *args, **kwargs): + render_function = salt.loader.render(__opts__, __salt__).get("nacl") + return render_function(pillar)
ce8dc2152a508ea359159a917ea469a106065503
modules/neural_network.py
modules/neural_network.py
#!/usr/bin/env python2 # -*- coding: utf-8 -*- """ Created on Thu Apr 27 14:10:49 2017 @author: daniele """ import tensorflow as tf import os import numpy as np import matplotlib.pyplot as plt
Add (currently empty) neural netowrks module
feat: Add (currently empty) neural netowrks module
Python
mit
dangall/Kaggle-MobileODT-Cancer-Screening
--- +++ @@ -0,0 +1,17 @@ +#!/usr/bin/env python2 +# -*- coding: utf-8 -*- +""" +Created on Thu Apr 27 14:10:49 2017 + +@author: daniele +""" + +import tensorflow as tf + +import os + +import numpy as np +import matplotlib.pyplot as plt + + +
f8b6d9a1bb12087c25e60df03f9f40c435f1a949
tests/test_archive.py
tests/test_archive.py
#!/usr/bin/env python # -*- coding: utf8 -*- try: import unittest2 as unittest except ImportError: import unittest import agate import agateremote class TestArchive(agate.AgateTestCase): def setUp(self): self.archive = agateremote.Archive('https://github.com/vincentarelbundock/Rdatasets/raw/master/csv/') def test_get_table(self): table = self.archive.get_table('sandwich/PublicSchools.csv') self.assertColumnNames(table, ['', 'Expenditure', 'Income']) self.assertColumnTypes(table, [agate.Text, agate.Number, agate.Number]) self.assertEqual(len(table.rows), 51)
#!/usr/bin/env python # -*- coding: utf8 -*- try: import unittest2 as unittest except ImportError: import unittest import agate import agateremote class TestArchive(agate.AgateTestCase): def setUp(self): self.archive = agateremote.Archive('https://github.com/vincentarelbundock/Rdatasets/raw/master/csv/') def test_get_table(self): table = self.archive.get_table('sandwich/PublicSchools.csv') self.assertColumnNames(table, ('a', 'Expenditure', 'Income')) self.assertColumnTypes(table, [agate.Text, agate.Number, agate.Number]) self.assertEqual(len(table.rows), 51)
Update test for recent version of agate
Update test for recent version of agate
Python
mit
wireservice/agate-remote,onyxfish/agate-remote
--- +++ @@ -16,6 +16,6 @@ def test_get_table(self): table = self.archive.get_table('sandwich/PublicSchools.csv') - self.assertColumnNames(table, ['', 'Expenditure', 'Income']) + self.assertColumnNames(table, ('a', 'Expenditure', 'Income')) self.assertColumnTypes(table, [agate.Text, agate.Number, agate.Number]) self.assertEqual(len(table.rows), 51)
533160e303fd73a8e6f53ebc7e6430bba8888bbb
tests/test_player_draft.py
tests/test_player_draft.py
#!/usr/bin/env python # -*- coding: utf-8 -*- from db.player_draft import PlayerDraft from db.team import Team def test_find_by_id(): pdft = PlayerDraft.find_by_player_id(8475883) # Frederik Andersen assert len(pdft) == 2 pdft = PlayerDraft.find_by_player_id(8466145) # Nick Boynton assert len(pdft) == 2 def test_find(): pdft = PlayerDraft.find(8479318, 10, 2016) # Auston Matthews assert pdft.round == 1 assert pdft.overall == 1 def test_constructor(): pdft = PlayerDraft(8999444, 1, 2018, 3, 75) # fictional player assert pdft.player_id == 8999444 assert Team.find_by_id(pdft.team_id).name == 'New Jersey Devils' assert pdft.year == 2018 assert pdft.round == 3 assert pdft.overall == 75 def test_comparison_operators(): pdft_kopitar = PlayerDraft.find_by_player_id(8471685).pop(0) # 2005, 11 pdft_toews = PlayerDraft.find_by_player_id(8473604).pop(0) # 2006, 3 pdft_kessel = PlayerDraft.find_by_player_id(8473548).pop(0) # 2006, 5 pdft_stamkos = PlayerDraft.find_by_player_id(8474564).pop(0) # 2008, 1 ordered = sorted([pdft_kessel, pdft_kopitar, pdft_stamkos, pdft_toews]) assert ordered[0] == pdft_kopitar assert ordered[1] == pdft_toews assert ordered[2] == pdft_kessel assert ordered[3] == pdft_stamkos
Add tests for player draft items
Add tests for player draft items
Python
mit
leaffan/pynhldb
--- +++ @@ -0,0 +1,39 @@ +#!/usr/bin/env python +# -*- coding: utf-8 -*- + +from db.player_draft import PlayerDraft +from db.team import Team + + +def test_find_by_id(): + pdft = PlayerDraft.find_by_player_id(8475883) # Frederik Andersen + assert len(pdft) == 2 + pdft = PlayerDraft.find_by_player_id(8466145) # Nick Boynton + assert len(pdft) == 2 + + +def test_find(): + pdft = PlayerDraft.find(8479318, 10, 2016) # Auston Matthews + assert pdft.round == 1 + assert pdft.overall == 1 + + +def test_constructor(): + pdft = PlayerDraft(8999444, 1, 2018, 3, 75) # fictional player + assert pdft.player_id == 8999444 + assert Team.find_by_id(pdft.team_id).name == 'New Jersey Devils' + assert pdft.year == 2018 + assert pdft.round == 3 + assert pdft.overall == 75 + + +def test_comparison_operators(): + pdft_kopitar = PlayerDraft.find_by_player_id(8471685).pop(0) # 2005, 11 + pdft_toews = PlayerDraft.find_by_player_id(8473604).pop(0) # 2006, 3 + pdft_kessel = PlayerDraft.find_by_player_id(8473548).pop(0) # 2006, 5 + pdft_stamkos = PlayerDraft.find_by_player_id(8474564).pop(0) # 2008, 1 + ordered = sorted([pdft_kessel, pdft_kopitar, pdft_stamkos, pdft_toews]) + assert ordered[0] == pdft_kopitar + assert ordered[1] == pdft_toews + assert ordered[2] == pdft_kessel + assert ordered[3] == pdft_stamkos
517c29e34a1bc43df60cbabad998a1c0581e7b21
pybossa/error/__init__.py
pybossa/error/__init__.py
# -*- coding: utf8 -*- # This file is part of PyBossa. # # Copyright (C) 2013 SF Isle of Man Limited # # PyBossa is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # PyBossa is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with PyBossa. If not, see <http://www.gnu.org/licenses/>. """ PyBossa error module for processing error status. This package adds GET, POST, PUT and DELETE errors for the API: * applications, * tasks and * task_runs """ import json from flask import Response class ErrorStatus(object): """ Class for formatting error status in JSON format. This class has the following methods: * format_exception: returns a Flask Response with the error. """ error_status = {"Forbidden": 403, "NotFound": 404, "Unauthorized": 401, "TypeError": 415, "ValueError": 415, "DataError": 415, "AttributeError": 415, "IntegrityError": 415, "TooManyRequests": 429} def format_exception(self, e, target, action): """ Format the exception to a valid JSON object. Returns a Flask Response with the error. """ exception_cls = e.__class__.__name__ if self.error_status.get(exception_cls): status = self.error_status.get(exception_cls) else: status = 200 error = dict(action=action.upper(), status="failed", status_code=status, target=target, exception_cls=exception_cls, exception_msg=e.message) return Response(json.dumps(error), status=status, mimetype='application/json')
Refactor HTTP Error status in a module
Refactor HTTP Error status in a module
Python
agpl-3.0
stefanhahmann/pybossa,jean/pybossa,geotagx/geotagx-pybossa-archive,harihpr/tweetclickers,OpenNewsLabs/pybossa,inteligencia-coletiva-lsd/pybossa,geotagx/geotagx-pybossa-archive,inteligencia-coletiva-lsd/pybossa,geotagx/pybossa,OpenNewsLabs/pybossa,PyBossa/pybossa,harihpr/tweetclickers,geotagx/pybossa,stefanhahmann/pybossa,geotagx/geotagx-pybossa-archive,CulturePlex/pybossa,geotagx/geotagx-pybossa-archive,geotagx/geotagx-pybossa-archive,proyectos-analizo-info/pybossa-analizo-info,Scifabric/pybossa,CulturePlex/pybossa,proyectos-analizo-info/pybossa-analizo-info,Scifabric/pybossa,proyectos-analizo-info/pybossa-analizo-info,CulturePlex/pybossa,jean/pybossa,PyBossa/pybossa
--- +++ @@ -0,0 +1,70 @@ +# -*- coding: utf8 -*- +# This file is part of PyBossa. +# +# Copyright (C) 2013 SF Isle of Man Limited +# +# PyBossa is free software: you can redistribute it and/or modify +# it under the terms of the GNU Affero General Public License as published by +# the Free Software Foundation, either version 3 of the License, or +# (at your option) any later version. +# +# PyBossa is distributed in the hope that it will be useful, +# but WITHOUT ANY WARRANTY; without even the implied warranty of +# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the +# GNU Affero General Public License for more details. +# +# You should have received a copy of the GNU Affero General Public License +# along with PyBossa. If not, see <http://www.gnu.org/licenses/>. +""" +PyBossa error module for processing error status. + +This package adds GET, POST, PUT and DELETE errors for the API: + * applications, + * tasks and + * task_runs + +""" +import json +from flask import Response + + +class ErrorStatus(object): + + """ + Class for formatting error status in JSON format. + + This class has the following methods: + * format_exception: returns a Flask Response with the error. + + """ + + error_status = {"Forbidden": 403, + "NotFound": 404, + "Unauthorized": 401, + "TypeError": 415, + "ValueError": 415, + "DataError": 415, + "AttributeError": 415, + "IntegrityError": 415, + "TooManyRequests": 429} + + def format_exception(self, e, target, action): + """ + Format the exception to a valid JSON object. + + Returns a Flask Response with the error. + + """ + exception_cls = e.__class__.__name__ + if self.error_status.get(exception_cls): + status = self.error_status.get(exception_cls) + else: + status = 200 + error = dict(action=action.upper(), + status="failed", + status_code=status, + target=target, + exception_cls=exception_cls, + exception_msg=e.message) + return Response(json.dumps(error), status=status, + mimetype='application/json')
17c69212ff5ea1a5991dc89ba1ca7365c8d666c2
firmware/common/binsize.py
firmware/common/binsize.py
#!/usr/bin/env python from subprocess import Popen, PIPE class app(object): def __init__(self): pass motolink = app() motolink.name = "ch" motolink.path = "build/ch.elf" motolink.max_ccm = 4*1024 motolink.max_ram = 12*1024 motolink.max_rom = 64*1024 APPS = [motolink] for app in APPS: ccm = 0 ram = 0 rom = 0 p = Popen(["arm-none-eabi-size", "-A", app.path], stdout=PIPE) if p.wait() == 0: output = p.stdout.read() lines = filter(None, output.split("\n")) for line in lines: columns = filter(None, line.split(" ")) if ".stacks" in columns[0]: ram += int(columns[1]) elif ".ram4" in columns[0]: ccm += int(columns[1]) rom += int(columns[1]) elif ".bss" in columns[0]: ram += int(columns[1]) elif ".data" in columns[0]: ram += int(columns[1]) rom += int(columns[1]) elif ".text" in columns[0]: rom += int(columns[1]) elif ".startup" in columns[0]: rom += int(columns[1]) print "" print app.name print "CCM used: {}% - {:4.1f}/{}k".format((ccm*100)/app.max_ccm, ccm/1024.0, app.max_ccm/1024.0) print "RAM used: {}% - {:4.1f}/{}k".format((ram*100)/app.max_ram, ram/1024.0, app.max_ram/1024.0) print "ROM used: {}% - {:4.1f}/{}k".format((rom*100)/app.max_rom, rom/1024.0, app.max_rom/1024.0)
Add a Python script to get ROM and RAM usage.
Add a Python script to get ROM and RAM usage.
Python
apache-2.0
romainreignier/robot2017,romainreignier/robot2017,romainreignier/robot2017,romainreignier/robot2017
--- +++ @@ -0,0 +1,62 @@ +#!/usr/bin/env python + +from subprocess import Popen, PIPE + + +class app(object): + def __init__(self): + pass + + +motolink = app() +motolink.name = "ch" +motolink.path = "build/ch.elf" +motolink.max_ccm = 4*1024 +motolink.max_ram = 12*1024 +motolink.max_rom = 64*1024 + +APPS = [motolink] + +for app in APPS: + + ccm = 0 + ram = 0 + rom = 0 + + p = Popen(["arm-none-eabi-size", "-A", app.path], stdout=PIPE) + + if p.wait() == 0: + + output = p.stdout.read() + lines = filter(None, output.split("\n")) + + for line in lines: + columns = filter(None, line.split(" ")) + if ".stacks" in columns[0]: + ram += int(columns[1]) + elif ".ram4" in columns[0]: + ccm += int(columns[1]) + rom += int(columns[1]) + elif ".bss" in columns[0]: + ram += int(columns[1]) + elif ".data" in columns[0]: + ram += int(columns[1]) + rom += int(columns[1]) + elif ".text" in columns[0]: + rom += int(columns[1]) + elif ".startup" in columns[0]: + rom += int(columns[1]) + + print "" + print app.name + print "CCM used: {}% - {:4.1f}/{}k".format((ccm*100)/app.max_ccm, + ccm/1024.0, + app.max_ccm/1024.0) + + print "RAM used: {}% - {:4.1f}/{}k".format((ram*100)/app.max_ram, + ram/1024.0, + app.max_ram/1024.0) + + print "ROM used: {}% - {:4.1f}/{}k".format((rom*100)/app.max_rom, + rom/1024.0, + app.max_rom/1024.0)
4f189c20fc535f780e905e6fee7351329240d8fc
tyr/clusters/mongo.py
tyr/clusters/mongo.py
import logging class MongoCluster(object): log = logging.getLogger('Clusters.Mongo') log.setLevel(logging.DEBUG) ch = logging.StreamHandler() ch.setLevel(logging.DEBUG) formatter = logging.Formatter( '%(asctime)s [%(name)s] %(levelname)s: %(message)s', datefmt = '%H:%M:%S') ch.setFormatter(formatter) log.addHandler(ch)
Set up logging for a MongoDB cluster
Set up logging for a MongoDB cluster
Python
unlicense
hudl/Tyr
--- +++ @@ -0,0 +1,14 @@ +import logging + +class MongoCluster(object): + + log = logging.getLogger('Clusters.Mongo') + log.setLevel(logging.DEBUG) + ch = logging.StreamHandler() + ch.setLevel(logging.DEBUG) + formatter = logging.Formatter( + '%(asctime)s [%(name)s] %(levelname)s: %(message)s', + datefmt = '%H:%M:%S') + ch.setFormatter(formatter) + log.addHandler(ch) +
feafbefbc7de1f71cfa50af51139c4a09a306012
socketio-client.py
socketio-client.py
import websocket, httplib, sys, asyncore ''' connect to the socketio server 1. perform the HTTP handshake 2. open a websocket connection ''' def connect(server, port): print("connecting to: %s:%d" %(server, port)) conn = httplib.HTTPConnection(server + ":" + str(port)) conn.request('POST','/socket.io/1/') resp = conn.getresponse() hskey = resp.read().split(':')[0] ws = websocket.WebSocket( 'ws://'+server+':'+str(port)+'/socket.io/1/websocket/'+hskey, onopen = _onopen, onmessage = _onmessage, onclose = _onclose) return ws def _onopen(): print("opened!") def _onmessage(msg): print("msg: " + str(msg)) def _onclose(): print("closed!") if __name__ == '__main__': if len(sys.argv) != 3: sys.stderr.write('usage: python client.py <server> <port>\n') sys.exit(1) server = sys.argv[1] port = int(sys.argv[2]) ws = connect(server, port) try: asyncore.loop() except KeyboardInterrupt: ws.close()
Add socket io client example
Add socket io client example
Python
mit
voidabhi/python-scripts,voidabhi/python-scripts,voidabhi/python-scripts,voidabhi/python-scripts,voidabhi/python-scripts
--- +++ @@ -0,0 +1,49 @@ +import websocket, httplib, sys, asyncore + +''' + connect to the socketio server + 1. perform the HTTP handshake + 2. open a websocket connection ''' +def connect(server, port): + + print("connecting to: %s:%d" %(server, port)) + + conn = httplib.HTTPConnection(server + ":" + str(port)) + conn.request('POST','/socket.io/1/') + resp = conn.getresponse() + hskey = resp.read().split(':')[0] + + ws = websocket.WebSocket( + 'ws://'+server+':'+str(port)+'/socket.io/1/websocket/'+hskey, + onopen = _onopen, + onmessage = _onmessage, + onclose = _onclose) + + return ws + +def _onopen(): + print("opened!") + +def _onmessage(msg): + print("msg: " + str(msg)) + +def _onclose(): + print("closed!") + + +if __name__ == '__main__': + if len(sys.argv) != 3: + sys.stderr.write('usage: python client.py <server> <port>\n') + sys.exit(1) + + server = sys.argv[1] + port = int(sys.argv[2]) + + ws = connect(server, port) + + try: + asyncore.loop() + except KeyboardInterrupt: + ws.close() + +
c56efa68fbc878d307391910de9c521e5f2a6673
python/dynamic_class.py
python/dynamic_class.py
# -*- coding:utf-8 -*- class CanDoSomething(object): message = "You don't have permission" def has_permission(self, request, view): return True class CanDoSomethingOrReadOnly(CanDoSomething): def has_permission(self, request, view): return False class HasPermissionToDo(object): p_class = CanDoSomething name = "HasPermissionToDo%s" def __init__(self, permission_code, *args, **kwargs): pass def __new__(cls, permission_code, *args, **kwargs): code = permission_code.split('.')[1] name = ''.join(x.title() for x in code.split('_')) cls.p_class.message= "hello world" cls.p_class.__name__ = cls.name % name return cls.p_class class HasPermissionToDoOrReadOnly(HasPermissionToDo): p_class = CanDoSomethingOrReadOnly name = "HasPermissionToDo%sOrReadOnly"
Add python dynamic class sample
Add python dynamic class sample
Python
mit
aiden0z/snippets,aiden0z/snippets,aiden0z/snippets,aiden0z/snippets,aiden0z/snippets,aiden0z/snippets
--- +++ @@ -0,0 +1,36 @@ +# -*- coding:utf-8 -*- + + +class CanDoSomething(object): + + message = "You don't have permission" + + def has_permission(self, request, view): + return True + + +class CanDoSomethingOrReadOnly(CanDoSomething): + + def has_permission(self, request, view): + return False + + +class HasPermissionToDo(object): + + p_class = CanDoSomething + name = "HasPermissionToDo%s" + + def __init__(self, permission_code, *args, **kwargs): + pass + + def __new__(cls, permission_code, *args, **kwargs): + code = permission_code.split('.')[1] + name = ''.join(x.title() for x in code.split('_')) + cls.p_class.message= "hello world" + cls.p_class.__name__ = cls.name % name + return cls.p_class + + +class HasPermissionToDoOrReadOnly(HasPermissionToDo): + p_class = CanDoSomethingOrReadOnly + name = "HasPermissionToDo%sOrReadOnly"
1dd8120dab6cdec4097bea1193a5b5b68d3bfe4f
salt/modules/win_groupadd.py
salt/modules/win_groupadd.py
''' Manage groups on Windows ''' def __virtual__(): ''' Set the group module if the kernel is Windows ''' return 'group' if __grains__['kernel'] == 'Windows' else False def add(name): ''' Add the specified group CLI Example:: salt '*' group.add foo ''' cmd = 'net localgroup {0} /add'.format(name) ret = __salt__['cmd.run_all'](cmd) return not ret['retcode'] def delete(name): ''' Remove the named group CLI Example:: salt '*' group.delete foo ''' ret = __salt__['cmd.run_all']('net localgroup {0} /delete'.format(name)) return not ret['retcode'] def info(name): ''' Return information about a group CLI Example:: salt '*' group.info foo ''' lines = __salt__['cmd.run']('net localgroup {0}'.format(name)).split('\n') memberline = False gr_mem = [] gr_name = '' for line in lines: if 'Alias name' in line: comps = line.split(' ', 1) gr_name = comps[1].strip() if 'successfully' in line: memberline = False if memberline: gr_mem.append(line.strip()) if '---' in line: memberline = True if not gr_name: return False return {'name': gr_name, 'passwd': None, 'gid': None, 'members': gr_mem} def getent(): ''' Return info on all groups CLI Example:: salt '*' group.getent ''' ret = [] lines = __salt__['cmd.run']('net localgroup').split('\n') groupline = False for line in lines: if 'successfully' in line: groupline = False if groupline: ret.append(line.strip('*').strip()) if '---' in line: groupline = True return ret
Add Windows support to group add Add, remove and get info on Local Windows Groups
Add Windows support to group add Add, remove and get info on Local Windows Groups
Python
apache-2.0
saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt
--- +++ @@ -0,0 +1,90 @@ +''' +Manage groups on Windows +''' + +def __virtual__(): + ''' + Set the group module if the kernel is Windows + ''' + return 'group' if __grains__['kernel'] == 'Windows' else False + + +def add(name): + ''' + Add the specified group + + CLI Example:: + + salt '*' group.add foo + ''' + cmd = 'net localgroup {0} /add'.format(name) + + ret = __salt__['cmd.run_all'](cmd) + + return not ret['retcode'] + + +def delete(name): + ''' + Remove the named group + + CLI Example:: + + salt '*' group.delete foo + ''' + ret = __salt__['cmd.run_all']('net localgroup {0} /delete'.format(name)) + + return not ret['retcode'] + + +def info(name): + ''' + Return information about a group + + CLI Example:: + + salt '*' group.info foo + ''' + lines = __salt__['cmd.run']('net localgroup {0}'.format(name)).split('\n') + memberline = False + gr_mem = [] + gr_name = '' + for line in lines: + if 'Alias name' in line: + comps = line.split(' ', 1) + gr_name = comps[1].strip() + if 'successfully' in line: + memberline = False + if memberline: + gr_mem.append(line.strip()) + if '---' in line: + memberline = True + if not gr_name: + return False + + return {'name': gr_name, + 'passwd': None, + 'gid': None, + 'members': gr_mem} + + +def getent(): + ''' + Return info on all groups + + CLI Example:: + + salt '*' group.getent + ''' + ret = [] + lines = __salt__['cmd.run']('net localgroup').split('\n') + groupline = False + for line in lines: + if 'successfully' in line: + groupline = False + if groupline: + ret.append(line.strip('*').strip()) + if '---' in line: + groupline = True + + return ret
a04dddb09c276300c9a11dcbbe331ce2a32cff29
convert.py
convert.py
from AppKit import NSApplication, NSImage, NSImageCurrentFrame, NSGIFFileType; import sys, os dirname = sys.argv[1] files = os.listdir(dirname) for f in files: if '.gif' not in f: continue fName = os.path.join(dirname, f) tName=os.path.basename(fName) dir='/tmp/frames/' os.system('rm -rf %s && mkdir -p %s' % (dir,dir)) app=NSApplication.sharedApplication() img=NSImage.alloc().initWithContentsOfFile_(fName) if img: gifRep=img.representations()[0] frames=gifRep.valueForProperty_('NSImageFrameCount') if frames: for i in range(frames.intValue()): gifRep.setProperty_withValue_(NSImageCurrentFrame, i) gifRep.representationUsingType_properties_(NSGIFFileType, None).writeToFile_atomically_(dir + 'tmp' + str(i + 1).zfill(2) + '.gif', True) for i in range(9): os.system('mv %stmp0%d.gif %stmp%d.gif' % (dir, i + 1, dir, i+1)) v = "montage " for i in range(frames.intValue()): v += '/tmp/frames/tmp%d.gif ' % (i + 1) v += " -tile x1 -geometry +0+0 -alpha On -background \"rgba(0, 0, 0, 0.0)\" -quality 100 %s" % (fName.replace('gif', 'png'),) os.system(v)
Convert script for .gif -> spritesheet.
Convert script for .gif -> spritesheet.
Python
mit
Mytherin/PokemonAI,Mytherin/PokemonAI,Mytherin/PokemonAI
--- +++ @@ -0,0 +1,27 @@ +from AppKit import NSApplication, NSImage, NSImageCurrentFrame, NSGIFFileType; import sys, os + +dirname = sys.argv[1] +files = os.listdir(dirname) +for f in files: + if '.gif' not in f: continue + fName = os.path.join(dirname, f) + tName=os.path.basename(fName) + dir='/tmp/frames/' + os.system('rm -rf %s && mkdir -p %s' % (dir,dir)) + app=NSApplication.sharedApplication() + img=NSImage.alloc().initWithContentsOfFile_(fName) + if img: + gifRep=img.representations()[0] + frames=gifRep.valueForProperty_('NSImageFrameCount') + if frames: + for i in range(frames.intValue()): + gifRep.setProperty_withValue_(NSImageCurrentFrame, i) + gifRep.representationUsingType_properties_(NSGIFFileType, None).writeToFile_atomically_(dir + 'tmp' + str(i + 1).zfill(2) + '.gif', True) + for i in range(9): + os.system('mv %stmp0%d.gif %stmp%d.gif' % (dir, i + 1, dir, i+1)) + + v = "montage " + for i in range(frames.intValue()): + v += '/tmp/frames/tmp%d.gif ' % (i + 1) + v += " -tile x1 -geometry +0+0 -alpha On -background \"rgba(0, 0, 0, 0.0)\" -quality 100 %s" % (fName.replace('gif', 'png'),) + os.system(v)
ed7cc112adb6d54f1c2b79b941ab5455a17d9442
FlaskApp/app_ls.py
FlaskApp/app_ls.py
import subprocess import datetime from flask import Flask, render_template, redirect, url_for, request app = Flask(__name__) @app.route('/') def index(): return render_template('index.html') @app.route('/outlet') def rf_outlet(): return render_template('rf-outlet.html') @app.route('/hello/<name>') def hello(name): now = datetime.datetime.now() timeString = now.strftime("%Y-%m-%d %H:%M") templateData = { 'name': name, 'time': timeString } return render_template('hello.html', **templateData) # route for handling the login page logic @app.route('/login', methods=['GET', 'POST']) def login(): error = None if request.method == 'POST': if request.form['username'] != 'admin' or request.form['password'] != 'admin': error = 'Invalid Credentials. Please try again.' else: return redirect(url_for('/')) return render_template('login.html', error=error) @app.route('/postmethod', methods=['POST']) def get_post(): outlet, status = request.form['outlet'], request.form['status'] now = datetime.datetime.now() time = now.strftime("%Y-%m-%d %H:%M") print('Time: %s | Outlet: %s | Status: %s' % (time, outlet, status)) if status == 'on': out = subprocess.check_output('ls') print(out) else: out = subprocess.check_output('pwd') print(out) return outlet if __name__ == '__main__': app.run(host='0.0.0.0')
Add subprocess trial for webserver app
Add subprocess trial for webserver app
Python
bsd-3-clause
kbsezginel/raspberry-pi,kbsezginel/raspberry-pi,kbsezginel/raspberry-pi,kbsezginel/raspberry-pi
--- +++ @@ -0,0 +1,57 @@ +import subprocess +import datetime +from flask import Flask, render_template, redirect, url_for, request + +app = Flask(__name__) + + +@app.route('/') +def index(): + return render_template('index.html') + + +@app.route('/outlet') +def rf_outlet(): + return render_template('rf-outlet.html') + + +@app.route('/hello/<name>') +def hello(name): + now = datetime.datetime.now() + timeString = now.strftime("%Y-%m-%d %H:%M") + templateData = { + 'name': name, + 'time': timeString + } + return render_template('hello.html', **templateData) + + +# route for handling the login page logic +@app.route('/login', methods=['GET', 'POST']) +def login(): + error = None + if request.method == 'POST': + if request.form['username'] != 'admin' or request.form['password'] != 'admin': + error = 'Invalid Credentials. Please try again.' + else: + return redirect(url_for('/')) + return render_template('login.html', error=error) + + +@app.route('/postmethod', methods=['POST']) +def get_post(): + outlet, status = request.form['outlet'], request.form['status'] + now = datetime.datetime.now() + time = now.strftime("%Y-%m-%d %H:%M") + print('Time: %s | Outlet: %s | Status: %s' % (time, outlet, status)) + if status == 'on': + out = subprocess.check_output('ls') + print(out) + else: + out = subprocess.check_output('pwd') + print(out) + return outlet + + +if __name__ == '__main__': + app.run(host='0.0.0.0')
b6e8d93b661b4e65ecc6f5c150fcc1b1a4f26dc3
utils.py
utils.py
from numpy import array, cross, eye, dot from scipy.linalg import expm3,norm from math import sin, cos def rotate(axis, vector, angle): if axis == 'x': m = array([ [1, 0, 0], [0, cos(angle), -sin(angle)], [0, sin(angle), cos(angle)] ]) elif axis == 'y': m = array([ [cos(angle), 0, sin(angle)], [0, 1, 0], [-sin(angle), 0, cos(angle)] ]) elif axis == 'z': m = array([ [cos(angle), -sin(angle), 0], [sin(angle), cos(angle), 0], [0, 0, 1] ]) else: raise Exception("axis needs to be x, y or z") return dot(m, vector)
Add silly un-mathy rotation function
Add silly un-mathy rotation function
Python
mit
kirberich/3dscanner
--- +++ @@ -0,0 +1,27 @@ +from numpy import array, cross, eye, dot +from scipy.linalg import expm3,norm +from math import sin, cos + +def rotate(axis, vector, angle): + if axis == 'x': + m = array([ + [1, 0, 0], + [0, cos(angle), -sin(angle)], + [0, sin(angle), cos(angle)] + ]) + elif axis == 'y': + m = array([ + [cos(angle), 0, sin(angle)], + [0, 1, 0], + [-sin(angle), 0, cos(angle)] + ]) + elif axis == 'z': + m = array([ + [cos(angle), -sin(angle), 0], + [sin(angle), cos(angle), 0], + [0, 0, 1] + ]) + else: + raise Exception("axis needs to be x, y or z") + + return dot(m, vector)
a1f13dd46cdd68f72002ba13d4875af659694f94
models/base_olims_model.py
models/base_olims_model.py
import copy def add_a_field(cls, field): setattr(cls, field.string, field) pass def add_a_getter(cls, field): fieldname = field.string getter_template = copy.deepcopy(field.get_getterTemplate()) getter_template.__doc__ = "get%s - method to get value of field: %s" % (fieldname,fieldname) getter_template.__name__ = "get%s" % fieldname setattr(cls, getter_template.__name__, classmethod(getter_template)) pass class BaseOLiMSModel(object): ''' This is the base class of all OLiMS model. In addition to models.Model, every OLiMS model will be derived from this class and get all its code from a Bika Content class i.e. a Bika content class will be converted to OLiMS model class by removing irrelevant code, and adding derivation from this class, with the calling of following in the end DerivedModel_Of_BaseOLiMSModel.initialze(bika_schema) ''' @classmethod def initialze(cls, schema): # take out each filed from the passed schema and: # 1. make a model variable for it, and assign this field to this variable # 2. make getter methods for each model variable defined in step 1 for field in schema: add_a_field(cls, field) add_a_getter(cls, field) pass pass pass
import copy def add_a_field(cls, field): setattr(cls, field.string, field) pass def add_a_getter(cls, field): fieldname = field.string getter_template = copy.deepcopy(field.get_getterTemplate()) getter_template.__doc__ = "get%s - method to get value of field: %s" % (fieldname,fieldname) getter_template.__name__ = "get%s" % fieldname setattr(cls, getter_template.__name__, classmethod(getter_template)) pass class BaseOLiMSModel(object): ''' This is the base class of all OLiMS model. In addition to models.Model, every OLiMS model will be derived from this class and get all its code from a Bika Content class i.e. a Bika content class will be converted to OLiMS model class by removing irrelevant code, and adding derivation from this class, with the calling of following in the end DerivedModel_Of_BaseOLiMSModel.initialze(bika_schema) ''' @classmethod def initialze(cls, schema): # take out each filed from the passed schema and: # 1. make a model variable for it, and assign this field to this variable # 2. make getter methods for each model variable defined in step 1 for field in schema: add_a_field(cls, field) # add_a_getter(cls, field) pass pass pass
Make comment of 'add_a_getter' method
Make comment of 'add_a_getter' method
Python
agpl-3.0
sciCloud/OLiMS,yasir1brahim/OLiMS,sciCloud/OLiMS,sciCloud/OLiMS
--- +++ @@ -29,7 +29,7 @@ # 2. make getter methods for each model variable defined in step 1 for field in schema: add_a_field(cls, field) - add_a_getter(cls, field) +# add_a_getter(cls, field) pass pass pass
ba9214f7fca609948130d2ae56bb19805b79d59c
tests/rules/test_git_branch_list.py
tests/rules/test_git_branch_list.py
from thefuck import shells from thefuck.rules.git_branch_list import match, get_new_command from tests.utils import Command def test_match(): assert match(Command('git branch list'), None) def test_not_match(): assert not match(Command(), None) assert not match(Command('git commit'), None) assert not match(Command('git branch'), None) assert not match(Command('git stash list'), None) def test_get_new_command(): assert (get_new_command(Command('git branch list'), None) == shells.and_('git branch --delete list', 'git branch'))
Add a test for git_branch_list rule
Add a test for git_branch_list rule
Python
mit
beni55/thefuck,mlk/thefuck,zhangzhishan/thefuck,AntonChankin/thefuck,suxinde2009/thefuck,nvbn/thefuck,NguyenHoaiNam/thefuck,levythu/thefuck,PLNech/thefuck,redreamality/thefuck,thinkerchan/thefuck,princeofdarkness76/thefuck,BertieJim/thefuck,ostree/thefuck,bigplus/thefuck,vanita5/thefuck,beni55/thefuck,sekaiamber/thefuck,barneyElDinosaurio/thefuck,roth1002/thefuck,Clpsplug/thefuck,subajat1/thefuck,subajat1/thefuck,AntonChankin/thefuck,nvbn/thefuck,vanita5/thefuck,thesoulkiller/thefuck,PLNech/thefuck,LawrenceHan/thefuck,scorphus/thefuck,qingying5810/thefuck,thinkerchan/thefuck,SimenB/thefuck,barneyElDinosaurio/thefuck,BertieJim/thefuck,artiya4u/thefuck,levythu/thefuck,bigplus/thefuck,Clpsplug/thefuck,bugaevc/thefuck,hxddh/thefuck,gaurav9991/thefuck,ytjiang/thefuck,scorphus/thefuck,roth1002/thefuck,Aeron/thefuck,princeofdarkness76/thefuck,LawrenceHan/thefuck,ostree/thefuck,gogobebe2/thefuck,SimenB/thefuck,MJerty/thefuck,thesoulkiller/thefuck,petr-tichy/thefuck,mcarton/thefuck,MJerty/thefuck,manashmndl/thefuck,mlk/thefuck,lawrencebenson/thefuck,manashmndl/thefuck,mbbill/thefuck,lawrencebenson/thefuck,redreamality/thefuck,mcarton/thefuck,hxddh/thefuck
--- +++ @@ -0,0 +1,18 @@ +from thefuck import shells +from thefuck.rules.git_branch_list import match, get_new_command +from tests.utils import Command + +def test_match(): + assert match(Command('git branch list'), None) + + +def test_not_match(): + assert not match(Command(), None) + assert not match(Command('git commit'), None) + assert not match(Command('git branch'), None) + assert not match(Command('git stash list'), None) + + +def test_get_new_command(): + assert (get_new_command(Command('git branch list'), None) == + shells.and_('git branch --delete list', 'git branch'))
cecd767ef94f4bc890b8b19d2404528a7c4170bb
setup.py
setup.py
from setuptools import setup setup(name='LpSchedule', version='0.1', description='A API for Lviv Polytechnik schedule', author='Stepanov Valentyn', author_email='mr.valentyn.stepanov@gmail.com', url='http://example.com', install_requires=['Flask>=0.10.1', 'Flask-Script>=2.0.5','grab>=0.6.30','gunicorn>=19.5.0', 'itsdangerous>=0.24','Jinja2>=2.8','lxml>=3.6.0', 'MarkupSafe>=0.23', 'pycurl>=7.43.0','pytils>=0.3', 'selection>=0.0.11','six>=1.10.0','user-agent>=0.1.4','weblib>=0.1.20', 'Werkzeug>=0.11.9]'] )
Add pythonic way to install
Add pythonic way to install
Python
mit
stenvix/lpschedule,stepanov-valentin/lpschedule,stepanov-valentin/lpschedule,stenvix/lpschedule,stepanov-valentin/lpschedule,stepanov-valentin/lpschedule,stenvix/lpschedule,stenvix/lpschedule
--- +++ @@ -0,0 +1,15 @@ +from setuptools import setup + +setup(name='LpSchedule', + version='0.1', + description='A API for Lviv Polytechnik schedule', + author='Stepanov Valentyn', + author_email='mr.valentyn.stepanov@gmail.com', + url='http://example.com', + install_requires=['Flask>=0.10.1', + 'Flask-Script>=2.0.5','grab>=0.6.30','gunicorn>=19.5.0', + 'itsdangerous>=0.24','Jinja2>=2.8','lxml>=3.6.0', + 'MarkupSafe>=0.23', 'pycurl>=7.43.0','pytils>=0.3', + 'selection>=0.0.11','six>=1.10.0','user-agent>=0.1.4','weblib>=0.1.20', + 'Werkzeug>=0.11.9]'] + )
95edd6f6d3076e78b995b9b02e7d32938734cbf2
setup.py
setup.py
from distutils.core import setup setup( name='udiskie', version='0.3.8', description='Removable disk automounter for udisks', author='Byron Clark', author_email='byron@theclarkfamily.name', url='http://bitbucket.org/byronclark/udiskie', license='MIT', packages=[ 'udiskie', ], scripts=[ 'bin/udiskie', 'bin/udiskie-umount', ], )
from distutils.core import setup setup( name='udiskie', version='0.3.9', description='Removable disk automounter for udisks', author='Byron Clark', author_email='byron@theclarkfamily.name', url='http://bitbucket.org/byronclark/udiskie', license='MIT', packages=[ 'udiskie', ], scripts=[ 'bin/udiskie', 'bin/udiskie-umount', ], )
Prepare for next development cycle.
Prepare for next development cycle.
Python
mit
coldfix/udiskie,khardix/udiskie,coldfix/udiskie,pstray/udiskie,pstray/udiskie,mathstuf/udiskie
--- +++ @@ -2,7 +2,7 @@ setup( name='udiskie', - version='0.3.8', + version='0.3.9', description='Removable disk automounter for udisks', author='Byron Clark', author_email='byron@theclarkfamily.name',
acf0b48055c339e67ead4b85c90a07ffdce60bf1
py/max-consecutive-ones.py
py/max-consecutive-ones.py
class Solution(object): def findMaxConsecutiveOnes(self, nums): """ :type nums: List[int] :rtype: int """ n = 0 m = 0 for c in nums: if c == 0: n = 0 else: n += 1 m = max(n, m) return m
Add py solution for 485. Max Consecutive Ones
Add py solution for 485. Max Consecutive Ones 485. Max Consecutive Ones: https://leetcode.com/problems/max-consecutive-ones/
Python
apache-2.0
ckclark/leetcode,ckclark/leetcode,ckclark/leetcode,ckclark/leetcode,ckclark/leetcode,ckclark/leetcode
--- +++ @@ -0,0 +1,16 @@ +class Solution(object): + def findMaxConsecutiveOnes(self, nums): + """ + :type nums: List[int] + :rtype: int + """ + n = 0 + m = 0 + for c in nums: + if c == 0: + n = 0 + else: + n += 1 + m = max(n, m) + return m +
d86d117963ac87bfb11e731ea98d1803b2fcd609
test/test_utils.py
test/test_utils.py
# Copyright 2021 Sean Vig # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import pytest from pywayland.utils import AnonymousFile def test_anonymous_file(): with AnonymousFile(10) as fd: assert fd > 0 f = AnonymousFile(10) f.close() f.open() with pytest.raises(IOError, match="File is already open"): f.open() f.close() f.close()
Add test on anonymous file
Add test on anonymous file
Python
apache-2.0
flacjacket/pywayland
--- +++ @@ -0,0 +1,31 @@ +# Copyright 2021 Sean Vig +# +# Licensed under the Apache License, Version 2.0 (the "License"); +# you may not use this file except in compliance with the License. +# You may obtain a copy of the License at +# +# http://www.apache.org/licenses/LICENSE-2.0 +# +# Unless required by applicable law or agreed to in writing, software +# distributed under the License is distributed on an "AS IS" BASIS, +# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. +# See the License for the specific language governing permissions and +# limitations under the License. + +import pytest + +from pywayland.utils import AnonymousFile + + +def test_anonymous_file(): + with AnonymousFile(10) as fd: + assert fd > 0 + + f = AnonymousFile(10) + f.close() + + f.open() + with pytest.raises(IOError, match="File is already open"): + f.open() + f.close() + f.close()
43cb60efabfdaab77fa2c8eee1a5d8730a321db1
samples/query_interfaces.py
samples/query_interfaces.py
#!/usr/bin/env python import requests from orionsdk import SwisClient npm_server = 'localhost' username = 'admin' password = '' verify = False if not verify: from requests.packages.urllib3.exceptions import InsecureRequestWarning requests.packages.urllib3.disable_warnings(InsecureRequestWarning) def main(): swis = SwisClient(npm_server, username, password) query = """ SELECT TOP 10 n.NodeID, n.Caption AS NodeName, i.InterfaceID, i.Caption AS InterfaceName FROM Orion.Nodes n JOIN Orion.NPM.Interfaces i ON n.NodeID = i.NodeID """ results = swis.query(query) for row in results['results']: print("{NodeID} [{NodeName}] : {InterfaceID} [{InterfaceName}]".format(**row)) if __name__ == '__main__': main()
Add example for pulling interfaces.
Add example for pulling interfaces.
Python
apache-2.0
solarwinds/orionsdk-python
--- +++ @@ -0,0 +1,39 @@ +#!/usr/bin/env python +import requests +from orionsdk import SwisClient + + +npm_server = 'localhost' +username = 'admin' +password = '' + +verify = False +if not verify: + from requests.packages.urllib3.exceptions import InsecureRequestWarning + requests.packages.urllib3.disable_warnings(InsecureRequestWarning) + + +def main(): + swis = SwisClient(npm_server, username, password) + + query = """ + SELECT TOP 10 + n.NodeID, + n.Caption AS NodeName, + i.InterfaceID, + i.Caption AS InterfaceName + FROM + Orion.Nodes n + JOIN + Orion.NPM.Interfaces i ON n.NodeID = i.NodeID + """ + + results = swis.query(query) + + for row in results['results']: + print("{NodeID} [{NodeName}] : {InterfaceID} [{InterfaceName}]".format(**row)) + + +if __name__ == '__main__': + main() +
358cdd4b89221cbb02e7b04fc83cebb06570b03a
mezzanine/twitter/defaults.py
mezzanine/twitter/defaults.py
""" Default settings for the ``mezzanine.twitter`` app. Each of these can be overridden in your project's settings module, just like regular Django settings. The ``editable`` argument for each controls whether the setting is editable via Django's admin. Thought should be given to how a setting is actually used before making it editable, as it may be inappropriate - for example settings that are only read during startup shouldn't be editable, since changing them would require an application reload. """ from django.utils.translation import ugettext_lazy as _ from mezzanine.conf import register_setting from mezzanine.twitter import QUERY_TYPE_CHOICES, QUERY_TYPE_SEARCH register_setting( name="TWITTER_DEFAULT_QUERY_TYPE", label=_("Default Twitter Query Type"), description=_("Type of query that will be used to retrieve tweets for " "the default Twitter feed."), editable=True, default=QUERY_TYPE_SEARCH, choices=QUERY_TYPE_CHOICES, ) register_setting( name="TWITTER_DEFAULT_QUERY", label=_("Default Twitter Query"), description=_("Twitter query to use for the default query type."), editable=True, default="#django", ) register_setting( name="TWITTER_DEFAULT_NUM_TWEETS", label=_("Default Number of Tweets"), description=_("Number of tweets to display in the default Twitter feed."), editable=True, default=3, )
""" Default settings for the ``mezzanine.twitter`` app. Each of these can be overridden in your project's settings module, just like regular Django settings. The ``editable`` argument for each controls whether the setting is editable via Django's admin. Thought should be given to how a setting is actually used before making it editable, as it may be inappropriate - for example settings that are only read during startup shouldn't be editable, since changing them would require an application reload. """ from django.utils.translation import ugettext_lazy as _ from mezzanine.conf import register_setting from mezzanine.twitter import QUERY_TYPE_CHOICES, QUERY_TYPE_SEARCH register_setting( name="TWITTER_DEFAULT_QUERY_TYPE", label=_("Default Twitter Query Type"), description=_("Type of query that will be used to retrieve tweets for " "the default Twitter feed."), editable=True, default=QUERY_TYPE_SEARCH, choices=QUERY_TYPE_CHOICES, ) register_setting( name="TWITTER_DEFAULT_QUERY", label=_("Default Twitter Query"), description=_("Twitter query to use for the default query type."), editable=True, default="django mezzanine", ) register_setting( name="TWITTER_DEFAULT_NUM_TWEETS", label=_("Default Number of Tweets"), description=_("Number of tweets to display in the default Twitter feed."), editable=True, default=3, )
Update the default twitter query since it's been flooded by movie tweets.
Update the default twitter query since it's been flooded by movie tweets.
Python
bsd-2-clause
readevalprint/mezzanine,theclanks/mezzanine,dovydas/mezzanine,scarcry/snm-mezzanine,industrydive/mezzanine,ryneeverett/mezzanine,mush42/mezzanine,AlexHill/mezzanine,industrydive/mezzanine,spookylukey/mezzanine,eino-makitalo/mezzanine,Cajoline/mezzanine,Cajoline/mezzanine,webounty/mezzanine,gradel/mezzanine,dekomote/mezzanine-modeltranslation-backport,spookylukey/mezzanine,sjuxax/mezzanine,douglaskastle/mezzanine,vladir/mezzanine,Cicero-Zhao/mezzanine,geodesign/mezzanine,fusionbox/mezzanine,stbarnabas/mezzanine,orlenko/sfpirg,frankier/mezzanine,douglaskastle/mezzanine,damnfine/mezzanine,PegasusWang/mezzanine,stephenmcd/mezzanine,eino-makitalo/mezzanine,spookylukey/mezzanine,wbtuomela/mezzanine,agepoly/mezzanine,molokov/mezzanine,sjdines/mezzanine,christianwgd/mezzanine,dsanders11/mezzanine,readevalprint/mezzanine,batpad/mezzanine,frankchin/mezzanine,AlexHill/mezzanine,frankier/mezzanine,damnfine/mezzanine,Cicero-Zhao/mezzanine,wbtuomela/mezzanine,industrydive/mezzanine,SoLoHiC/mezzanine,frankchin/mezzanine,jjz/mezzanine,Skytorn86/mezzanine,mush42/mezzanine,ZeroXn/mezzanine,biomassives/mezzanine,gradel/mezzanine,emile2016/mezzanine,scarcry/snm-mezzanine,geodesign/mezzanine,viaregio/mezzanine,fusionbox/mezzanine,christianwgd/mezzanine,webounty/mezzanine,ZeroXn/mezzanine,Kniyl/mezzanine,cccs-web/mezzanine,molokov/mezzanine,tuxinhang1989/mezzanine,saintbird/mezzanine,ryneeverett/mezzanine,frankier/mezzanine,adrian-the-git/mezzanine,dsanders11/mezzanine,wyzex/mezzanine,sjdines/mezzanine,Skytorn86/mezzanine,Kniyl/mezzanine,orlenko/sfpirg,theclanks/mezzanine,orlenko/plei,dustinrb/mezzanine,agepoly/mezzanine,eino-makitalo/mezzanine,dovydas/mezzanine,joshcartme/mezzanine,sjdines/mezzanine,promil23/mezzanine,Kniyl/mezzanine,jjz/mezzanine,nikolas/mezzanine,sjuxax/mezzanine,joshcartme/mezzanine,molokov/mezzanine,dustinrb/mezzanine,vladir/mezzanine,tuxinhang1989/mezzanine,nikolas/mezzanine,wrwrwr/mezzanine,wyzex/mezzanine,saintbird/mezzanine,orlenko/plei,scarcry/snm-mezzanine,dovydas/mezzanine,nikolas/mezzanine,stephenmcd/mezzanine,ZeroXn/mezzanine,joshcartme/mezzanine,tuxinhang1989/mezzanine,dekomote/mezzanine-modeltranslation-backport,theclanks/mezzanine,ryneeverett/mezzanine,PegasusWang/mezzanine,damnfine/mezzanine,vladir/mezzanine,biomassives/mezzanine,adrian-the-git/mezzanine,orlenko/sfpirg,promil23/mezzanine,jjz/mezzanine,dustinrb/mezzanine,douglaskastle/mezzanine,SoLoHiC/mezzanine,stbarnabas/mezzanine,agepoly/mezzanine,dsanders11/mezzanine,batpad/mezzanine,wyzex/mezzanine,frankchin/mezzanine,adrian-the-git/mezzanine,PegasusWang/mezzanine,orlenko/plei,jerivas/mezzanine,wrwrwr/mezzanine,christianwgd/mezzanine,cccs-web/mezzanine,saintbird/mezzanine,jerivas/mezzanine,mush42/mezzanine,viaregio/mezzanine,webounty/mezzanine,dekomote/mezzanine-modeltranslation-backport,stephenmcd/mezzanine,jerivas/mezzanine,wbtuomela/mezzanine,biomassives/mezzanine,sjuxax/mezzanine,viaregio/mezzanine,promil23/mezzanine,gradel/mezzanine,SoLoHiC/mezzanine,emile2016/mezzanine,geodesign/mezzanine,Cajoline/mezzanine,Skytorn86/mezzanine,readevalprint/mezzanine,emile2016/mezzanine
--- +++ @@ -31,7 +31,7 @@ label=_("Default Twitter Query"), description=_("Twitter query to use for the default query type."), editable=True, - default="#django", + default="django mezzanine", ) register_setting(
d79a7f4e8be3cc9f0eb27da562bd0f1143005368
examples/power_on_swarm.py
examples/power_on_swarm.py
#!/usr/bin/python import time import sys from psphere.client import Client from psphere.managedobjects import VirtualMachine scatter_secs = 8 nodes = sys.argv[1:] client = Client() print("Powering on %s VMs" % len(nodes)) print("Estimated run time with %s seconds sleep between each power on: %s" % (scatter_secs, scatter_secs*len(nodes))) for node in nodes: try: vm = VirtualMachine.get(client, name=node, properties=["name", "runtime"]) except ObjectNotFoundError: print("WARNING: Could not find VM with name %s" % node) pass print("Powering on %s" % vm.name) if vm.runtime.powerState == "poweredOn": print("%s is already powered on." % vm.name) continue task = vm.PowerOnVM_Task() time.sleep(scatter_secs)
Add a new example of powering on many VMs at once
Add a new example of powering on many VMs at once
Python
apache-2.0
graphite-server/psphere,jkinred/psphere
--- +++ @@ -0,0 +1,32 @@ +#!/usr/bin/python + +import time +import sys + +from psphere.client import Client +from psphere.managedobjects import VirtualMachine + +scatter_secs = 8 + +nodes = sys.argv[1:] + +client = Client() + +print("Powering on %s VMs" % len(nodes)) +print("Estimated run time with %s seconds sleep between each power on: %s" % + (scatter_secs, scatter_secs*len(nodes))) + +for node in nodes: + try: + vm = VirtualMachine.get(client, name=node, properties=["name", "runtime"]) + except ObjectNotFoundError: + print("WARNING: Could not find VM with name %s" % node) + pass + + print("Powering on %s" % vm.name) + if vm.runtime.powerState == "poweredOn": + print("%s is already powered on." % vm.name) + continue + + task = vm.PowerOnVM_Task() + time.sleep(scatter_secs)
eb77b1335afb9d9e29d39ca9ff4e3f0fa3d89a8d
Lib/test/test_pwd.py
Lib/test/test_pwd.py
import pwd import string verbose = 0 if __name__ == '__main__': verbose = 1 entries = pwd.getpwall() for e in entries: name = e[0] uid = e[2] if verbose: print name, uid dbuid = pwd.getpwuid(uid) if dbuid[0] <> name: print 'Mismatch in pwd.getpwuid()' dbname = pwd.getpwnam(name) if dbname[2] <> uid: print 'Mismatch in pwd.getpwnam()' break # try to get some errors bynames = {} byuids = {} for n, p, u, g, gecos, d, s in entries: bynames[n] = u byuids[u] = n allnames = bynames.keys() namei = 0 fakename = allnames[namei] while bynames.has_key(fakename): chars = map(None, fakename) for i in range(len(chars)): if chars[i] == 'z': chars[i] = 'A' break elif chars[i] == 'Z': continue else: chars[i] = chr(ord(chars[i]) + 1) break else: namei = namei + 1 try: fakename = allnames[namei] except IndexError: # should never happen... if so, just forget it break fakename = string.join(map(None, chars), '') try: pwd.getpwnam(fakename) except KeyError: pass else: print 'fakename', fakename, 'did not except pwd.getpwnam()' uids = byuids.keys() uids.sort() uids.reverse() fakeuid = uids[0] + 1 try: pwd.getpwuid(fakeuid) except KeyError: pass else: print 'fakeuid', fakeuid, 'did not except pwd.getpwuid()'
Test of the pwd module
Test of the pwd module
Python
mit
sk-/python2.7-type-annotator,sk-/python2.7-type-annotator,sk-/python2.7-type-annotator
--- +++ @@ -0,0 +1,70 @@ +import pwd +import string + +verbose = 0 +if __name__ == '__main__': + verbose = 1 + +entries = pwd.getpwall() + +for e in entries: + name = e[0] + uid = e[2] + if verbose: + print name, uid + dbuid = pwd.getpwuid(uid) + if dbuid[0] <> name: + print 'Mismatch in pwd.getpwuid()' + dbname = pwd.getpwnam(name) + if dbname[2] <> uid: + print 'Mismatch in pwd.getpwnam()' + break + +# try to get some errors +bynames = {} +byuids = {} +for n, p, u, g, gecos, d, s in entries: + bynames[n] = u + byuids[u] = n + +allnames = bynames.keys() +namei = 0 +fakename = allnames[namei] +while bynames.has_key(fakename): + chars = map(None, fakename) + for i in range(len(chars)): + if chars[i] == 'z': + chars[i] = 'A' + break + elif chars[i] == 'Z': + continue + else: + chars[i] = chr(ord(chars[i]) + 1) + break + else: + namei = namei + 1 + try: + fakename = allnames[namei] + except IndexError: + # should never happen... if so, just forget it + break + fakename = string.join(map(None, chars), '') + +try: + pwd.getpwnam(fakename) +except KeyError: + pass +else: + print 'fakename', fakename, 'did not except pwd.getpwnam()' + +uids = byuids.keys() +uids.sort() +uids.reverse() +fakeuid = uids[0] + 1 + +try: + pwd.getpwuid(fakeuid) +except KeyError: + pass +else: + print 'fakeuid', fakeuid, 'did not except pwd.getpwuid()'
ea5bd5fe05c6d2a9580f5ca6c2238971c28e36bc
apps/plea/migrations/0026_caseoffencefilter.py
apps/plea/migrations/0026_caseoffencefilter.py
# -*- coding: utf-8 -*- from __future__ import unicode_literals from django.db import models, migrations class Migration(migrations.Migration): dependencies = [ ('plea', '0025_auto_20151210_1526'), ] operations = [ migrations.CreateModel( name='CaseOffenceFilter', fields=[ ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)), ('filter_match', models.CharField(max_length=20)), ('description', models.CharField(max_length=500, null=True, blank=True)), ], ), ]
Add migration for CaseOffenceFilter model
Add migration for CaseOffenceFilter model
Python
mit
ministryofjustice/manchester_traffic_offences_pleas,ministryofjustice/manchester_traffic_offences_pleas,ministryofjustice/manchester_traffic_offences_pleas,ministryofjustice/manchester_traffic_offences_pleas
--- +++ @@ -0,0 +1,22 @@ +# -*- coding: utf-8 -*- +from __future__ import unicode_literals + +from django.db import models, migrations + + +class Migration(migrations.Migration): + + dependencies = [ + ('plea', '0025_auto_20151210_1526'), + ] + + operations = [ + migrations.CreateModel( + name='CaseOffenceFilter', + fields=[ + ('id', models.AutoField(verbose_name='ID', serialize=False, auto_created=True, primary_key=True)), + ('filter_match', models.CharField(max_length=20)), + ('description', models.CharField(max_length=500, null=True, blank=True)), + ], + ), + ]