commit
stringlengths
40
40
old_file
stringlengths
4
118
new_file
stringlengths
4
118
old_contents
stringlengths
0
2.94k
new_contents
stringlengths
1
4.43k
subject
stringlengths
15
444
message
stringlengths
16
3.45k
lang
stringclasses
1 value
license
stringclasses
13 values
repos
stringlengths
5
43.2k
prompt
stringlengths
17
4.58k
response
stringlengths
1
4.43k
prompt_tagged
stringlengths
58
4.62k
response_tagged
stringlengths
1
4.43k
text
stringlengths
132
7.29k
text_tagged
stringlengths
173
7.33k
1e148822b4611403add32154fbe47bc8775f8001
py/garage/garage/sql/sqlite.py
py/garage/garage/sql/sqlite.py
__all__ = [ 'create_engine', ] import sqlalchemy def create_engine(db_uri, check_same_thread=False, echo=False): engine = sqlalchemy.create_engine( db_uri, echo=echo, connect_args={ 'check_same_thread': check_same_thread, }, ) @sqlalchemy.event.listens_for(engine, 'connect') def do_connect(dbapi_connection, _): # Stop pysqlite issue commit automatically. dbapi_connection.isolation_level = None @sqlalchemy.event.listens_for(engine, 'begin') def do_begin(connection): connection.execute('BEGIN EXCLUSIVE') return engine
__all__ = [ 'create_engine', ] import sqlalchemy def create_engine(db_uri, check_same_thread=False, echo=False): engine = sqlalchemy.create_engine( db_uri, echo=echo, connect_args={ 'check_same_thread': check_same_thread, }, ) @sqlalchemy.event.listens_for(engine, 'connect') def do_connect(dbapi_connection, _): # Stop pysqlite issue commit automatically. dbapi_connection.isolation_level = None # Enable foreign key. cursor = dbapi_connection.cursor() cursor.execute("PRAGMA foreign_keys = ON") cursor.close() @sqlalchemy.event.listens_for(engine, 'begin') def do_begin(connection): connection.execute('BEGIN EXCLUSIVE') return engine
Enable foreign key constraint for SQLite
Enable foreign key constraint for SQLite
Python
mit
clchiou/garage,clchiou/garage,clchiou/garage,clchiou/garage
__all__ = [ 'create_engine', ] import sqlalchemy def create_engine(db_uri, check_same_thread=False, echo=False): engine = sqlalchemy.create_engine( db_uri, echo=echo, connect_args={ 'check_same_thread': check_same_thread, }, ) @sqlalchemy.event.listens_for(engine, 'connect') def do_connect(dbapi_connection, _): # Stop pysqlite issue commit automatically. dbapi_connection.isolation_level = None @sqlalchemy.event.listens_for(engine, 'begin') def do_begin(connection): connection.execute('BEGIN EXCLUSIVE') return engine Enable foreign key constraint for SQLite
__all__ = [ 'create_engine', ] import sqlalchemy def create_engine(db_uri, check_same_thread=False, echo=False): engine = sqlalchemy.create_engine( db_uri, echo=echo, connect_args={ 'check_same_thread': check_same_thread, }, ) @sqlalchemy.event.listens_for(engine, 'connect') def do_connect(dbapi_connection, _): # Stop pysqlite issue commit automatically. dbapi_connection.isolation_level = None # Enable foreign key. cursor = dbapi_connection.cursor() cursor.execute("PRAGMA foreign_keys = ON") cursor.close() @sqlalchemy.event.listens_for(engine, 'begin') def do_begin(connection): connection.execute('BEGIN EXCLUSIVE') return engine
<commit_before>__all__ = [ 'create_engine', ] import sqlalchemy def create_engine(db_uri, check_same_thread=False, echo=False): engine = sqlalchemy.create_engine( db_uri, echo=echo, connect_args={ 'check_same_thread': check_same_thread, }, ) @sqlalchemy.event.listens_for(engine, 'connect') def do_connect(dbapi_connection, _): # Stop pysqlite issue commit automatically. dbapi_connection.isolation_level = None @sqlalchemy.event.listens_for(engine, 'begin') def do_begin(connection): connection.execute('BEGIN EXCLUSIVE') return engine <commit_msg>Enable foreign key constraint for SQLite<commit_after>
__all__ = [ 'create_engine', ] import sqlalchemy def create_engine(db_uri, check_same_thread=False, echo=False): engine = sqlalchemy.create_engine( db_uri, echo=echo, connect_args={ 'check_same_thread': check_same_thread, }, ) @sqlalchemy.event.listens_for(engine, 'connect') def do_connect(dbapi_connection, _): # Stop pysqlite issue commit automatically. dbapi_connection.isolation_level = None # Enable foreign key. cursor = dbapi_connection.cursor() cursor.execute("PRAGMA foreign_keys = ON") cursor.close() @sqlalchemy.event.listens_for(engine, 'begin') def do_begin(connection): connection.execute('BEGIN EXCLUSIVE') return engine
__all__ = [ 'create_engine', ] import sqlalchemy def create_engine(db_uri, check_same_thread=False, echo=False): engine = sqlalchemy.create_engine( db_uri, echo=echo, connect_args={ 'check_same_thread': check_same_thread, }, ) @sqlalchemy.event.listens_for(engine, 'connect') def do_connect(dbapi_connection, _): # Stop pysqlite issue commit automatically. dbapi_connection.isolation_level = None @sqlalchemy.event.listens_for(engine, 'begin') def do_begin(connection): connection.execute('BEGIN EXCLUSIVE') return engine Enable foreign key constraint for SQLite__all__ = [ 'create_engine', ] import sqlalchemy def create_engine(db_uri, check_same_thread=False, echo=False): engine = sqlalchemy.create_engine( db_uri, echo=echo, connect_args={ 'check_same_thread': check_same_thread, }, ) @sqlalchemy.event.listens_for(engine, 'connect') def do_connect(dbapi_connection, _): # Stop pysqlite issue commit automatically. dbapi_connection.isolation_level = None # Enable foreign key. cursor = dbapi_connection.cursor() cursor.execute("PRAGMA foreign_keys = ON") cursor.close() @sqlalchemy.event.listens_for(engine, 'begin') def do_begin(connection): connection.execute('BEGIN EXCLUSIVE') return engine
<commit_before>__all__ = [ 'create_engine', ] import sqlalchemy def create_engine(db_uri, check_same_thread=False, echo=False): engine = sqlalchemy.create_engine( db_uri, echo=echo, connect_args={ 'check_same_thread': check_same_thread, }, ) @sqlalchemy.event.listens_for(engine, 'connect') def do_connect(dbapi_connection, _): # Stop pysqlite issue commit automatically. dbapi_connection.isolation_level = None @sqlalchemy.event.listens_for(engine, 'begin') def do_begin(connection): connection.execute('BEGIN EXCLUSIVE') return engine <commit_msg>Enable foreign key constraint for SQLite<commit_after>__all__ = [ 'create_engine', ] import sqlalchemy def create_engine(db_uri, check_same_thread=False, echo=False): engine = sqlalchemy.create_engine( db_uri, echo=echo, connect_args={ 'check_same_thread': check_same_thread, }, ) @sqlalchemy.event.listens_for(engine, 'connect') def do_connect(dbapi_connection, _): # Stop pysqlite issue commit automatically. dbapi_connection.isolation_level = None # Enable foreign key. cursor = dbapi_connection.cursor() cursor.execute("PRAGMA foreign_keys = ON") cursor.close() @sqlalchemy.event.listens_for(engine, 'begin') def do_begin(connection): connection.execute('BEGIN EXCLUSIVE') return engine
7e2835f76474f6153d8972a983c5d45f9c4f11ee
tests/Physics/TestLight.py
tests/Physics/TestLight.py
#!/usr/bin/env python3 # -*- coding: utf-8 -*- from numpy.testing import assert_approx_equal from UliEngineering.Physics.Light import * from UliEngineering.EngineerIO import auto_format import unittest class TestJohnsonNyquistNoise(unittest.TestCase): def test_lumen_to_candela_by_apex_angle(self): v = lumen_to_candela_by_apex_angle("25 lm", "120°") assert_approx_equal(v, 7.9577471546, significant=5) self.assertEqual(auto_format(lumen_to_candela_by_apex_angle, "25 lm", "120°"), "7.96 cd")
#!/usr/bin/env python3 # -*- coding: utf-8 -*- from numpy.testing import assert_approx_equal from UliEngineering.Physics.Light import * from UliEngineering.EngineerIO import auto_format import unittest class TestLight(unittest.TestCase): def test_lumen_to_candela_by_apex_angle(self): v = lumen_to_candela_by_apex_angle("25 lm", "120°") assert_approx_equal(v, 7.9577471546, significant=5) self.assertEqual(auto_format(lumen_to_candela_by_apex_angle, "25 lm", "120°"), "7.96 cd")
Fix badly named test class
Fix badly named test class
Python
apache-2.0
ulikoehler/UliEngineering
#!/usr/bin/env python3 # -*- coding: utf-8 -*- from numpy.testing import assert_approx_equal from UliEngineering.Physics.Light import * from UliEngineering.EngineerIO import auto_format import unittest class TestJohnsonNyquistNoise(unittest.TestCase): def test_lumen_to_candela_by_apex_angle(self): v = lumen_to_candela_by_apex_angle("25 lm", "120°") assert_approx_equal(v, 7.9577471546, significant=5) self.assertEqual(auto_format(lumen_to_candela_by_apex_angle, "25 lm", "120°"), "7.96 cd") Fix badly named test class
#!/usr/bin/env python3 # -*- coding: utf-8 -*- from numpy.testing import assert_approx_equal from UliEngineering.Physics.Light import * from UliEngineering.EngineerIO import auto_format import unittest class TestLight(unittest.TestCase): def test_lumen_to_candela_by_apex_angle(self): v = lumen_to_candela_by_apex_angle("25 lm", "120°") assert_approx_equal(v, 7.9577471546, significant=5) self.assertEqual(auto_format(lumen_to_candela_by_apex_angle, "25 lm", "120°"), "7.96 cd")
<commit_before>#!/usr/bin/env python3 # -*- coding: utf-8 -*- from numpy.testing import assert_approx_equal from UliEngineering.Physics.Light import * from UliEngineering.EngineerIO import auto_format import unittest class TestJohnsonNyquistNoise(unittest.TestCase): def test_lumen_to_candela_by_apex_angle(self): v = lumen_to_candela_by_apex_angle("25 lm", "120°") assert_approx_equal(v, 7.9577471546, significant=5) self.assertEqual(auto_format(lumen_to_candela_by_apex_angle, "25 lm", "120°"), "7.96 cd") <commit_msg>Fix badly named test class<commit_after>
#!/usr/bin/env python3 # -*- coding: utf-8 -*- from numpy.testing import assert_approx_equal from UliEngineering.Physics.Light import * from UliEngineering.EngineerIO import auto_format import unittest class TestLight(unittest.TestCase): def test_lumen_to_candela_by_apex_angle(self): v = lumen_to_candela_by_apex_angle("25 lm", "120°") assert_approx_equal(v, 7.9577471546, significant=5) self.assertEqual(auto_format(lumen_to_candela_by_apex_angle, "25 lm", "120°"), "7.96 cd")
#!/usr/bin/env python3 # -*- coding: utf-8 -*- from numpy.testing import assert_approx_equal from UliEngineering.Physics.Light import * from UliEngineering.EngineerIO import auto_format import unittest class TestJohnsonNyquistNoise(unittest.TestCase): def test_lumen_to_candela_by_apex_angle(self): v = lumen_to_candela_by_apex_angle("25 lm", "120°") assert_approx_equal(v, 7.9577471546, significant=5) self.assertEqual(auto_format(lumen_to_candela_by_apex_angle, "25 lm", "120°"), "7.96 cd") Fix badly named test class#!/usr/bin/env python3 # -*- coding: utf-8 -*- from numpy.testing import assert_approx_equal from UliEngineering.Physics.Light import * from UliEngineering.EngineerIO import auto_format import unittest class TestLight(unittest.TestCase): def test_lumen_to_candela_by_apex_angle(self): v = lumen_to_candela_by_apex_angle("25 lm", "120°") assert_approx_equal(v, 7.9577471546, significant=5) self.assertEqual(auto_format(lumen_to_candela_by_apex_angle, "25 lm", "120°"), "7.96 cd")
<commit_before>#!/usr/bin/env python3 # -*- coding: utf-8 -*- from numpy.testing import assert_approx_equal from UliEngineering.Physics.Light import * from UliEngineering.EngineerIO import auto_format import unittest class TestJohnsonNyquistNoise(unittest.TestCase): def test_lumen_to_candela_by_apex_angle(self): v = lumen_to_candela_by_apex_angle("25 lm", "120°") assert_approx_equal(v, 7.9577471546, significant=5) self.assertEqual(auto_format(lumen_to_candela_by_apex_angle, "25 lm", "120°"), "7.96 cd") <commit_msg>Fix badly named test class<commit_after>#!/usr/bin/env python3 # -*- coding: utf-8 -*- from numpy.testing import assert_approx_equal from UliEngineering.Physics.Light import * from UliEngineering.EngineerIO import auto_format import unittest class TestLight(unittest.TestCase): def test_lumen_to_candela_by_apex_angle(self): v = lumen_to_candela_by_apex_angle("25 lm", "120°") assert_approx_equal(v, 7.9577471546, significant=5) self.assertEqual(auto_format(lumen_to_candela_by_apex_angle, "25 lm", "120°"), "7.96 cd")
e2c9d39dd30a60c5c54521d7d11773430cae1bd1
tests/test_image_access.py
tests/test_image_access.py
import pytest import imghdr from io import BytesIO from PIL import Image from pikepdf import _qpdf as qpdf def test_jpeg(resources, outdir): pdf = qpdf.Pdf.open(resources / 'congress.pdf') # If you are looking at this as example code, Im0 is not necessarily the # name of any image. pdfimage = pdf.pages[0].Resources.XObject.Im0 raw_stream = pdf.pages[0].Resources.XObject.Im0.read_raw_stream() with pytest.raises(RuntimeError): pdf.pages[0].Resources.XObject.Im0.read_stream() assert imghdr.what('', h=raw_stream) == 'jpeg' im = Image.open(BytesIO(raw_stream)) assert im.size == (pdfimage.Width, pdfimage.Height) assert im.mode == 'RGB'
import pytest import imghdr from io import BytesIO from PIL import Image import zlib from pikepdf import Pdf, Object def test_jpeg(resources, outdir): pdf = Pdf.open(resources / 'congress.pdf') # If you are looking at this as example code, Im0 is not necessarily the # name of any image. pdfimage = pdf.pages[0].Resources.XObject['/Im0'] raw_bytes = pdfimage.read_raw_bytes() with pytest.raises(RuntimeError): pdfimage.read_bytes() assert imghdr.what('', h=raw_bytes) == 'jpeg' im = Image.open(BytesIO(raw_bytes)) assert im.size == (pdfimage.Width, pdfimage.Height) assert im.mode == 'RGB' def test_replace_jpeg(resources, outdir): pdf = Pdf.open(resources / 'congress.pdf') pdfimage = pdf.pages[0].Resources.XObject['/Im0'] raw_bytes = pdfimage.read_raw_bytes() im = Image.open(BytesIO(raw_bytes)) grayscale = im.convert('L') #newimage = Object.Stream(pdf, grayscale.tobytes()) pdfimage.write(zlib.compress(grayscale.tobytes()), Object.Name("/FlateDecode"), Object.Null()) pdfimage.ColorSpace = Object.Name('/DeviceGray') pdf.save(outdir / 'congress_gray.pdf')
Add manual experiment that replaces a RGB image with grayscale
Add manual experiment that replaces a RGB image with grayscale
Python
mpl-2.0
pikepdf/pikepdf,pikepdf/pikepdf,pikepdf/pikepdf
import pytest import imghdr from io import BytesIO from PIL import Image from pikepdf import _qpdf as qpdf def test_jpeg(resources, outdir): pdf = qpdf.Pdf.open(resources / 'congress.pdf') # If you are looking at this as example code, Im0 is not necessarily the # name of any image. pdfimage = pdf.pages[0].Resources.XObject.Im0 raw_stream = pdf.pages[0].Resources.XObject.Im0.read_raw_stream() with pytest.raises(RuntimeError): pdf.pages[0].Resources.XObject.Im0.read_stream() assert imghdr.what('', h=raw_stream) == 'jpeg' im = Image.open(BytesIO(raw_stream)) assert im.size == (pdfimage.Width, pdfimage.Height) assert im.mode == 'RGB'Add manual experiment that replaces a RGB image with grayscale
import pytest import imghdr from io import BytesIO from PIL import Image import zlib from pikepdf import Pdf, Object def test_jpeg(resources, outdir): pdf = Pdf.open(resources / 'congress.pdf') # If you are looking at this as example code, Im0 is not necessarily the # name of any image. pdfimage = pdf.pages[0].Resources.XObject['/Im0'] raw_bytes = pdfimage.read_raw_bytes() with pytest.raises(RuntimeError): pdfimage.read_bytes() assert imghdr.what('', h=raw_bytes) == 'jpeg' im = Image.open(BytesIO(raw_bytes)) assert im.size == (pdfimage.Width, pdfimage.Height) assert im.mode == 'RGB' def test_replace_jpeg(resources, outdir): pdf = Pdf.open(resources / 'congress.pdf') pdfimage = pdf.pages[0].Resources.XObject['/Im0'] raw_bytes = pdfimage.read_raw_bytes() im = Image.open(BytesIO(raw_bytes)) grayscale = im.convert('L') #newimage = Object.Stream(pdf, grayscale.tobytes()) pdfimage.write(zlib.compress(grayscale.tobytes()), Object.Name("/FlateDecode"), Object.Null()) pdfimage.ColorSpace = Object.Name('/DeviceGray') pdf.save(outdir / 'congress_gray.pdf')
<commit_before>import pytest import imghdr from io import BytesIO from PIL import Image from pikepdf import _qpdf as qpdf def test_jpeg(resources, outdir): pdf = qpdf.Pdf.open(resources / 'congress.pdf') # If you are looking at this as example code, Im0 is not necessarily the # name of any image. pdfimage = pdf.pages[0].Resources.XObject.Im0 raw_stream = pdf.pages[0].Resources.XObject.Im0.read_raw_stream() with pytest.raises(RuntimeError): pdf.pages[0].Resources.XObject.Im0.read_stream() assert imghdr.what('', h=raw_stream) == 'jpeg' im = Image.open(BytesIO(raw_stream)) assert im.size == (pdfimage.Width, pdfimage.Height) assert im.mode == 'RGB'<commit_msg>Add manual experiment that replaces a RGB image with grayscale<commit_after>
import pytest import imghdr from io import BytesIO from PIL import Image import zlib from pikepdf import Pdf, Object def test_jpeg(resources, outdir): pdf = Pdf.open(resources / 'congress.pdf') # If you are looking at this as example code, Im0 is not necessarily the # name of any image. pdfimage = pdf.pages[0].Resources.XObject['/Im0'] raw_bytes = pdfimage.read_raw_bytes() with pytest.raises(RuntimeError): pdfimage.read_bytes() assert imghdr.what('', h=raw_bytes) == 'jpeg' im = Image.open(BytesIO(raw_bytes)) assert im.size == (pdfimage.Width, pdfimage.Height) assert im.mode == 'RGB' def test_replace_jpeg(resources, outdir): pdf = Pdf.open(resources / 'congress.pdf') pdfimage = pdf.pages[0].Resources.XObject['/Im0'] raw_bytes = pdfimage.read_raw_bytes() im = Image.open(BytesIO(raw_bytes)) grayscale = im.convert('L') #newimage = Object.Stream(pdf, grayscale.tobytes()) pdfimage.write(zlib.compress(grayscale.tobytes()), Object.Name("/FlateDecode"), Object.Null()) pdfimage.ColorSpace = Object.Name('/DeviceGray') pdf.save(outdir / 'congress_gray.pdf')
import pytest import imghdr from io import BytesIO from PIL import Image from pikepdf import _qpdf as qpdf def test_jpeg(resources, outdir): pdf = qpdf.Pdf.open(resources / 'congress.pdf') # If you are looking at this as example code, Im0 is not necessarily the # name of any image. pdfimage = pdf.pages[0].Resources.XObject.Im0 raw_stream = pdf.pages[0].Resources.XObject.Im0.read_raw_stream() with pytest.raises(RuntimeError): pdf.pages[0].Resources.XObject.Im0.read_stream() assert imghdr.what('', h=raw_stream) == 'jpeg' im = Image.open(BytesIO(raw_stream)) assert im.size == (pdfimage.Width, pdfimage.Height) assert im.mode == 'RGB'Add manual experiment that replaces a RGB image with grayscaleimport pytest import imghdr from io import BytesIO from PIL import Image import zlib from pikepdf import Pdf, Object def test_jpeg(resources, outdir): pdf = Pdf.open(resources / 'congress.pdf') # If you are looking at this as example code, Im0 is not necessarily the # name of any image. pdfimage = pdf.pages[0].Resources.XObject['/Im0'] raw_bytes = pdfimage.read_raw_bytes() with pytest.raises(RuntimeError): pdfimage.read_bytes() assert imghdr.what('', h=raw_bytes) == 'jpeg' im = Image.open(BytesIO(raw_bytes)) assert im.size == (pdfimage.Width, pdfimage.Height) assert im.mode == 'RGB' def test_replace_jpeg(resources, outdir): pdf = Pdf.open(resources / 'congress.pdf') pdfimage = pdf.pages[0].Resources.XObject['/Im0'] raw_bytes = pdfimage.read_raw_bytes() im = Image.open(BytesIO(raw_bytes)) grayscale = im.convert('L') #newimage = Object.Stream(pdf, grayscale.tobytes()) pdfimage.write(zlib.compress(grayscale.tobytes()), Object.Name("/FlateDecode"), Object.Null()) pdfimage.ColorSpace = Object.Name('/DeviceGray') pdf.save(outdir / 'congress_gray.pdf')
<commit_before>import pytest import imghdr from io import BytesIO from PIL import Image from pikepdf import _qpdf as qpdf def test_jpeg(resources, outdir): pdf = qpdf.Pdf.open(resources / 'congress.pdf') # If you are looking at this as example code, Im0 is not necessarily the # name of any image. pdfimage = pdf.pages[0].Resources.XObject.Im0 raw_stream = pdf.pages[0].Resources.XObject.Im0.read_raw_stream() with pytest.raises(RuntimeError): pdf.pages[0].Resources.XObject.Im0.read_stream() assert imghdr.what('', h=raw_stream) == 'jpeg' im = Image.open(BytesIO(raw_stream)) assert im.size == (pdfimage.Width, pdfimage.Height) assert im.mode == 'RGB'<commit_msg>Add manual experiment that replaces a RGB image with grayscale<commit_after>import pytest import imghdr from io import BytesIO from PIL import Image import zlib from pikepdf import Pdf, Object def test_jpeg(resources, outdir): pdf = Pdf.open(resources / 'congress.pdf') # If you are looking at this as example code, Im0 is not necessarily the # name of any image. pdfimage = pdf.pages[0].Resources.XObject['/Im0'] raw_bytes = pdfimage.read_raw_bytes() with pytest.raises(RuntimeError): pdfimage.read_bytes() assert imghdr.what('', h=raw_bytes) == 'jpeg' im = Image.open(BytesIO(raw_bytes)) assert im.size == (pdfimage.Width, pdfimage.Height) assert im.mode == 'RGB' def test_replace_jpeg(resources, outdir): pdf = Pdf.open(resources / 'congress.pdf') pdfimage = pdf.pages[0].Resources.XObject['/Im0'] raw_bytes = pdfimage.read_raw_bytes() im = Image.open(BytesIO(raw_bytes)) grayscale = im.convert('L') #newimage = Object.Stream(pdf, grayscale.tobytes()) pdfimage.write(zlib.compress(grayscale.tobytes()), Object.Name("/FlateDecode"), Object.Null()) pdfimage.ColorSpace = Object.Name('/DeviceGray') pdf.save(outdir / 'congress_gray.pdf')
6874ff82ddf5e9d803a45676c45d83be65ad3b33
core/app.py
core/app.py
import os from flask import Flask from flask_login import LoginManager from flask_sqlalchemy import SQLAlchemy template_directory = os.path.join( os.path.dirname(os.path.abspath(__file__)), 'frontend/templates' ) static_directory = 'frontend/static' app = Flask(__name__, template_folder=template_directory, static_folder=static_directory) app.config.from_object('config.flask_config') def init_db(): """ Initialize the SQLAlchemy database object. :return: A SQLAlchemy instance used universally for database operations. """ return SQLAlchemy(app, session_options={ 'expire_on_commit': False, }) def init_login_manager(): """ Initialize the login manager. :return: A LoginManager instance used universally for flask-login loaders. """ login_manager = LoginManager() login_manager.init_app(app) return login_manager
import os from flask import Flask from flask_login import LoginManager from flask_sqlalchemy import SQLAlchemy template_directory = os.path.join( os.path.dirname(os.path.abspath(__file__)), '../frontend/templates' ) static_directory = '../frontend/static' app = Flask(__name__, template_folder=template_directory, static_folder=static_directory) app.config.from_object('config.flask_config') def init_db(): """ Initialize the SQLAlchemy database object. :return: A SQLAlchemy instance used universally for database operations. """ return SQLAlchemy(app, session_options={ 'expire_on_commit': False, }) def init_login_manager(): """ Initialize the login manager. :return: A LoginManager instance used universally for flask-login loaders. """ login_manager = LoginManager() login_manager.init_app(app) return login_manager
Fix template and static paths
Fix template and static paths
Python
mit
LINKIWI/linkr,LINKIWI/linkr,LINKIWI/linkr
import os from flask import Flask from flask_login import LoginManager from flask_sqlalchemy import SQLAlchemy template_directory = os.path.join( os.path.dirname(os.path.abspath(__file__)), 'frontend/templates' ) static_directory = 'frontend/static' app = Flask(__name__, template_folder=template_directory, static_folder=static_directory) app.config.from_object('config.flask_config') def init_db(): """ Initialize the SQLAlchemy database object. :return: A SQLAlchemy instance used universally for database operations. """ return SQLAlchemy(app, session_options={ 'expire_on_commit': False, }) def init_login_manager(): """ Initialize the login manager. :return: A LoginManager instance used universally for flask-login loaders. """ login_manager = LoginManager() login_manager.init_app(app) return login_manager Fix template and static paths
import os from flask import Flask from flask_login import LoginManager from flask_sqlalchemy import SQLAlchemy template_directory = os.path.join( os.path.dirname(os.path.abspath(__file__)), '../frontend/templates' ) static_directory = '../frontend/static' app = Flask(__name__, template_folder=template_directory, static_folder=static_directory) app.config.from_object('config.flask_config') def init_db(): """ Initialize the SQLAlchemy database object. :return: A SQLAlchemy instance used universally for database operations. """ return SQLAlchemy(app, session_options={ 'expire_on_commit': False, }) def init_login_manager(): """ Initialize the login manager. :return: A LoginManager instance used universally for flask-login loaders. """ login_manager = LoginManager() login_manager.init_app(app) return login_manager
<commit_before>import os from flask import Flask from flask_login import LoginManager from flask_sqlalchemy import SQLAlchemy template_directory = os.path.join( os.path.dirname(os.path.abspath(__file__)), 'frontend/templates' ) static_directory = 'frontend/static' app = Flask(__name__, template_folder=template_directory, static_folder=static_directory) app.config.from_object('config.flask_config') def init_db(): """ Initialize the SQLAlchemy database object. :return: A SQLAlchemy instance used universally for database operations. """ return SQLAlchemy(app, session_options={ 'expire_on_commit': False, }) def init_login_manager(): """ Initialize the login manager. :return: A LoginManager instance used universally for flask-login loaders. """ login_manager = LoginManager() login_manager.init_app(app) return login_manager <commit_msg>Fix template and static paths<commit_after>
import os from flask import Flask from flask_login import LoginManager from flask_sqlalchemy import SQLAlchemy template_directory = os.path.join( os.path.dirname(os.path.abspath(__file__)), '../frontend/templates' ) static_directory = '../frontend/static' app = Flask(__name__, template_folder=template_directory, static_folder=static_directory) app.config.from_object('config.flask_config') def init_db(): """ Initialize the SQLAlchemy database object. :return: A SQLAlchemy instance used universally for database operations. """ return SQLAlchemy(app, session_options={ 'expire_on_commit': False, }) def init_login_manager(): """ Initialize the login manager. :return: A LoginManager instance used universally for flask-login loaders. """ login_manager = LoginManager() login_manager.init_app(app) return login_manager
import os from flask import Flask from flask_login import LoginManager from flask_sqlalchemy import SQLAlchemy template_directory = os.path.join( os.path.dirname(os.path.abspath(__file__)), 'frontend/templates' ) static_directory = 'frontend/static' app = Flask(__name__, template_folder=template_directory, static_folder=static_directory) app.config.from_object('config.flask_config') def init_db(): """ Initialize the SQLAlchemy database object. :return: A SQLAlchemy instance used universally for database operations. """ return SQLAlchemy(app, session_options={ 'expire_on_commit': False, }) def init_login_manager(): """ Initialize the login manager. :return: A LoginManager instance used universally for flask-login loaders. """ login_manager = LoginManager() login_manager.init_app(app) return login_manager Fix template and static pathsimport os from flask import Flask from flask_login import LoginManager from flask_sqlalchemy import SQLAlchemy template_directory = os.path.join( os.path.dirname(os.path.abspath(__file__)), '../frontend/templates' ) static_directory = '../frontend/static' app = Flask(__name__, template_folder=template_directory, static_folder=static_directory) app.config.from_object('config.flask_config') def init_db(): """ Initialize the SQLAlchemy database object. :return: A SQLAlchemy instance used universally for database operations. """ return SQLAlchemy(app, session_options={ 'expire_on_commit': False, }) def init_login_manager(): """ Initialize the login manager. :return: A LoginManager instance used universally for flask-login loaders. """ login_manager = LoginManager() login_manager.init_app(app) return login_manager
<commit_before>import os from flask import Flask from flask_login import LoginManager from flask_sqlalchemy import SQLAlchemy template_directory = os.path.join( os.path.dirname(os.path.abspath(__file__)), 'frontend/templates' ) static_directory = 'frontend/static' app = Flask(__name__, template_folder=template_directory, static_folder=static_directory) app.config.from_object('config.flask_config') def init_db(): """ Initialize the SQLAlchemy database object. :return: A SQLAlchemy instance used universally for database operations. """ return SQLAlchemy(app, session_options={ 'expire_on_commit': False, }) def init_login_manager(): """ Initialize the login manager. :return: A LoginManager instance used universally for flask-login loaders. """ login_manager = LoginManager() login_manager.init_app(app) return login_manager <commit_msg>Fix template and static paths<commit_after>import os from flask import Flask from flask_login import LoginManager from flask_sqlalchemy import SQLAlchemy template_directory = os.path.join( os.path.dirname(os.path.abspath(__file__)), '../frontend/templates' ) static_directory = '../frontend/static' app = Flask(__name__, template_folder=template_directory, static_folder=static_directory) app.config.from_object('config.flask_config') def init_db(): """ Initialize the SQLAlchemy database object. :return: A SQLAlchemy instance used universally for database operations. """ return SQLAlchemy(app, session_options={ 'expire_on_commit': False, }) def init_login_manager(): """ Initialize the login manager. :return: A LoginManager instance used universally for flask-login loaders. """ login_manager = LoginManager() login_manager.init_app(app) return login_manager
f5fe10907d1ecffb89f62a56e8b7f8e34f4fcf2a
urls.py
urls.py
from django.conf.urls.defaults import * # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() urlpatterns = patterns('', # Example: # (r'^store/', include('store.foo.urls')), # Uncomment the admin/doc line below to enable admin documentation: (r'^admin/doc/', include('django.contrib.admindocs.urls')), # Uncomment the next line to enable the admin: (r'^admin/', include(admin.site.urls)), (r'^/', 'store.views.frontpage'), )
from django.conf.urls.defaults import * # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() urlpatterns = patterns('', # Example: # (r'^store/', include('store.foo.urls')), # Uncomment the admin/doc line below to enable admin documentation: (r'^admin/doc/', include('django.contrib.admindocs.urls')), # Uncomment the next line to enable the admin: (r'^admin/', include(admin.site.urls)), (r'', 'store.views.frontpage'), )
Make catch-all actually catch all
Make catch-all actually catch all
Python
bsd-3-clause
willmurnane/store
from django.conf.urls.defaults import * # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() urlpatterns = patterns('', # Example: # (r'^store/', include('store.foo.urls')), # Uncomment the admin/doc line below to enable admin documentation: (r'^admin/doc/', include('django.contrib.admindocs.urls')), # Uncomment the next line to enable the admin: (r'^admin/', include(admin.site.urls)), (r'^/', 'store.views.frontpage'), ) Make catch-all actually catch all
from django.conf.urls.defaults import * # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() urlpatterns = patterns('', # Example: # (r'^store/', include('store.foo.urls')), # Uncomment the admin/doc line below to enable admin documentation: (r'^admin/doc/', include('django.contrib.admindocs.urls')), # Uncomment the next line to enable the admin: (r'^admin/', include(admin.site.urls)), (r'', 'store.views.frontpage'), )
<commit_before>from django.conf.urls.defaults import * # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() urlpatterns = patterns('', # Example: # (r'^store/', include('store.foo.urls')), # Uncomment the admin/doc line below to enable admin documentation: (r'^admin/doc/', include('django.contrib.admindocs.urls')), # Uncomment the next line to enable the admin: (r'^admin/', include(admin.site.urls)), (r'^/', 'store.views.frontpage'), ) <commit_msg>Make catch-all actually catch all<commit_after>
from django.conf.urls.defaults import * # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() urlpatterns = patterns('', # Example: # (r'^store/', include('store.foo.urls')), # Uncomment the admin/doc line below to enable admin documentation: (r'^admin/doc/', include('django.contrib.admindocs.urls')), # Uncomment the next line to enable the admin: (r'^admin/', include(admin.site.urls)), (r'', 'store.views.frontpage'), )
from django.conf.urls.defaults import * # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() urlpatterns = patterns('', # Example: # (r'^store/', include('store.foo.urls')), # Uncomment the admin/doc line below to enable admin documentation: (r'^admin/doc/', include('django.contrib.admindocs.urls')), # Uncomment the next line to enable the admin: (r'^admin/', include(admin.site.urls)), (r'^/', 'store.views.frontpage'), ) Make catch-all actually catch allfrom django.conf.urls.defaults import * # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() urlpatterns = patterns('', # Example: # (r'^store/', include('store.foo.urls')), # Uncomment the admin/doc line below to enable admin documentation: (r'^admin/doc/', include('django.contrib.admindocs.urls')), # Uncomment the next line to enable the admin: (r'^admin/', include(admin.site.urls)), (r'', 'store.views.frontpage'), )
<commit_before>from django.conf.urls.defaults import * # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() urlpatterns = patterns('', # Example: # (r'^store/', include('store.foo.urls')), # Uncomment the admin/doc line below to enable admin documentation: (r'^admin/doc/', include('django.contrib.admindocs.urls')), # Uncomment the next line to enable the admin: (r'^admin/', include(admin.site.urls)), (r'^/', 'store.views.frontpage'), ) <commit_msg>Make catch-all actually catch all<commit_after>from django.conf.urls.defaults import * # Uncomment the next two lines to enable the admin: from django.contrib import admin admin.autodiscover() urlpatterns = patterns('', # Example: # (r'^store/', include('store.foo.urls')), # Uncomment the admin/doc line below to enable admin documentation: (r'^admin/doc/', include('django.contrib.admindocs.urls')), # Uncomment the next line to enable the admin: (r'^admin/', include(admin.site.urls)), (r'', 'store.views.frontpage'), )
cebd4d613cc95ef6e775581a6f77f4850e39020a
src/mdformat/_conf.py
src/mdformat/_conf.py
from __future__ import annotations import functools from pathlib import Path from typing import Mapping import tomli DEFAULT_OPTS = { "wrap": "keep", "number": False, "end_of_line": "lf", } class InvalidConfError(Exception): """Error raised given invalid TOML or a key that is not valid for mdformat.""" @functools.lru_cache() def read_toml_opts(conf_dir: Path) -> Mapping: conf_path = conf_dir / ".mdformat.toml" if not conf_path.is_file(): parent_dir = conf_dir.parent if conf_dir == parent_dir: return {} return read_toml_opts(parent_dir) with open(conf_path, "rb") as f: try: toml_opts = tomli.load(f) except tomli.TOMLDecodeError as e: raise InvalidConfError(f"Invalid TOML syntax: {e}") for key in toml_opts: if key not in DEFAULT_OPTS: raise InvalidConfError(f"Invalid key {key!r} in {conf_path}") return toml_opts
from __future__ import annotations import functools from pathlib import Path from typing import Mapping import tomli DEFAULT_OPTS = { "wrap": "keep", "number": False, "end_of_line": "lf", } class InvalidConfError(Exception): """Error raised given invalid TOML or a key that is not valid for mdformat.""" @functools.lru_cache() def read_toml_opts(conf_dir: Path) -> Mapping: conf_path = conf_dir / ".mdformat.toml" if not conf_path.is_file(): parent_dir = conf_dir.parent if conf_dir == parent_dir: return {} return read_toml_opts(parent_dir) with open(conf_path, "rb") as f: try: toml_opts = tomli.load(f) except tomli.TOMLDecodeError as e: raise InvalidConfError(f"Invalid TOML syntax: {e}") for key in toml_opts: if key not in DEFAULT_OPTS: raise InvalidConfError( f"Invalid key {key!r} in {conf_path}." f" Keys must be one of {set(DEFAULT_OPTS)}." ) return toml_opts
Improve error message on invalid conf key
Improve error message on invalid conf key
Python
mit
executablebooks/mdformat
from __future__ import annotations import functools from pathlib import Path from typing import Mapping import tomli DEFAULT_OPTS = { "wrap": "keep", "number": False, "end_of_line": "lf", } class InvalidConfError(Exception): """Error raised given invalid TOML or a key that is not valid for mdformat.""" @functools.lru_cache() def read_toml_opts(conf_dir: Path) -> Mapping: conf_path = conf_dir / ".mdformat.toml" if not conf_path.is_file(): parent_dir = conf_dir.parent if conf_dir == parent_dir: return {} return read_toml_opts(parent_dir) with open(conf_path, "rb") as f: try: toml_opts = tomli.load(f) except tomli.TOMLDecodeError as e: raise InvalidConfError(f"Invalid TOML syntax: {e}") for key in toml_opts: if key not in DEFAULT_OPTS: raise InvalidConfError(f"Invalid key {key!r} in {conf_path}") return toml_opts Improve error message on invalid conf key
from __future__ import annotations import functools from pathlib import Path from typing import Mapping import tomli DEFAULT_OPTS = { "wrap": "keep", "number": False, "end_of_line": "lf", } class InvalidConfError(Exception): """Error raised given invalid TOML or a key that is not valid for mdformat.""" @functools.lru_cache() def read_toml_opts(conf_dir: Path) -> Mapping: conf_path = conf_dir / ".mdformat.toml" if not conf_path.is_file(): parent_dir = conf_dir.parent if conf_dir == parent_dir: return {} return read_toml_opts(parent_dir) with open(conf_path, "rb") as f: try: toml_opts = tomli.load(f) except tomli.TOMLDecodeError as e: raise InvalidConfError(f"Invalid TOML syntax: {e}") for key in toml_opts: if key not in DEFAULT_OPTS: raise InvalidConfError( f"Invalid key {key!r} in {conf_path}." f" Keys must be one of {set(DEFAULT_OPTS)}." ) return toml_opts
<commit_before>from __future__ import annotations import functools from pathlib import Path from typing import Mapping import tomli DEFAULT_OPTS = { "wrap": "keep", "number": False, "end_of_line": "lf", } class InvalidConfError(Exception): """Error raised given invalid TOML or a key that is not valid for mdformat.""" @functools.lru_cache() def read_toml_opts(conf_dir: Path) -> Mapping: conf_path = conf_dir / ".mdformat.toml" if not conf_path.is_file(): parent_dir = conf_dir.parent if conf_dir == parent_dir: return {} return read_toml_opts(parent_dir) with open(conf_path, "rb") as f: try: toml_opts = tomli.load(f) except tomli.TOMLDecodeError as e: raise InvalidConfError(f"Invalid TOML syntax: {e}") for key in toml_opts: if key not in DEFAULT_OPTS: raise InvalidConfError(f"Invalid key {key!r} in {conf_path}") return toml_opts <commit_msg>Improve error message on invalid conf key<commit_after>
from __future__ import annotations import functools from pathlib import Path from typing import Mapping import tomli DEFAULT_OPTS = { "wrap": "keep", "number": False, "end_of_line": "lf", } class InvalidConfError(Exception): """Error raised given invalid TOML or a key that is not valid for mdformat.""" @functools.lru_cache() def read_toml_opts(conf_dir: Path) -> Mapping: conf_path = conf_dir / ".mdformat.toml" if not conf_path.is_file(): parent_dir = conf_dir.parent if conf_dir == parent_dir: return {} return read_toml_opts(parent_dir) with open(conf_path, "rb") as f: try: toml_opts = tomli.load(f) except tomli.TOMLDecodeError as e: raise InvalidConfError(f"Invalid TOML syntax: {e}") for key in toml_opts: if key not in DEFAULT_OPTS: raise InvalidConfError( f"Invalid key {key!r} in {conf_path}." f" Keys must be one of {set(DEFAULT_OPTS)}." ) return toml_opts
from __future__ import annotations import functools from pathlib import Path from typing import Mapping import tomli DEFAULT_OPTS = { "wrap": "keep", "number": False, "end_of_line": "lf", } class InvalidConfError(Exception): """Error raised given invalid TOML or a key that is not valid for mdformat.""" @functools.lru_cache() def read_toml_opts(conf_dir: Path) -> Mapping: conf_path = conf_dir / ".mdformat.toml" if not conf_path.is_file(): parent_dir = conf_dir.parent if conf_dir == parent_dir: return {} return read_toml_opts(parent_dir) with open(conf_path, "rb") as f: try: toml_opts = tomli.load(f) except tomli.TOMLDecodeError as e: raise InvalidConfError(f"Invalid TOML syntax: {e}") for key in toml_opts: if key not in DEFAULT_OPTS: raise InvalidConfError(f"Invalid key {key!r} in {conf_path}") return toml_opts Improve error message on invalid conf keyfrom __future__ import annotations import functools from pathlib import Path from typing import Mapping import tomli DEFAULT_OPTS = { "wrap": "keep", "number": False, "end_of_line": "lf", } class InvalidConfError(Exception): """Error raised given invalid TOML or a key that is not valid for mdformat.""" @functools.lru_cache() def read_toml_opts(conf_dir: Path) -> Mapping: conf_path = conf_dir / ".mdformat.toml" if not conf_path.is_file(): parent_dir = conf_dir.parent if conf_dir == parent_dir: return {} return read_toml_opts(parent_dir) with open(conf_path, "rb") as f: try: toml_opts = tomli.load(f) except tomli.TOMLDecodeError as e: raise InvalidConfError(f"Invalid TOML syntax: {e}") for key in toml_opts: if key not in DEFAULT_OPTS: raise InvalidConfError( f"Invalid key {key!r} in {conf_path}." f" Keys must be one of {set(DEFAULT_OPTS)}." ) return toml_opts
<commit_before>from __future__ import annotations import functools from pathlib import Path from typing import Mapping import tomli DEFAULT_OPTS = { "wrap": "keep", "number": False, "end_of_line": "lf", } class InvalidConfError(Exception): """Error raised given invalid TOML or a key that is not valid for mdformat.""" @functools.lru_cache() def read_toml_opts(conf_dir: Path) -> Mapping: conf_path = conf_dir / ".mdformat.toml" if not conf_path.is_file(): parent_dir = conf_dir.parent if conf_dir == parent_dir: return {} return read_toml_opts(parent_dir) with open(conf_path, "rb") as f: try: toml_opts = tomli.load(f) except tomli.TOMLDecodeError as e: raise InvalidConfError(f"Invalid TOML syntax: {e}") for key in toml_opts: if key not in DEFAULT_OPTS: raise InvalidConfError(f"Invalid key {key!r} in {conf_path}") return toml_opts <commit_msg>Improve error message on invalid conf key<commit_after>from __future__ import annotations import functools from pathlib import Path from typing import Mapping import tomli DEFAULT_OPTS = { "wrap": "keep", "number": False, "end_of_line": "lf", } class InvalidConfError(Exception): """Error raised given invalid TOML or a key that is not valid for mdformat.""" @functools.lru_cache() def read_toml_opts(conf_dir: Path) -> Mapping: conf_path = conf_dir / ".mdformat.toml" if not conf_path.is_file(): parent_dir = conf_dir.parent if conf_dir == parent_dir: return {} return read_toml_opts(parent_dir) with open(conf_path, "rb") as f: try: toml_opts = tomli.load(f) except tomli.TOMLDecodeError as e: raise InvalidConfError(f"Invalid TOML syntax: {e}") for key in toml_opts: if key not in DEFAULT_OPTS: raise InvalidConfError( f"Invalid key {key!r} in {conf_path}." f" Keys must be one of {set(DEFAULT_OPTS)}." ) return toml_opts
20151a50424bbb6c4edcab5f19b97e3d7fb838b9
plugins/GCodeReader/__init__.py
plugins/GCodeReader/__init__.py
# Copyright (c) 2015 Ultimaker B.V. # Cura is released under the terms of the AGPLv3 or higher. from . import GCodeReader from UM.i18n import i18nCatalog catalog = i18nCatalog("cura") def getMetaData(): return { "plugin": { "name": catalog.i18nc("@label", "GCode Reader"), "author": "Ultimaker", "version": "1.0", "description": catalog.i18nc("@info:whatsthis", "Provides support for reading GCode files."), "api": 2 }, "mesh_reader": [ { "extension": "gcode", "description": catalog.i18nc("@item:inlistbox", "Gcode File") } ] } def register(app): return { "mesh_reader": GCodeReader.GCodeReader() }
# Copyright (c) 2015 Ultimaker B.V. # Cura is released under the terms of the AGPLv3 or higher. from . import GCodeReader from UM.i18n import i18nCatalog catalog = i18nCatalog("cura") def getMetaData(): return { "plugin": { "name": catalog.i18nc("@label", "GCode Reader"), "author": "Ultimaker", "version": "1.0", "description": catalog.i18nc("@info:whatsthis", "Provides support for reading GCode files."), "api": 2 }, "profile_reader": { "extension": "gcode", "description": catalog.i18nc("@item:inlistbox", "Gcode File") } } def register(app): return { "profile_reader": GCodeReader.GCodeReader() }
Change plugin type to profile_reader
Change plugin type to profile_reader This repairs the profile reading at startup. It should not be a mesh reader. Contributes to issue CURA-34.
Python
agpl-3.0
Curahelper/Cura,fieldOfView/Cura,hmflash/Cura,ynotstartups/Wanhao,hmflash/Cura,senttech/Cura,Curahelper/Cura,ynotstartups/Wanhao,fieldOfView/Cura,totalretribution/Cura,totalretribution/Cura,senttech/Cura
# Copyright (c) 2015 Ultimaker B.V. # Cura is released under the terms of the AGPLv3 or higher. from . import GCodeReader from UM.i18n import i18nCatalog catalog = i18nCatalog("cura") def getMetaData(): return { "plugin": { "name": catalog.i18nc("@label", "GCode Reader"), "author": "Ultimaker", "version": "1.0", "description": catalog.i18nc("@info:whatsthis", "Provides support for reading GCode files."), "api": 2 }, "mesh_reader": [ { "extension": "gcode", "description": catalog.i18nc("@item:inlistbox", "Gcode File") } ] } def register(app): return { "mesh_reader": GCodeReader.GCodeReader() } Change plugin type to profile_reader This repairs the profile reading at startup. It should not be a mesh reader. Contributes to issue CURA-34.
# Copyright (c) 2015 Ultimaker B.V. # Cura is released under the terms of the AGPLv3 or higher. from . import GCodeReader from UM.i18n import i18nCatalog catalog = i18nCatalog("cura") def getMetaData(): return { "plugin": { "name": catalog.i18nc("@label", "GCode Reader"), "author": "Ultimaker", "version": "1.0", "description": catalog.i18nc("@info:whatsthis", "Provides support for reading GCode files."), "api": 2 }, "profile_reader": { "extension": "gcode", "description": catalog.i18nc("@item:inlistbox", "Gcode File") } } def register(app): return { "profile_reader": GCodeReader.GCodeReader() }
<commit_before># Copyright (c) 2015 Ultimaker B.V. # Cura is released under the terms of the AGPLv3 or higher. from . import GCodeReader from UM.i18n import i18nCatalog catalog = i18nCatalog("cura") def getMetaData(): return { "plugin": { "name": catalog.i18nc("@label", "GCode Reader"), "author": "Ultimaker", "version": "1.0", "description": catalog.i18nc("@info:whatsthis", "Provides support for reading GCode files."), "api": 2 }, "mesh_reader": [ { "extension": "gcode", "description": catalog.i18nc("@item:inlistbox", "Gcode File") } ] } def register(app): return { "mesh_reader": GCodeReader.GCodeReader() } <commit_msg>Change plugin type to profile_reader This repairs the profile reading at startup. It should not be a mesh reader. Contributes to issue CURA-34.<commit_after>
# Copyright (c) 2015 Ultimaker B.V. # Cura is released under the terms of the AGPLv3 or higher. from . import GCodeReader from UM.i18n import i18nCatalog catalog = i18nCatalog("cura") def getMetaData(): return { "plugin": { "name": catalog.i18nc("@label", "GCode Reader"), "author": "Ultimaker", "version": "1.0", "description": catalog.i18nc("@info:whatsthis", "Provides support for reading GCode files."), "api": 2 }, "profile_reader": { "extension": "gcode", "description": catalog.i18nc("@item:inlistbox", "Gcode File") } } def register(app): return { "profile_reader": GCodeReader.GCodeReader() }
# Copyright (c) 2015 Ultimaker B.V. # Cura is released under the terms of the AGPLv3 or higher. from . import GCodeReader from UM.i18n import i18nCatalog catalog = i18nCatalog("cura") def getMetaData(): return { "plugin": { "name": catalog.i18nc("@label", "GCode Reader"), "author": "Ultimaker", "version": "1.0", "description": catalog.i18nc("@info:whatsthis", "Provides support for reading GCode files."), "api": 2 }, "mesh_reader": [ { "extension": "gcode", "description": catalog.i18nc("@item:inlistbox", "Gcode File") } ] } def register(app): return { "mesh_reader": GCodeReader.GCodeReader() } Change plugin type to profile_reader This repairs the profile reading at startup. It should not be a mesh reader. Contributes to issue CURA-34.# Copyright (c) 2015 Ultimaker B.V. # Cura is released under the terms of the AGPLv3 or higher. from . import GCodeReader from UM.i18n import i18nCatalog catalog = i18nCatalog("cura") def getMetaData(): return { "plugin": { "name": catalog.i18nc("@label", "GCode Reader"), "author": "Ultimaker", "version": "1.0", "description": catalog.i18nc("@info:whatsthis", "Provides support for reading GCode files."), "api": 2 }, "profile_reader": { "extension": "gcode", "description": catalog.i18nc("@item:inlistbox", "Gcode File") } } def register(app): return { "profile_reader": GCodeReader.GCodeReader() }
<commit_before># Copyright (c) 2015 Ultimaker B.V. # Cura is released under the terms of the AGPLv3 or higher. from . import GCodeReader from UM.i18n import i18nCatalog catalog = i18nCatalog("cura") def getMetaData(): return { "plugin": { "name": catalog.i18nc("@label", "GCode Reader"), "author": "Ultimaker", "version": "1.0", "description": catalog.i18nc("@info:whatsthis", "Provides support for reading GCode files."), "api": 2 }, "mesh_reader": [ { "extension": "gcode", "description": catalog.i18nc("@item:inlistbox", "Gcode File") } ] } def register(app): return { "mesh_reader": GCodeReader.GCodeReader() } <commit_msg>Change plugin type to profile_reader This repairs the profile reading at startup. It should not be a mesh reader. Contributes to issue CURA-34.<commit_after># Copyright (c) 2015 Ultimaker B.V. # Cura is released under the terms of the AGPLv3 or higher. from . import GCodeReader from UM.i18n import i18nCatalog catalog = i18nCatalog("cura") def getMetaData(): return { "plugin": { "name": catalog.i18nc("@label", "GCode Reader"), "author": "Ultimaker", "version": "1.0", "description": catalog.i18nc("@info:whatsthis", "Provides support for reading GCode files."), "api": 2 }, "profile_reader": { "extension": "gcode", "description": catalog.i18nc("@item:inlistbox", "Gcode File") } } def register(app): return { "profile_reader": GCodeReader.GCodeReader() }
715098531f823c3b2932e6a03d2e4b113bd53ed9
tests/test_grammar.py
tests/test_grammar.py
import viper.grammar as vg import viper.grammar.languages as vgl import viper.lexer as vl import pytest @pytest.mark.parametrize('line,sppf', [ ('foo', vgl.SPPF(vgl.ParseTreeChar(vl.Name('foo')))), ('2', vgl.SPPF(vgl.ParseTreeChar(vl.Number('2')))), ('...', vgl.SPPF(vgl.ParseTreeChar(vl.Operator('...')))), ('Zilch', vgl.SPPF(vgl.ParseTreeChar(vl.Class('Zilch')))), ('True', vgl.SPPF(vgl.ParseTreeChar(vl.Class('True')))), ('False', vgl.SPPF(vgl.ParseTreeChar(vl.Class('False')))), ]) def test_atom(line: str, sppf: vgl.SPPF): atom = vg.GRAMMAR.get_rule('atom') lexemes = vl.lex_line(line) assert sppf == vg.make_sppf(atom, lexemes)
import viper.grammar as vg import viper.lexer as vl from viper.grammar.languages import ( SPPF, ParseTreeEmpty as PTE, ParseTreeChar as PTC, ParseTreePair as PTP, ParseTreeRep as PTR ) import pytest @pytest.mark.parametrize('line,sppf', [ ('foo', SPPF(PTC(vl.Name('foo')))), ('42', SPPF(PTC(vl.Number('42')))), ('...', SPPF(PTC(vl.Operator('...')))), ('Zilch', SPPF(PTC(vl.Class('Zilch')))), ('True', SPPF(PTC(vl.Class('True')))), ('False', SPPF(PTC(vl.Class('False')))), ('()', SPPF(PTP(SPPF(PTC(vl.OpenParen())), SPPF(PTC(vl.CloseParen()))))), ('(foo)', SPPF(PTP(SPPF(PTC(vl.OpenParen())), SPPF(PTP(SPPF(PTP(SPPF(PTC(vl.Name('foo'))), SPPF(PTR(SPPF())))), SPPF(PTC(vl.CloseParen()))))))), ]) def test_atom(line: str, sppf: SPPF): atom = vg.GRAMMAR.get_rule('atom') lexemes = vl.lex_line(line) assert sppf == vg.make_sppf(atom, lexemes)
Revise grammar tests for atom
Revise grammar tests for atom
Python
apache-2.0
pdarragh/Viper
import viper.grammar as vg import viper.grammar.languages as vgl import viper.lexer as vl import pytest @pytest.mark.parametrize('line,sppf', [ ('foo', vgl.SPPF(vgl.ParseTreeChar(vl.Name('foo')))), ('2', vgl.SPPF(vgl.ParseTreeChar(vl.Number('2')))), ('...', vgl.SPPF(vgl.ParseTreeChar(vl.Operator('...')))), ('Zilch', vgl.SPPF(vgl.ParseTreeChar(vl.Class('Zilch')))), ('True', vgl.SPPF(vgl.ParseTreeChar(vl.Class('True')))), ('False', vgl.SPPF(vgl.ParseTreeChar(vl.Class('False')))), ]) def test_atom(line: str, sppf: vgl.SPPF): atom = vg.GRAMMAR.get_rule('atom') lexemes = vl.lex_line(line) assert sppf == vg.make_sppf(atom, lexemes) Revise grammar tests for atom
import viper.grammar as vg import viper.lexer as vl from viper.grammar.languages import ( SPPF, ParseTreeEmpty as PTE, ParseTreeChar as PTC, ParseTreePair as PTP, ParseTreeRep as PTR ) import pytest @pytest.mark.parametrize('line,sppf', [ ('foo', SPPF(PTC(vl.Name('foo')))), ('42', SPPF(PTC(vl.Number('42')))), ('...', SPPF(PTC(vl.Operator('...')))), ('Zilch', SPPF(PTC(vl.Class('Zilch')))), ('True', SPPF(PTC(vl.Class('True')))), ('False', SPPF(PTC(vl.Class('False')))), ('()', SPPF(PTP(SPPF(PTC(vl.OpenParen())), SPPF(PTC(vl.CloseParen()))))), ('(foo)', SPPF(PTP(SPPF(PTC(vl.OpenParen())), SPPF(PTP(SPPF(PTP(SPPF(PTC(vl.Name('foo'))), SPPF(PTR(SPPF())))), SPPF(PTC(vl.CloseParen()))))))), ]) def test_atom(line: str, sppf: SPPF): atom = vg.GRAMMAR.get_rule('atom') lexemes = vl.lex_line(line) assert sppf == vg.make_sppf(atom, lexemes)
<commit_before>import viper.grammar as vg import viper.grammar.languages as vgl import viper.lexer as vl import pytest @pytest.mark.parametrize('line,sppf', [ ('foo', vgl.SPPF(vgl.ParseTreeChar(vl.Name('foo')))), ('2', vgl.SPPF(vgl.ParseTreeChar(vl.Number('2')))), ('...', vgl.SPPF(vgl.ParseTreeChar(vl.Operator('...')))), ('Zilch', vgl.SPPF(vgl.ParseTreeChar(vl.Class('Zilch')))), ('True', vgl.SPPF(vgl.ParseTreeChar(vl.Class('True')))), ('False', vgl.SPPF(vgl.ParseTreeChar(vl.Class('False')))), ]) def test_atom(line: str, sppf: vgl.SPPF): atom = vg.GRAMMAR.get_rule('atom') lexemes = vl.lex_line(line) assert sppf == vg.make_sppf(atom, lexemes) <commit_msg>Revise grammar tests for atom<commit_after>
import viper.grammar as vg import viper.lexer as vl from viper.grammar.languages import ( SPPF, ParseTreeEmpty as PTE, ParseTreeChar as PTC, ParseTreePair as PTP, ParseTreeRep as PTR ) import pytest @pytest.mark.parametrize('line,sppf', [ ('foo', SPPF(PTC(vl.Name('foo')))), ('42', SPPF(PTC(vl.Number('42')))), ('...', SPPF(PTC(vl.Operator('...')))), ('Zilch', SPPF(PTC(vl.Class('Zilch')))), ('True', SPPF(PTC(vl.Class('True')))), ('False', SPPF(PTC(vl.Class('False')))), ('()', SPPF(PTP(SPPF(PTC(vl.OpenParen())), SPPF(PTC(vl.CloseParen()))))), ('(foo)', SPPF(PTP(SPPF(PTC(vl.OpenParen())), SPPF(PTP(SPPF(PTP(SPPF(PTC(vl.Name('foo'))), SPPF(PTR(SPPF())))), SPPF(PTC(vl.CloseParen()))))))), ]) def test_atom(line: str, sppf: SPPF): atom = vg.GRAMMAR.get_rule('atom') lexemes = vl.lex_line(line) assert sppf == vg.make_sppf(atom, lexemes)
import viper.grammar as vg import viper.grammar.languages as vgl import viper.lexer as vl import pytest @pytest.mark.parametrize('line,sppf', [ ('foo', vgl.SPPF(vgl.ParseTreeChar(vl.Name('foo')))), ('2', vgl.SPPF(vgl.ParseTreeChar(vl.Number('2')))), ('...', vgl.SPPF(vgl.ParseTreeChar(vl.Operator('...')))), ('Zilch', vgl.SPPF(vgl.ParseTreeChar(vl.Class('Zilch')))), ('True', vgl.SPPF(vgl.ParseTreeChar(vl.Class('True')))), ('False', vgl.SPPF(vgl.ParseTreeChar(vl.Class('False')))), ]) def test_atom(line: str, sppf: vgl.SPPF): atom = vg.GRAMMAR.get_rule('atom') lexemes = vl.lex_line(line) assert sppf == vg.make_sppf(atom, lexemes) Revise grammar tests for atomimport viper.grammar as vg import viper.lexer as vl from viper.grammar.languages import ( SPPF, ParseTreeEmpty as PTE, ParseTreeChar as PTC, ParseTreePair as PTP, ParseTreeRep as PTR ) import pytest @pytest.mark.parametrize('line,sppf', [ ('foo', SPPF(PTC(vl.Name('foo')))), ('42', SPPF(PTC(vl.Number('42')))), ('...', SPPF(PTC(vl.Operator('...')))), ('Zilch', SPPF(PTC(vl.Class('Zilch')))), ('True', SPPF(PTC(vl.Class('True')))), ('False', SPPF(PTC(vl.Class('False')))), ('()', SPPF(PTP(SPPF(PTC(vl.OpenParen())), SPPF(PTC(vl.CloseParen()))))), ('(foo)', SPPF(PTP(SPPF(PTC(vl.OpenParen())), SPPF(PTP(SPPF(PTP(SPPF(PTC(vl.Name('foo'))), SPPF(PTR(SPPF())))), SPPF(PTC(vl.CloseParen()))))))), ]) def test_atom(line: str, sppf: SPPF): atom = vg.GRAMMAR.get_rule('atom') lexemes = vl.lex_line(line) assert sppf == vg.make_sppf(atom, lexemes)
<commit_before>import viper.grammar as vg import viper.grammar.languages as vgl import viper.lexer as vl import pytest @pytest.mark.parametrize('line,sppf', [ ('foo', vgl.SPPF(vgl.ParseTreeChar(vl.Name('foo')))), ('2', vgl.SPPF(vgl.ParseTreeChar(vl.Number('2')))), ('...', vgl.SPPF(vgl.ParseTreeChar(vl.Operator('...')))), ('Zilch', vgl.SPPF(vgl.ParseTreeChar(vl.Class('Zilch')))), ('True', vgl.SPPF(vgl.ParseTreeChar(vl.Class('True')))), ('False', vgl.SPPF(vgl.ParseTreeChar(vl.Class('False')))), ]) def test_atom(line: str, sppf: vgl.SPPF): atom = vg.GRAMMAR.get_rule('atom') lexemes = vl.lex_line(line) assert sppf == vg.make_sppf(atom, lexemes) <commit_msg>Revise grammar tests for atom<commit_after>import viper.grammar as vg import viper.lexer as vl from viper.grammar.languages import ( SPPF, ParseTreeEmpty as PTE, ParseTreeChar as PTC, ParseTreePair as PTP, ParseTreeRep as PTR ) import pytest @pytest.mark.parametrize('line,sppf', [ ('foo', SPPF(PTC(vl.Name('foo')))), ('42', SPPF(PTC(vl.Number('42')))), ('...', SPPF(PTC(vl.Operator('...')))), ('Zilch', SPPF(PTC(vl.Class('Zilch')))), ('True', SPPF(PTC(vl.Class('True')))), ('False', SPPF(PTC(vl.Class('False')))), ('()', SPPF(PTP(SPPF(PTC(vl.OpenParen())), SPPF(PTC(vl.CloseParen()))))), ('(foo)', SPPF(PTP(SPPF(PTC(vl.OpenParen())), SPPF(PTP(SPPF(PTP(SPPF(PTC(vl.Name('foo'))), SPPF(PTR(SPPF())))), SPPF(PTC(vl.CloseParen()))))))), ]) def test_atom(line: str, sppf: SPPF): atom = vg.GRAMMAR.get_rule('atom') lexemes = vl.lex_line(line) assert sppf == vg.make_sppf(atom, lexemes)
eca9ee90bf64b14c6a8eacdb4197825790ab7825
tests/test_helpers.py
tests/test_helpers.py
""" Tests for the NURBS-Python package Released under The MIT License. See LICENSE file for details. Copyright (c) 2018 Onur Rauf Bingol Tests geomdl.helpers module. """ from geomdl import helpers GEOMDL_DELTA = 10e-8 def test_basis_function_one(): degree = 2 knot_vector = [0, 0, 0, 1, 2, 3, 4, 4, 5, 5, 5] knot = 5. / 2. span = 3 to_check = helpers.basis_function_one(degree, knot_vector, span, knot) result = 0.75 assert to_check == result def test_basis_function_ders_one(): degree = 2 knot_vector = [0, 0, 0, 1, 2, 3, 4, 4, 5, 5, 5] knot = 5. / 2. span = 4 to_check = helpers.basis_function_ders_one(degree, knot_vector, 4, knot, 2) result = [0.125, 0.5, 1.0] assert to_check == result
Add tests for A2.4 and A2.5
Add tests for A2.4 and A2.5
Python
mit
orbingol/NURBS-Python,orbingol/NURBS-Python
Add tests for A2.4 and A2.5
""" Tests for the NURBS-Python package Released under The MIT License. See LICENSE file for details. Copyright (c) 2018 Onur Rauf Bingol Tests geomdl.helpers module. """ from geomdl import helpers GEOMDL_DELTA = 10e-8 def test_basis_function_one(): degree = 2 knot_vector = [0, 0, 0, 1, 2, 3, 4, 4, 5, 5, 5] knot = 5. / 2. span = 3 to_check = helpers.basis_function_one(degree, knot_vector, span, knot) result = 0.75 assert to_check == result def test_basis_function_ders_one(): degree = 2 knot_vector = [0, 0, 0, 1, 2, 3, 4, 4, 5, 5, 5] knot = 5. / 2. span = 4 to_check = helpers.basis_function_ders_one(degree, knot_vector, 4, knot, 2) result = [0.125, 0.5, 1.0] assert to_check == result
<commit_before><commit_msg>Add tests for A2.4 and A2.5<commit_after>
""" Tests for the NURBS-Python package Released under The MIT License. See LICENSE file for details. Copyright (c) 2018 Onur Rauf Bingol Tests geomdl.helpers module. """ from geomdl import helpers GEOMDL_DELTA = 10e-8 def test_basis_function_one(): degree = 2 knot_vector = [0, 0, 0, 1, 2, 3, 4, 4, 5, 5, 5] knot = 5. / 2. span = 3 to_check = helpers.basis_function_one(degree, knot_vector, span, knot) result = 0.75 assert to_check == result def test_basis_function_ders_one(): degree = 2 knot_vector = [0, 0, 0, 1, 2, 3, 4, 4, 5, 5, 5] knot = 5. / 2. span = 4 to_check = helpers.basis_function_ders_one(degree, knot_vector, 4, knot, 2) result = [0.125, 0.5, 1.0] assert to_check == result
Add tests for A2.4 and A2.5""" Tests for the NURBS-Python package Released under The MIT License. See LICENSE file for details. Copyright (c) 2018 Onur Rauf Bingol Tests geomdl.helpers module. """ from geomdl import helpers GEOMDL_DELTA = 10e-8 def test_basis_function_one(): degree = 2 knot_vector = [0, 0, 0, 1, 2, 3, 4, 4, 5, 5, 5] knot = 5. / 2. span = 3 to_check = helpers.basis_function_one(degree, knot_vector, span, knot) result = 0.75 assert to_check == result def test_basis_function_ders_one(): degree = 2 knot_vector = [0, 0, 0, 1, 2, 3, 4, 4, 5, 5, 5] knot = 5. / 2. span = 4 to_check = helpers.basis_function_ders_one(degree, knot_vector, 4, knot, 2) result = [0.125, 0.5, 1.0] assert to_check == result
<commit_before><commit_msg>Add tests for A2.4 and A2.5<commit_after>""" Tests for the NURBS-Python package Released under The MIT License. See LICENSE file for details. Copyright (c) 2018 Onur Rauf Bingol Tests geomdl.helpers module. """ from geomdl import helpers GEOMDL_DELTA = 10e-8 def test_basis_function_one(): degree = 2 knot_vector = [0, 0, 0, 1, 2, 3, 4, 4, 5, 5, 5] knot = 5. / 2. span = 3 to_check = helpers.basis_function_one(degree, knot_vector, span, knot) result = 0.75 assert to_check == result def test_basis_function_ders_one(): degree = 2 knot_vector = [0, 0, 0, 1, 2, 3, 4, 4, 5, 5, 5] knot = 5. / 2. span = 4 to_check = helpers.basis_function_ders_one(degree, knot_vector, 4, knot, 2) result = [0.125, 0.5, 1.0] assert to_check == result
99683d16551450686397f953668b7d6bf4167a5e
tests/test_methods.py
tests/test_methods.py
import interleaving as il import numpy as np np.random.seed(0) class TestMethods(object): def assert_almost_equal(self, a, b, error_rate=0.1): half_error_rate = error_rate / 2.0 lower_bound = (1.0 - half_error_rate) * a upper_bound = (1.0 + half_error_rate) * a assert lower_bound <= b and b <= upper_bound def interleave(self, method, a, b, ideals, num=100): results = [] for i in range(num): res = method().interleave(a, b) results.append(tuple(res)) results = set(results) possible_results = set([tuple(i) for i in ideals]) assert results == possible_results def multileave(self, method, a, b, ideals, num=100): results = [] for i in range(num): res = method().multileave(a, b) results.append(tuple(res)) results = set(results) possible_results = set([tuple(i) for i in ideals]) assert results == possible_results def evaluate(self, method, ranking, clicks, result): res = method().evaluate(ranking, clicks) assert set(res) == set(result)
import interleaving as il import numpy as np np.random.seed(0) class TestMethods(object): def assert_almost_equal(self, a, b, error_rate=0.1): half_error_rate = error_rate / 2.0 lower_bound = (1.0 - half_error_rate) * a upper_bound = (1.0 + half_error_rate) * a assert lower_bound <= b and b <= upper_bound def interleave(self, method, k, a, b, ideals, num=100): results = [] for i in range(num): res = method().interleave(k, a, b) results.append(tuple(res)) results = set(results) possible_results = set([tuple(i) for i in ideals]) assert results == possible_results def multileave(self, method, a, b, ideals, num=100): results = [] for i in range(num): res = method().multileave(a, b) results.append(tuple(res)) results = set(results) possible_results = set([tuple(i) for i in ideals]) assert results == possible_results def evaluate(self, method, ranking, clicks, result): res = method().evaluate(ranking, clicks) assert set(res) == set(result)
Add a length argument in TestMethods.interleave
Add a length argument in TestMethods.interleave
Python
mit
mpkato/interleaving
import interleaving as il import numpy as np np.random.seed(0) class TestMethods(object): def assert_almost_equal(self, a, b, error_rate=0.1): half_error_rate = error_rate / 2.0 lower_bound = (1.0 - half_error_rate) * a upper_bound = (1.0 + half_error_rate) * a assert lower_bound <= b and b <= upper_bound def interleave(self, method, a, b, ideals, num=100): results = [] for i in range(num): res = method().interleave(a, b) results.append(tuple(res)) results = set(results) possible_results = set([tuple(i) for i in ideals]) assert results == possible_results def multileave(self, method, a, b, ideals, num=100): results = [] for i in range(num): res = method().multileave(a, b) results.append(tuple(res)) results = set(results) possible_results = set([tuple(i) for i in ideals]) assert results == possible_results def evaluate(self, method, ranking, clicks, result): res = method().evaluate(ranking, clicks) assert set(res) == set(result) Add a length argument in TestMethods.interleave
import interleaving as il import numpy as np np.random.seed(0) class TestMethods(object): def assert_almost_equal(self, a, b, error_rate=0.1): half_error_rate = error_rate / 2.0 lower_bound = (1.0 - half_error_rate) * a upper_bound = (1.0 + half_error_rate) * a assert lower_bound <= b and b <= upper_bound def interleave(self, method, k, a, b, ideals, num=100): results = [] for i in range(num): res = method().interleave(k, a, b) results.append(tuple(res)) results = set(results) possible_results = set([tuple(i) for i in ideals]) assert results == possible_results def multileave(self, method, a, b, ideals, num=100): results = [] for i in range(num): res = method().multileave(a, b) results.append(tuple(res)) results = set(results) possible_results = set([tuple(i) for i in ideals]) assert results == possible_results def evaluate(self, method, ranking, clicks, result): res = method().evaluate(ranking, clicks) assert set(res) == set(result)
<commit_before>import interleaving as il import numpy as np np.random.seed(0) class TestMethods(object): def assert_almost_equal(self, a, b, error_rate=0.1): half_error_rate = error_rate / 2.0 lower_bound = (1.0 - half_error_rate) * a upper_bound = (1.0 + half_error_rate) * a assert lower_bound <= b and b <= upper_bound def interleave(self, method, a, b, ideals, num=100): results = [] for i in range(num): res = method().interleave(a, b) results.append(tuple(res)) results = set(results) possible_results = set([tuple(i) for i in ideals]) assert results == possible_results def multileave(self, method, a, b, ideals, num=100): results = [] for i in range(num): res = method().multileave(a, b) results.append(tuple(res)) results = set(results) possible_results = set([tuple(i) for i in ideals]) assert results == possible_results def evaluate(self, method, ranking, clicks, result): res = method().evaluate(ranking, clicks) assert set(res) == set(result) <commit_msg>Add a length argument in TestMethods.interleave<commit_after>
import interleaving as il import numpy as np np.random.seed(0) class TestMethods(object): def assert_almost_equal(self, a, b, error_rate=0.1): half_error_rate = error_rate / 2.0 lower_bound = (1.0 - half_error_rate) * a upper_bound = (1.0 + half_error_rate) * a assert lower_bound <= b and b <= upper_bound def interleave(self, method, k, a, b, ideals, num=100): results = [] for i in range(num): res = method().interleave(k, a, b) results.append(tuple(res)) results = set(results) possible_results = set([tuple(i) for i in ideals]) assert results == possible_results def multileave(self, method, a, b, ideals, num=100): results = [] for i in range(num): res = method().multileave(a, b) results.append(tuple(res)) results = set(results) possible_results = set([tuple(i) for i in ideals]) assert results == possible_results def evaluate(self, method, ranking, clicks, result): res = method().evaluate(ranking, clicks) assert set(res) == set(result)
import interleaving as il import numpy as np np.random.seed(0) class TestMethods(object): def assert_almost_equal(self, a, b, error_rate=0.1): half_error_rate = error_rate / 2.0 lower_bound = (1.0 - half_error_rate) * a upper_bound = (1.0 + half_error_rate) * a assert lower_bound <= b and b <= upper_bound def interleave(self, method, a, b, ideals, num=100): results = [] for i in range(num): res = method().interleave(a, b) results.append(tuple(res)) results = set(results) possible_results = set([tuple(i) for i in ideals]) assert results == possible_results def multileave(self, method, a, b, ideals, num=100): results = [] for i in range(num): res = method().multileave(a, b) results.append(tuple(res)) results = set(results) possible_results = set([tuple(i) for i in ideals]) assert results == possible_results def evaluate(self, method, ranking, clicks, result): res = method().evaluate(ranking, clicks) assert set(res) == set(result) Add a length argument in TestMethods.interleaveimport interleaving as il import numpy as np np.random.seed(0) class TestMethods(object): def assert_almost_equal(self, a, b, error_rate=0.1): half_error_rate = error_rate / 2.0 lower_bound = (1.0 - half_error_rate) * a upper_bound = (1.0 + half_error_rate) * a assert lower_bound <= b and b <= upper_bound def interleave(self, method, k, a, b, ideals, num=100): results = [] for i in range(num): res = method().interleave(k, a, b) results.append(tuple(res)) results = set(results) possible_results = set([tuple(i) for i in ideals]) assert results == possible_results def multileave(self, method, a, b, ideals, num=100): results = [] for i in range(num): res = method().multileave(a, b) results.append(tuple(res)) results = set(results) possible_results = set([tuple(i) for i in ideals]) assert results == possible_results def evaluate(self, method, ranking, clicks, result): res = method().evaluate(ranking, clicks) assert set(res) == set(result)
<commit_before>import interleaving as il import numpy as np np.random.seed(0) class TestMethods(object): def assert_almost_equal(self, a, b, error_rate=0.1): half_error_rate = error_rate / 2.0 lower_bound = (1.0 - half_error_rate) * a upper_bound = (1.0 + half_error_rate) * a assert lower_bound <= b and b <= upper_bound def interleave(self, method, a, b, ideals, num=100): results = [] for i in range(num): res = method().interleave(a, b) results.append(tuple(res)) results = set(results) possible_results = set([tuple(i) for i in ideals]) assert results == possible_results def multileave(self, method, a, b, ideals, num=100): results = [] for i in range(num): res = method().multileave(a, b) results.append(tuple(res)) results = set(results) possible_results = set([tuple(i) for i in ideals]) assert results == possible_results def evaluate(self, method, ranking, clicks, result): res = method().evaluate(ranking, clicks) assert set(res) == set(result) <commit_msg>Add a length argument in TestMethods.interleave<commit_after>import interleaving as il import numpy as np np.random.seed(0) class TestMethods(object): def assert_almost_equal(self, a, b, error_rate=0.1): half_error_rate = error_rate / 2.0 lower_bound = (1.0 - half_error_rate) * a upper_bound = (1.0 + half_error_rate) * a assert lower_bound <= b and b <= upper_bound def interleave(self, method, k, a, b, ideals, num=100): results = [] for i in range(num): res = method().interleave(k, a, b) results.append(tuple(res)) results = set(results) possible_results = set([tuple(i) for i in ideals]) assert results == possible_results def multileave(self, method, a, b, ideals, num=100): results = [] for i in range(num): res = method().multileave(a, b) results.append(tuple(res)) results = set(results) possible_results = set([tuple(i) for i in ideals]) assert results == possible_results def evaluate(self, method, ranking, clicks, result): res = method().evaluate(ranking, clicks) assert set(res) == set(result)
94529f62757886d2291cf90596a179dc2d0b6642
yutu.py
yutu.py
import discord from discord.ext.commands import Bot import json client = Bot("~", game=discord.Game(name="~help")) @client.event async def on_ready(): print('We have logged in as {0.user}'.format(client)) @client.command() async def highfive(ctx): ''' Give Yutu a high-five ''' await ctx.send('{0.mention} :pray: {1.mention}'.format(ctx.me, ctx.author)) @client.command() async def cute(ctx, member: discord.Member = None): if member is None: first = ctx.me second = ctx.author else: first = ctx.author second = member post = discord.Embed(description='**{0.name}** thinks that **{1.name}** is cute!'.format(first, second)) post.set_image(url="https://i.imgur.com/MuVAkV2.gif") await ctx.send(embed=post) if __name__ == "__main__": with open("cfg.json") as fh: token = json.load(fh)['token'] client.run(token)
import discord from discord.ext.commands import Bot import json client = Bot("~", game=discord.Game(name="~help")) @client.event async def on_ready(): print('We have logged in as {0.user}'.format(client)) @client.command() async def highfive(ctx): ''' Give Yutu a high-five ''' await ctx.send('{0.mention} :pray: {1.mention}'.format(ctx.me, ctx.author)) @client.command() async def cute(ctx, member: discord.Member = None): if member is None: first = ctx.me second = ctx.author else: first = ctx.author second = member post = discord.Embed(description='**{0.display_name}** thinks that **{1.display_name}** is cute!'.format(first, second)) post.set_image(url="https://i.imgur.com/MuVAkV2.gif") await ctx.send(embed=post) if __name__ == "__main__": with open("cfg.json") as fh: token = json.load(fh)['token'] client.run(token)
Make cute respect guild nicknames
Make cute respect guild nicknames
Python
mit
HarkonenBade/yutu
import discord from discord.ext.commands import Bot import json client = Bot("~", game=discord.Game(name="~help")) @client.event async def on_ready(): print('We have logged in as {0.user}'.format(client)) @client.command() async def highfive(ctx): ''' Give Yutu a high-five ''' await ctx.send('{0.mention} :pray: {1.mention}'.format(ctx.me, ctx.author)) @client.command() async def cute(ctx, member: discord.Member = None): if member is None: first = ctx.me second = ctx.author else: first = ctx.author second = member post = discord.Embed(description='**{0.name}** thinks that **{1.name}** is cute!'.format(first, second)) post.set_image(url="https://i.imgur.com/MuVAkV2.gif") await ctx.send(embed=post) if __name__ == "__main__": with open("cfg.json") as fh: token = json.load(fh)['token'] client.run(token) Make cute respect guild nicknames
import discord from discord.ext.commands import Bot import json client = Bot("~", game=discord.Game(name="~help")) @client.event async def on_ready(): print('We have logged in as {0.user}'.format(client)) @client.command() async def highfive(ctx): ''' Give Yutu a high-five ''' await ctx.send('{0.mention} :pray: {1.mention}'.format(ctx.me, ctx.author)) @client.command() async def cute(ctx, member: discord.Member = None): if member is None: first = ctx.me second = ctx.author else: first = ctx.author second = member post = discord.Embed(description='**{0.display_name}** thinks that **{1.display_name}** is cute!'.format(first, second)) post.set_image(url="https://i.imgur.com/MuVAkV2.gif") await ctx.send(embed=post) if __name__ == "__main__": with open("cfg.json") as fh: token = json.load(fh)['token'] client.run(token)
<commit_before>import discord from discord.ext.commands import Bot import json client = Bot("~", game=discord.Game(name="~help")) @client.event async def on_ready(): print('We have logged in as {0.user}'.format(client)) @client.command() async def highfive(ctx): ''' Give Yutu a high-five ''' await ctx.send('{0.mention} :pray: {1.mention}'.format(ctx.me, ctx.author)) @client.command() async def cute(ctx, member: discord.Member = None): if member is None: first = ctx.me second = ctx.author else: first = ctx.author second = member post = discord.Embed(description='**{0.name}** thinks that **{1.name}** is cute!'.format(first, second)) post.set_image(url="https://i.imgur.com/MuVAkV2.gif") await ctx.send(embed=post) if __name__ == "__main__": with open("cfg.json") as fh: token = json.load(fh)['token'] client.run(token) <commit_msg>Make cute respect guild nicknames<commit_after>
import discord from discord.ext.commands import Bot import json client = Bot("~", game=discord.Game(name="~help")) @client.event async def on_ready(): print('We have logged in as {0.user}'.format(client)) @client.command() async def highfive(ctx): ''' Give Yutu a high-five ''' await ctx.send('{0.mention} :pray: {1.mention}'.format(ctx.me, ctx.author)) @client.command() async def cute(ctx, member: discord.Member = None): if member is None: first = ctx.me second = ctx.author else: first = ctx.author second = member post = discord.Embed(description='**{0.display_name}** thinks that **{1.display_name}** is cute!'.format(first, second)) post.set_image(url="https://i.imgur.com/MuVAkV2.gif") await ctx.send(embed=post) if __name__ == "__main__": with open("cfg.json") as fh: token = json.load(fh)['token'] client.run(token)
import discord from discord.ext.commands import Bot import json client = Bot("~", game=discord.Game(name="~help")) @client.event async def on_ready(): print('We have logged in as {0.user}'.format(client)) @client.command() async def highfive(ctx): ''' Give Yutu a high-five ''' await ctx.send('{0.mention} :pray: {1.mention}'.format(ctx.me, ctx.author)) @client.command() async def cute(ctx, member: discord.Member = None): if member is None: first = ctx.me second = ctx.author else: first = ctx.author second = member post = discord.Embed(description='**{0.name}** thinks that **{1.name}** is cute!'.format(first, second)) post.set_image(url="https://i.imgur.com/MuVAkV2.gif") await ctx.send(embed=post) if __name__ == "__main__": with open("cfg.json") as fh: token = json.load(fh)['token'] client.run(token) Make cute respect guild nicknamesimport discord from discord.ext.commands import Bot import json client = Bot("~", game=discord.Game(name="~help")) @client.event async def on_ready(): print('We have logged in as {0.user}'.format(client)) @client.command() async def highfive(ctx): ''' Give Yutu a high-five ''' await ctx.send('{0.mention} :pray: {1.mention}'.format(ctx.me, ctx.author)) @client.command() async def cute(ctx, member: discord.Member = None): if member is None: first = ctx.me second = ctx.author else: first = ctx.author second = member post = discord.Embed(description='**{0.display_name}** thinks that **{1.display_name}** is cute!'.format(first, second)) post.set_image(url="https://i.imgur.com/MuVAkV2.gif") await ctx.send(embed=post) if __name__ == "__main__": with open("cfg.json") as fh: token = json.load(fh)['token'] client.run(token)
<commit_before>import discord from discord.ext.commands import Bot import json client = Bot("~", game=discord.Game(name="~help")) @client.event async def on_ready(): print('We have logged in as {0.user}'.format(client)) @client.command() async def highfive(ctx): ''' Give Yutu a high-five ''' await ctx.send('{0.mention} :pray: {1.mention}'.format(ctx.me, ctx.author)) @client.command() async def cute(ctx, member: discord.Member = None): if member is None: first = ctx.me second = ctx.author else: first = ctx.author second = member post = discord.Embed(description='**{0.name}** thinks that **{1.name}** is cute!'.format(first, second)) post.set_image(url="https://i.imgur.com/MuVAkV2.gif") await ctx.send(embed=post) if __name__ == "__main__": with open("cfg.json") as fh: token = json.load(fh)['token'] client.run(token) <commit_msg>Make cute respect guild nicknames<commit_after>import discord from discord.ext.commands import Bot import json client = Bot("~", game=discord.Game(name="~help")) @client.event async def on_ready(): print('We have logged in as {0.user}'.format(client)) @client.command() async def highfive(ctx): ''' Give Yutu a high-five ''' await ctx.send('{0.mention} :pray: {1.mention}'.format(ctx.me, ctx.author)) @client.command() async def cute(ctx, member: discord.Member = None): if member is None: first = ctx.me second = ctx.author else: first = ctx.author second = member post = discord.Embed(description='**{0.display_name}** thinks that **{1.display_name}** is cute!'.format(first, second)) post.set_image(url="https://i.imgur.com/MuVAkV2.gif") await ctx.send(embed=post) if __name__ == "__main__": with open("cfg.json") as fh: token = json.load(fh)['token'] client.run(token)
f8a42dc715c55d1f3faf3dd87d168e0687f87e5b
l10n_ch_payment_slip/report/__init__.py
l10n_ch_payment_slip/report/__init__.py
# -*- coding: utf-8 -*- ############################################################################## # # Copyright (c) 2010 Camptocamp SA (http://www.camptocamp.com) # All Right Reserved # # Author : Nicolas Bessi (Camptocamp) # # WARNING: This program as such is intended to be used by professional # programmers who take the whole responsability of assessing all potential # consequences resulting from its eventual inadequacies and bugs # End users who are looking for a ready-to-use solution with commercial # garantees and support are strongly adviced to contract a Free Software # Service Company # # This program is Free Software; you can redistribute it and/or # modify it under the terms of the GNU General Public License # as published by the Free Software Foundation; either version 2 # of the License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program; if not, write to the Free Software # Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. # ############################################################################## from . import payment_slip_from_invoice
# -*- coding: utf-8 -*- ############################################################################## # # Copyright (c) 2010 Camptocamp SA (http://www.camptocamp.com) # All Right Reserved # # Author : Nicolas Bessi (Camptocamp) # # WARNING: This program as such is intended to be used by professional # programmers who take the whole responsability of assessing all potential # consequences resulting from its eventual inadequacies and bugs # End users who are looking for a ready-to-use solution with commercial # garantees and support are strongly adviced to contract a Free Software # Service Company # # This program is Free Software; you can redistribute it and/or # modify it under the terms of the GNU General Public License # as published by the Free Software Foundation; either version 2 # of the License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program; if not, write to the Free Software # Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. # ############################################################################## from . import payment_slip_from_invoice from . import reports_common
Add common in import statement
Add common in import statement
Python
agpl-3.0
BT-ojossen/l10n-switzerland,cgaspoz/l10n-switzerland,CompassionCH/l10n-switzerland,BT-csanchez/l10n-switzerland,CompassionCH/l10n-switzerland,BT-fgarbely/l10n-switzerland,BT-ojossen/l10n-switzerland,BT-fgarbely/l10n-switzerland,cyp-opennet/ons_cyp_github,open-net-sarl/l10n-switzerland,open-net-sarl/l10n-switzerland,cyp-opennet/ons_cyp_github,BT-aestebanez/l10n-switzerland,michl/l10n-switzerland,ndtran/l10n-switzerland
# -*- coding: utf-8 -*- ############################################################################## # # Copyright (c) 2010 Camptocamp SA (http://www.camptocamp.com) # All Right Reserved # # Author : Nicolas Bessi (Camptocamp) # # WARNING: This program as such is intended to be used by professional # programmers who take the whole responsability of assessing all potential # consequences resulting from its eventual inadequacies and bugs # End users who are looking for a ready-to-use solution with commercial # garantees and support are strongly adviced to contract a Free Software # Service Company # # This program is Free Software; you can redistribute it and/or # modify it under the terms of the GNU General Public License # as published by the Free Software Foundation; either version 2 # of the License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program; if not, write to the Free Software # Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. # ############################################################################## from . import payment_slip_from_invoice Add common in import statement
# -*- coding: utf-8 -*- ############################################################################## # # Copyright (c) 2010 Camptocamp SA (http://www.camptocamp.com) # All Right Reserved # # Author : Nicolas Bessi (Camptocamp) # # WARNING: This program as such is intended to be used by professional # programmers who take the whole responsability of assessing all potential # consequences resulting from its eventual inadequacies and bugs # End users who are looking for a ready-to-use solution with commercial # garantees and support are strongly adviced to contract a Free Software # Service Company # # This program is Free Software; you can redistribute it and/or # modify it under the terms of the GNU General Public License # as published by the Free Software Foundation; either version 2 # of the License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program; if not, write to the Free Software # Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. # ############################################################################## from . import payment_slip_from_invoice from . import reports_common
<commit_before># -*- coding: utf-8 -*- ############################################################################## # # Copyright (c) 2010 Camptocamp SA (http://www.camptocamp.com) # All Right Reserved # # Author : Nicolas Bessi (Camptocamp) # # WARNING: This program as such is intended to be used by professional # programmers who take the whole responsability of assessing all potential # consequences resulting from its eventual inadequacies and bugs # End users who are looking for a ready-to-use solution with commercial # garantees and support are strongly adviced to contract a Free Software # Service Company # # This program is Free Software; you can redistribute it and/or # modify it under the terms of the GNU General Public License # as published by the Free Software Foundation; either version 2 # of the License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program; if not, write to the Free Software # Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. # ############################################################################## from . import payment_slip_from_invoice <commit_msg>Add common in import statement<commit_after>
# -*- coding: utf-8 -*- ############################################################################## # # Copyright (c) 2010 Camptocamp SA (http://www.camptocamp.com) # All Right Reserved # # Author : Nicolas Bessi (Camptocamp) # # WARNING: This program as such is intended to be used by professional # programmers who take the whole responsability of assessing all potential # consequences resulting from its eventual inadequacies and bugs # End users who are looking for a ready-to-use solution with commercial # garantees and support are strongly adviced to contract a Free Software # Service Company # # This program is Free Software; you can redistribute it and/or # modify it under the terms of the GNU General Public License # as published by the Free Software Foundation; either version 2 # of the License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program; if not, write to the Free Software # Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. # ############################################################################## from . import payment_slip_from_invoice from . import reports_common
# -*- coding: utf-8 -*- ############################################################################## # # Copyright (c) 2010 Camptocamp SA (http://www.camptocamp.com) # All Right Reserved # # Author : Nicolas Bessi (Camptocamp) # # WARNING: This program as such is intended to be used by professional # programmers who take the whole responsability of assessing all potential # consequences resulting from its eventual inadequacies and bugs # End users who are looking for a ready-to-use solution with commercial # garantees and support are strongly adviced to contract a Free Software # Service Company # # This program is Free Software; you can redistribute it and/or # modify it under the terms of the GNU General Public License # as published by the Free Software Foundation; either version 2 # of the License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program; if not, write to the Free Software # Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. # ############################################################################## from . import payment_slip_from_invoice Add common in import statement# -*- coding: utf-8 -*- ############################################################################## # # Copyright (c) 2010 Camptocamp SA (http://www.camptocamp.com) # All Right Reserved # # Author : Nicolas Bessi (Camptocamp) # # WARNING: This program as such is intended to be used by professional # programmers who take the whole responsability of assessing all potential # consequences resulting from its eventual inadequacies and bugs # End users who are looking for a ready-to-use solution with commercial # garantees and support are strongly adviced to contract a Free Software # Service Company # # This program is Free Software; you can redistribute it and/or # modify it under the terms of the GNU General Public License # as published by the Free Software Foundation; either version 2 # of the License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program; if not, write to the Free Software # Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. # ############################################################################## from . import payment_slip_from_invoice from . import reports_common
<commit_before># -*- coding: utf-8 -*- ############################################################################## # # Copyright (c) 2010 Camptocamp SA (http://www.camptocamp.com) # All Right Reserved # # Author : Nicolas Bessi (Camptocamp) # # WARNING: This program as such is intended to be used by professional # programmers who take the whole responsability of assessing all potential # consequences resulting from its eventual inadequacies and bugs # End users who are looking for a ready-to-use solution with commercial # garantees and support are strongly adviced to contract a Free Software # Service Company # # This program is Free Software; you can redistribute it and/or # modify it under the terms of the GNU General Public License # as published by the Free Software Foundation; either version 2 # of the License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program; if not, write to the Free Software # Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. # ############################################################################## from . import payment_slip_from_invoice <commit_msg>Add common in import statement<commit_after># -*- coding: utf-8 -*- ############################################################################## # # Copyright (c) 2010 Camptocamp SA (http://www.camptocamp.com) # All Right Reserved # # Author : Nicolas Bessi (Camptocamp) # # WARNING: This program as such is intended to be used by professional # programmers who take the whole responsability of assessing all potential # consequences resulting from its eventual inadequacies and bugs # End users who are looking for a ready-to-use solution with commercial # garantees and support are strongly adviced to contract a Free Software # Service Company # # This program is Free Software; you can redistribute it and/or # modify it under the terms of the GNU General Public License # as published by the Free Software Foundation; either version 2 # of the License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program; if not, write to the Free Software # Foundation, Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA. # ############################################################################## from . import payment_slip_from_invoice from . import reports_common
694a3e1aa5eec02e81cfde517ca126f00b21bd2f
rest_framework_docs/api_docs.py
rest_framework_docs/api_docs.py
from django.conf import settings from django.core.urlresolvers import RegexURLResolver, RegexURLPattern from rest_framework.views import APIView from rest_framework_docs.api_endpoint import ApiEndpoint class ApiDocumentation(object): def __init__(self): self.endpoints = [] root_urlconf = __import__(settings.ROOT_URLCONF) if hasattr(root_urlconf, 'urls'): self.get_all_view_names(root_urlconf.urls.urlpatterns) else: self.get_all_view_names(root_urlconf.urlpatterns) def get_all_view_names(self, urlpatterns, parent_pattern=None): for pattern in urlpatterns: if isinstance(pattern, RegexURLResolver): parent_pattern = None if pattern._regex == "^" else pattern self.get_all_view_names(urlpatterns=pattern.url_patterns, parent_pattern=parent_pattern) elif isinstance(pattern, RegexURLPattern) and self._is_drf_view(pattern): api_endpoint = ApiEndpoint(pattern, parent_pattern) self.endpoints.append(api_endpoint) def _is_drf_view(self, pattern): # Should check whether a pattern inherits from DRF's APIView return hasattr(pattern.callback, 'cls') and issubclass(pattern.callback.cls, APIView) def get_endpoints(self): return self.endpoints
from django.conf import settings from django.core.urlresolvers import RegexURLResolver, RegexURLPattern from django.utils.module_loading import import_string from rest_framework.views import APIView from rest_framework_docs.api_endpoint import ApiEndpoint class ApiDocumentation(object): def __init__(self): self.endpoints = [] root_urlconf = import_string(settings.ROOT_URLCONF) if hasattr(root_urlconf, 'urls'): self.get_all_view_names(root_urlconf.urls.urlpatterns) else: self.get_all_view_names(root_urlconf.urlpatterns) def get_all_view_names(self, urlpatterns, parent_pattern=None): for pattern in urlpatterns: if isinstance(pattern, RegexURLResolver): parent_pattern = None if pattern._regex == "^" else pattern self.get_all_view_names(urlpatterns=pattern.url_patterns, parent_pattern=parent_pattern) elif isinstance(pattern, RegexURLPattern) and self._is_drf_view(pattern): api_endpoint = ApiEndpoint(pattern, parent_pattern) self.endpoints.append(api_endpoint) def _is_drf_view(self, pattern): # Should check whether a pattern inherits from DRF's APIView return hasattr(pattern.callback, 'cls') and issubclass(pattern.callback.cls, APIView) def get_endpoints(self): return self.endpoints
Use Django's module loading rather than __import__
Use Django's module loading rather than __import__ __import__ doesn't deal well with dotted paths, in my instance my root url conf is in a few levels "appname.config.urls". unfortunately, for __import__ this means that just `appname` is imported, but `config.urls` is loaded and no other modules in between are usable. Switching to Django's `import_string` method allows importing the final module in the path string, in my case the equivalent would be `from appname.config import urls`.
Python
bsd-2-clause
manosim/django-rest-framework-docs,ekonstantinidis/django-rest-framework-docs,ekonstantinidis/django-rest-framework-docs,ekonstantinidis/django-rest-framework-docs,manosim/django-rest-framework-docs,manosim/django-rest-framework-docs
from django.conf import settings from django.core.urlresolvers import RegexURLResolver, RegexURLPattern from rest_framework.views import APIView from rest_framework_docs.api_endpoint import ApiEndpoint class ApiDocumentation(object): def __init__(self): self.endpoints = [] root_urlconf = __import__(settings.ROOT_URLCONF) if hasattr(root_urlconf, 'urls'): self.get_all_view_names(root_urlconf.urls.urlpatterns) else: self.get_all_view_names(root_urlconf.urlpatterns) def get_all_view_names(self, urlpatterns, parent_pattern=None): for pattern in urlpatterns: if isinstance(pattern, RegexURLResolver): parent_pattern = None if pattern._regex == "^" else pattern self.get_all_view_names(urlpatterns=pattern.url_patterns, parent_pattern=parent_pattern) elif isinstance(pattern, RegexURLPattern) and self._is_drf_view(pattern): api_endpoint = ApiEndpoint(pattern, parent_pattern) self.endpoints.append(api_endpoint) def _is_drf_view(self, pattern): # Should check whether a pattern inherits from DRF's APIView return hasattr(pattern.callback, 'cls') and issubclass(pattern.callback.cls, APIView) def get_endpoints(self): return self.endpoints Use Django's module loading rather than __import__ __import__ doesn't deal well with dotted paths, in my instance my root url conf is in a few levels "appname.config.urls". unfortunately, for __import__ this means that just `appname` is imported, but `config.urls` is loaded and no other modules in between are usable. Switching to Django's `import_string` method allows importing the final module in the path string, in my case the equivalent would be `from appname.config import urls`.
from django.conf import settings from django.core.urlresolvers import RegexURLResolver, RegexURLPattern from django.utils.module_loading import import_string from rest_framework.views import APIView from rest_framework_docs.api_endpoint import ApiEndpoint class ApiDocumentation(object): def __init__(self): self.endpoints = [] root_urlconf = import_string(settings.ROOT_URLCONF) if hasattr(root_urlconf, 'urls'): self.get_all_view_names(root_urlconf.urls.urlpatterns) else: self.get_all_view_names(root_urlconf.urlpatterns) def get_all_view_names(self, urlpatterns, parent_pattern=None): for pattern in urlpatterns: if isinstance(pattern, RegexURLResolver): parent_pattern = None if pattern._regex == "^" else pattern self.get_all_view_names(urlpatterns=pattern.url_patterns, parent_pattern=parent_pattern) elif isinstance(pattern, RegexURLPattern) and self._is_drf_view(pattern): api_endpoint = ApiEndpoint(pattern, parent_pattern) self.endpoints.append(api_endpoint) def _is_drf_view(self, pattern): # Should check whether a pattern inherits from DRF's APIView return hasattr(pattern.callback, 'cls') and issubclass(pattern.callback.cls, APIView) def get_endpoints(self): return self.endpoints
<commit_before>from django.conf import settings from django.core.urlresolvers import RegexURLResolver, RegexURLPattern from rest_framework.views import APIView from rest_framework_docs.api_endpoint import ApiEndpoint class ApiDocumentation(object): def __init__(self): self.endpoints = [] root_urlconf = __import__(settings.ROOT_URLCONF) if hasattr(root_urlconf, 'urls'): self.get_all_view_names(root_urlconf.urls.urlpatterns) else: self.get_all_view_names(root_urlconf.urlpatterns) def get_all_view_names(self, urlpatterns, parent_pattern=None): for pattern in urlpatterns: if isinstance(pattern, RegexURLResolver): parent_pattern = None if pattern._regex == "^" else pattern self.get_all_view_names(urlpatterns=pattern.url_patterns, parent_pattern=parent_pattern) elif isinstance(pattern, RegexURLPattern) and self._is_drf_view(pattern): api_endpoint = ApiEndpoint(pattern, parent_pattern) self.endpoints.append(api_endpoint) def _is_drf_view(self, pattern): # Should check whether a pattern inherits from DRF's APIView return hasattr(pattern.callback, 'cls') and issubclass(pattern.callback.cls, APIView) def get_endpoints(self): return self.endpoints <commit_msg>Use Django's module loading rather than __import__ __import__ doesn't deal well with dotted paths, in my instance my root url conf is in a few levels "appname.config.urls". unfortunately, for __import__ this means that just `appname` is imported, but `config.urls` is loaded and no other modules in between are usable. Switching to Django's `import_string` method allows importing the final module in the path string, in my case the equivalent would be `from appname.config import urls`.<commit_after>
from django.conf import settings from django.core.urlresolvers import RegexURLResolver, RegexURLPattern from django.utils.module_loading import import_string from rest_framework.views import APIView from rest_framework_docs.api_endpoint import ApiEndpoint class ApiDocumentation(object): def __init__(self): self.endpoints = [] root_urlconf = import_string(settings.ROOT_URLCONF) if hasattr(root_urlconf, 'urls'): self.get_all_view_names(root_urlconf.urls.urlpatterns) else: self.get_all_view_names(root_urlconf.urlpatterns) def get_all_view_names(self, urlpatterns, parent_pattern=None): for pattern in urlpatterns: if isinstance(pattern, RegexURLResolver): parent_pattern = None if pattern._regex == "^" else pattern self.get_all_view_names(urlpatterns=pattern.url_patterns, parent_pattern=parent_pattern) elif isinstance(pattern, RegexURLPattern) and self._is_drf_view(pattern): api_endpoint = ApiEndpoint(pattern, parent_pattern) self.endpoints.append(api_endpoint) def _is_drf_view(self, pattern): # Should check whether a pattern inherits from DRF's APIView return hasattr(pattern.callback, 'cls') and issubclass(pattern.callback.cls, APIView) def get_endpoints(self): return self.endpoints
from django.conf import settings from django.core.urlresolvers import RegexURLResolver, RegexURLPattern from rest_framework.views import APIView from rest_framework_docs.api_endpoint import ApiEndpoint class ApiDocumentation(object): def __init__(self): self.endpoints = [] root_urlconf = __import__(settings.ROOT_URLCONF) if hasattr(root_urlconf, 'urls'): self.get_all_view_names(root_urlconf.urls.urlpatterns) else: self.get_all_view_names(root_urlconf.urlpatterns) def get_all_view_names(self, urlpatterns, parent_pattern=None): for pattern in urlpatterns: if isinstance(pattern, RegexURLResolver): parent_pattern = None if pattern._regex == "^" else pattern self.get_all_view_names(urlpatterns=pattern.url_patterns, parent_pattern=parent_pattern) elif isinstance(pattern, RegexURLPattern) and self._is_drf_view(pattern): api_endpoint = ApiEndpoint(pattern, parent_pattern) self.endpoints.append(api_endpoint) def _is_drf_view(self, pattern): # Should check whether a pattern inherits from DRF's APIView return hasattr(pattern.callback, 'cls') and issubclass(pattern.callback.cls, APIView) def get_endpoints(self): return self.endpoints Use Django's module loading rather than __import__ __import__ doesn't deal well with dotted paths, in my instance my root url conf is in a few levels "appname.config.urls". unfortunately, for __import__ this means that just `appname` is imported, but `config.urls` is loaded and no other modules in between are usable. Switching to Django's `import_string` method allows importing the final module in the path string, in my case the equivalent would be `from appname.config import urls`.from django.conf import settings from django.core.urlresolvers import RegexURLResolver, RegexURLPattern from django.utils.module_loading import import_string from rest_framework.views import APIView from rest_framework_docs.api_endpoint import ApiEndpoint class ApiDocumentation(object): def __init__(self): self.endpoints = [] root_urlconf = import_string(settings.ROOT_URLCONF) if hasattr(root_urlconf, 'urls'): self.get_all_view_names(root_urlconf.urls.urlpatterns) else: self.get_all_view_names(root_urlconf.urlpatterns) def get_all_view_names(self, urlpatterns, parent_pattern=None): for pattern in urlpatterns: if isinstance(pattern, RegexURLResolver): parent_pattern = None if pattern._regex == "^" else pattern self.get_all_view_names(urlpatterns=pattern.url_patterns, parent_pattern=parent_pattern) elif isinstance(pattern, RegexURLPattern) and self._is_drf_view(pattern): api_endpoint = ApiEndpoint(pattern, parent_pattern) self.endpoints.append(api_endpoint) def _is_drf_view(self, pattern): # Should check whether a pattern inherits from DRF's APIView return hasattr(pattern.callback, 'cls') and issubclass(pattern.callback.cls, APIView) def get_endpoints(self): return self.endpoints
<commit_before>from django.conf import settings from django.core.urlresolvers import RegexURLResolver, RegexURLPattern from rest_framework.views import APIView from rest_framework_docs.api_endpoint import ApiEndpoint class ApiDocumentation(object): def __init__(self): self.endpoints = [] root_urlconf = __import__(settings.ROOT_URLCONF) if hasattr(root_urlconf, 'urls'): self.get_all_view_names(root_urlconf.urls.urlpatterns) else: self.get_all_view_names(root_urlconf.urlpatterns) def get_all_view_names(self, urlpatterns, parent_pattern=None): for pattern in urlpatterns: if isinstance(pattern, RegexURLResolver): parent_pattern = None if pattern._regex == "^" else pattern self.get_all_view_names(urlpatterns=pattern.url_patterns, parent_pattern=parent_pattern) elif isinstance(pattern, RegexURLPattern) and self._is_drf_view(pattern): api_endpoint = ApiEndpoint(pattern, parent_pattern) self.endpoints.append(api_endpoint) def _is_drf_view(self, pattern): # Should check whether a pattern inherits from DRF's APIView return hasattr(pattern.callback, 'cls') and issubclass(pattern.callback.cls, APIView) def get_endpoints(self): return self.endpoints <commit_msg>Use Django's module loading rather than __import__ __import__ doesn't deal well with dotted paths, in my instance my root url conf is in a few levels "appname.config.urls". unfortunately, for __import__ this means that just `appname` is imported, but `config.urls` is loaded and no other modules in between are usable. Switching to Django's `import_string` method allows importing the final module in the path string, in my case the equivalent would be `from appname.config import urls`.<commit_after>from django.conf import settings from django.core.urlresolvers import RegexURLResolver, RegexURLPattern from django.utils.module_loading import import_string from rest_framework.views import APIView from rest_framework_docs.api_endpoint import ApiEndpoint class ApiDocumentation(object): def __init__(self): self.endpoints = [] root_urlconf = import_string(settings.ROOT_URLCONF) if hasattr(root_urlconf, 'urls'): self.get_all_view_names(root_urlconf.urls.urlpatterns) else: self.get_all_view_names(root_urlconf.urlpatterns) def get_all_view_names(self, urlpatterns, parent_pattern=None): for pattern in urlpatterns: if isinstance(pattern, RegexURLResolver): parent_pattern = None if pattern._regex == "^" else pattern self.get_all_view_names(urlpatterns=pattern.url_patterns, parent_pattern=parent_pattern) elif isinstance(pattern, RegexURLPattern) and self._is_drf_view(pattern): api_endpoint = ApiEndpoint(pattern, parent_pattern) self.endpoints.append(api_endpoint) def _is_drf_view(self, pattern): # Should check whether a pattern inherits from DRF's APIView return hasattr(pattern.callback, 'cls') and issubclass(pattern.callback.cls, APIView) def get_endpoints(self): return self.endpoints
6ece85c530d9856a7db650729e57a9f2a92dfa4b
oscar/apps/offer/managers.py
oscar/apps/offer/managers.py
import datetime from django.db import models class ActiveOfferManager(models.Manager): """ For searching/creating ACTIVE offers only. """ def get_query_set(self): today = datetime.date.today() return super(ActiveOfferManager, self).get_query_set().filter( start_date__lte=today, end_date__gt=today)
import datetime from django.db import models class ActiveOfferManager(models.Manager): """ For searching/creating offers within their date range """ def get_query_set(self): today = datetime.date.today() return super(ActiveOfferManager, self).get_query_set().filter( start_date__lte=today, end_date__gte=today)
Fix queryset filtering for active offers
Fix queryset filtering for active offers
Python
bsd-3-clause
sasha0/django-oscar,thechampanurag/django-oscar,ka7eh/django-oscar,elliotthill/django-oscar,eddiep1101/django-oscar,john-parton/django-oscar,adamend/django-oscar,nfletton/django-oscar,jmt4/django-oscar,pdonadeo/django-oscar,bschuon/django-oscar,sonofatailor/django-oscar,QLGu/django-oscar,monikasulik/django-oscar,WillisXChen/django-oscar,josesanch/django-oscar,ka7eh/django-oscar,WillisXChen/django-oscar,WadeYuChen/django-oscar,taedori81/django-oscar,manevant/django-oscar,sonofatailor/django-oscar,binarydud/django-oscar,vovanbo/django-oscar,solarissmoke/django-oscar,adamend/django-oscar,QLGu/django-oscar,Jannes123/django-oscar,elliotthill/django-oscar,WadeYuChen/django-oscar,bnprk/django-oscar,rocopartners/django-oscar,jlmadurga/django-oscar,eddiep1101/django-oscar,bnprk/django-oscar,Bogh/django-oscar,lijoantony/django-oscar,django-oscar/django-oscar,solarissmoke/django-oscar,rocopartners/django-oscar,QLGu/django-oscar,spartonia/django-oscar,jinnykoo/wuyisj,Idematica/django-oscar,anentropic/django-oscar,saadatqadri/django-oscar,machtfit/django-oscar,itbabu/django-oscar,faratro/django-oscar,WadeYuChen/django-oscar,josesanch/django-oscar,jmt4/django-oscar,dongguangming/django-oscar,michaelkuty/django-oscar,sonofatailor/django-oscar,DrOctogon/unwash_ecom,mexeniz/django-oscar,jinnykoo/wuyisj,taedori81/django-oscar,josesanch/django-oscar,nfletton/django-oscar,DrOctogon/unwash_ecom,lijoantony/django-oscar,kapari/django-oscar,ka7eh/django-oscar,Bogh/django-oscar,MatthewWilkes/django-oscar,nfletton/django-oscar,ka7eh/django-oscar,django-oscar/django-oscar,DrOctogon/unwash_ecom,faratro/django-oscar,ademuk/django-oscar,ademuk/django-oscar,sasha0/django-oscar,makielab/django-oscar,michaelkuty/django-oscar,faratro/django-oscar,amirrpp/django-oscar,anentropic/django-oscar,vovanbo/django-oscar,nfletton/django-oscar,anentropic/django-oscar,michaelkuty/django-oscar,Bogh/django-oscar,john-parton/django-oscar,pasqualguerrero/django-oscar,john-parton/django-oscar,bschuon/django-oscar,bnprk/django-oscar,vovanbo/django-oscar,binarydud/django-oscar,mexeniz/django-oscar,thechampanurag/django-oscar,makielab/django-oscar,kapari/django-oscar,WillisXChen/django-oscar,nickpack/django-oscar,eddiep1101/django-oscar,mexeniz/django-oscar,binarydud/django-oscar,pasqualguerrero/django-oscar,marcoantoniooliveira/labweb,kapari/django-oscar,lijoantony/django-oscar,amirrpp/django-oscar,binarydud/django-oscar,okfish/django-oscar,sasha0/django-oscar,monikasulik/django-oscar,Jannes123/django-oscar,manevant/django-oscar,django-oscar/django-oscar,WillisXChen/django-oscar,WillisXChen/django-oscar,manevant/django-oscar,rocopartners/django-oscar,dongguangming/django-oscar,jmt4/django-oscar,MatthewWilkes/django-oscar,taedori81/django-oscar,jmt4/django-oscar,machtfit/django-oscar,lijoantony/django-oscar,jlmadurga/django-oscar,saadatqadri/django-oscar,solarissmoke/django-oscar,saadatqadri/django-oscar,ademuk/django-oscar,bnprk/django-oscar,saadatqadri/django-oscar,marcoantoniooliveira/labweb,django-oscar/django-oscar,rocopartners/django-oscar,manevant/django-oscar,john-parton/django-oscar,thechampanurag/django-oscar,jinnykoo/wuyisj.com,jlmadurga/django-oscar,itbabu/django-oscar,Bogh/django-oscar,mexeniz/django-oscar,jinnykoo/wuyisj.com,adamend/django-oscar,ahmetdaglarbas/e-commerce,MatthewWilkes/django-oscar,jlmadurga/django-oscar,Jannes123/django-oscar,faratro/django-oscar,spartonia/django-oscar,spartonia/django-oscar,itbabu/django-oscar,monikasulik/django-oscar,WillisXChen/django-oscar,MatthewWilkes/django-oscar,kapari/django-oscar,marcoantoniooliveira/labweb,sasha0/django-oscar,anentropic/django-oscar,kapt/django-oscar,vovanbo/django-oscar,Idematica/django-oscar,makielab/django-oscar,jinnykoo/christmas,ademuk/django-oscar,solarissmoke/django-oscar,kapt/django-oscar,ahmetdaglarbas/e-commerce,sonofatailor/django-oscar,jinnykoo/wuyisj,monikasulik/django-oscar,pdonadeo/django-oscar,okfish/django-oscar,amirrpp/django-oscar,bschuon/django-oscar,marcoantoniooliveira/labweb,jinnykoo/christmas,ahmetdaglarbas/e-commerce,machtfit/django-oscar,jinnykoo/wuyisj,nickpack/django-oscar,ahmetdaglarbas/e-commerce,Jannes123/django-oscar,itbabu/django-oscar,nickpack/django-oscar,jinnykoo/christmas,Idematica/django-oscar,pasqualguerrero/django-oscar,WadeYuChen/django-oscar,pdonadeo/django-oscar,QLGu/django-oscar,taedori81/django-oscar,jinnykoo/wuyisj.com,pdonadeo/django-oscar,adamend/django-oscar,makielab/django-oscar,okfish/django-oscar,pasqualguerrero/django-oscar,thechampanurag/django-oscar,michaelkuty/django-oscar,elliotthill/django-oscar,dongguangming/django-oscar,kapt/django-oscar,jinnykoo/wuyisj.com,dongguangming/django-oscar,okfish/django-oscar,eddiep1101/django-oscar,nickpack/django-oscar,bschuon/django-oscar,spartonia/django-oscar,amirrpp/django-oscar
import datetime from django.db import models class ActiveOfferManager(models.Manager): """ For searching/creating ACTIVE offers only. """ def get_query_set(self): today = datetime.date.today() return super(ActiveOfferManager, self).get_query_set().filter( start_date__lte=today, end_date__gt=today) Fix queryset filtering for active offers
import datetime from django.db import models class ActiveOfferManager(models.Manager): """ For searching/creating offers within their date range """ def get_query_set(self): today = datetime.date.today() return super(ActiveOfferManager, self).get_query_set().filter( start_date__lte=today, end_date__gte=today)
<commit_before>import datetime from django.db import models class ActiveOfferManager(models.Manager): """ For searching/creating ACTIVE offers only. """ def get_query_set(self): today = datetime.date.today() return super(ActiveOfferManager, self).get_query_set().filter( start_date__lte=today, end_date__gt=today) <commit_msg>Fix queryset filtering for active offers<commit_after>
import datetime from django.db import models class ActiveOfferManager(models.Manager): """ For searching/creating offers within their date range """ def get_query_set(self): today = datetime.date.today() return super(ActiveOfferManager, self).get_query_set().filter( start_date__lte=today, end_date__gte=today)
import datetime from django.db import models class ActiveOfferManager(models.Manager): """ For searching/creating ACTIVE offers only. """ def get_query_set(self): today = datetime.date.today() return super(ActiveOfferManager, self).get_query_set().filter( start_date__lte=today, end_date__gt=today) Fix queryset filtering for active offersimport datetime from django.db import models class ActiveOfferManager(models.Manager): """ For searching/creating offers within their date range """ def get_query_set(self): today = datetime.date.today() return super(ActiveOfferManager, self).get_query_set().filter( start_date__lte=today, end_date__gte=today)
<commit_before>import datetime from django.db import models class ActiveOfferManager(models.Manager): """ For searching/creating ACTIVE offers only. """ def get_query_set(self): today = datetime.date.today() return super(ActiveOfferManager, self).get_query_set().filter( start_date__lte=today, end_date__gt=today) <commit_msg>Fix queryset filtering for active offers<commit_after>import datetime from django.db import models class ActiveOfferManager(models.Manager): """ For searching/creating offers within their date range """ def get_query_set(self): today = datetime.date.today() return super(ActiveOfferManager, self).get_query_set().filter( start_date__lte=today, end_date__gte=today)
6e1337f7079ba48aafcde59e4d5806caabb0bc29
navigation_extensions.py
navigation_extensions.py
from django.utils.text import capfirst from django.utils.translation import ugettext_lazy as _ from feincms.module.page.extensions.navigation import NavigationExtension, PagePretender class ZivinetzNavigationExtension(NavigationExtension): name = _('Zivinetz navigation extension') def children(self, page, **kwargs): url = page.get_navigation_url() return [ PagePretender( title=capfirst(_('drudges')), url='%sdrudges/' % url, level=3, tree_id=page.tree_id, ), PagePretender( title=capfirst(_('regional offices')), url='%sregional_offices/' % url, level=3, tree_id=page.tree_id, ), PagePretender( title=capfirst(_('scope statements')), url='%sscope_statements/' % url, level=3, tree_id=page.tree_id, ), PagePretender( title=capfirst(_('assignments')), url='%sassignments/' % url, level=3, tree_id=page.tree_id, ), PagePretender( title=capfirst(_('expense reports')), url='%sexpense_reports/' % url, level=3, tree_id=page.tree_id, ), ]
from django.utils.text import capfirst from django.utils.translation import ugettext_lazy as _ from feincms.module.page.extensions.navigation import NavigationExtension, PagePretender class ZivinetzNavigationExtension(NavigationExtension): name = _('Zivinetz navigation extension') def children(self, page, **kwargs): url = page.get_navigation_url() return [ PagePretender( title=capfirst(_('drudges')), url='%sdrudges/' % url, level=page.level+1, tree_id=page.tree_id, ), PagePretender( title=capfirst(_('regional offices')), url='%sregional_offices/' % url, level=page.level+1, tree_id=page.tree_id, ), PagePretender( title=capfirst(_('scope statements')), url='%sscope_statements/' % url, level=page.level+1, tree_id=page.tree_id, ), PagePretender( title=capfirst(_('assignments')), url='%sassignments/' % url, level=page.level+1, tree_id=page.tree_id, ), PagePretender( title=capfirst(_('expense reports')), url='%sexpense_reports/' % url, level=page.level+1, tree_id=page.tree_id, ), ]
Stop hard-coding the navigation level in the extension
Stop hard-coding the navigation level in the extension
Python
mit
matthiask/zivinetz,matthiask/zivinetz,matthiask/zivinetz,matthiask/zivinetz
from django.utils.text import capfirst from django.utils.translation import ugettext_lazy as _ from feincms.module.page.extensions.navigation import NavigationExtension, PagePretender class ZivinetzNavigationExtension(NavigationExtension): name = _('Zivinetz navigation extension') def children(self, page, **kwargs): url = page.get_navigation_url() return [ PagePretender( title=capfirst(_('drudges')), url='%sdrudges/' % url, level=3, tree_id=page.tree_id, ), PagePretender( title=capfirst(_('regional offices')), url='%sregional_offices/' % url, level=3, tree_id=page.tree_id, ), PagePretender( title=capfirst(_('scope statements')), url='%sscope_statements/' % url, level=3, tree_id=page.tree_id, ), PagePretender( title=capfirst(_('assignments')), url='%sassignments/' % url, level=3, tree_id=page.tree_id, ), PagePretender( title=capfirst(_('expense reports')), url='%sexpense_reports/' % url, level=3, tree_id=page.tree_id, ), ] Stop hard-coding the navigation level in the extension
from django.utils.text import capfirst from django.utils.translation import ugettext_lazy as _ from feincms.module.page.extensions.navigation import NavigationExtension, PagePretender class ZivinetzNavigationExtension(NavigationExtension): name = _('Zivinetz navigation extension') def children(self, page, **kwargs): url = page.get_navigation_url() return [ PagePretender( title=capfirst(_('drudges')), url='%sdrudges/' % url, level=page.level+1, tree_id=page.tree_id, ), PagePretender( title=capfirst(_('regional offices')), url='%sregional_offices/' % url, level=page.level+1, tree_id=page.tree_id, ), PagePretender( title=capfirst(_('scope statements')), url='%sscope_statements/' % url, level=page.level+1, tree_id=page.tree_id, ), PagePretender( title=capfirst(_('assignments')), url='%sassignments/' % url, level=page.level+1, tree_id=page.tree_id, ), PagePretender( title=capfirst(_('expense reports')), url='%sexpense_reports/' % url, level=page.level+1, tree_id=page.tree_id, ), ]
<commit_before>from django.utils.text import capfirst from django.utils.translation import ugettext_lazy as _ from feincms.module.page.extensions.navigation import NavigationExtension, PagePretender class ZivinetzNavigationExtension(NavigationExtension): name = _('Zivinetz navigation extension') def children(self, page, **kwargs): url = page.get_navigation_url() return [ PagePretender( title=capfirst(_('drudges')), url='%sdrudges/' % url, level=3, tree_id=page.tree_id, ), PagePretender( title=capfirst(_('regional offices')), url='%sregional_offices/' % url, level=3, tree_id=page.tree_id, ), PagePretender( title=capfirst(_('scope statements')), url='%sscope_statements/' % url, level=3, tree_id=page.tree_id, ), PagePretender( title=capfirst(_('assignments')), url='%sassignments/' % url, level=3, tree_id=page.tree_id, ), PagePretender( title=capfirst(_('expense reports')), url='%sexpense_reports/' % url, level=3, tree_id=page.tree_id, ), ] <commit_msg>Stop hard-coding the navigation level in the extension<commit_after>
from django.utils.text import capfirst from django.utils.translation import ugettext_lazy as _ from feincms.module.page.extensions.navigation import NavigationExtension, PagePretender class ZivinetzNavigationExtension(NavigationExtension): name = _('Zivinetz navigation extension') def children(self, page, **kwargs): url = page.get_navigation_url() return [ PagePretender( title=capfirst(_('drudges')), url='%sdrudges/' % url, level=page.level+1, tree_id=page.tree_id, ), PagePretender( title=capfirst(_('regional offices')), url='%sregional_offices/' % url, level=page.level+1, tree_id=page.tree_id, ), PagePretender( title=capfirst(_('scope statements')), url='%sscope_statements/' % url, level=page.level+1, tree_id=page.tree_id, ), PagePretender( title=capfirst(_('assignments')), url='%sassignments/' % url, level=page.level+1, tree_id=page.tree_id, ), PagePretender( title=capfirst(_('expense reports')), url='%sexpense_reports/' % url, level=page.level+1, tree_id=page.tree_id, ), ]
from django.utils.text import capfirst from django.utils.translation import ugettext_lazy as _ from feincms.module.page.extensions.navigation import NavigationExtension, PagePretender class ZivinetzNavigationExtension(NavigationExtension): name = _('Zivinetz navigation extension') def children(self, page, **kwargs): url = page.get_navigation_url() return [ PagePretender( title=capfirst(_('drudges')), url='%sdrudges/' % url, level=3, tree_id=page.tree_id, ), PagePretender( title=capfirst(_('regional offices')), url='%sregional_offices/' % url, level=3, tree_id=page.tree_id, ), PagePretender( title=capfirst(_('scope statements')), url='%sscope_statements/' % url, level=3, tree_id=page.tree_id, ), PagePretender( title=capfirst(_('assignments')), url='%sassignments/' % url, level=3, tree_id=page.tree_id, ), PagePretender( title=capfirst(_('expense reports')), url='%sexpense_reports/' % url, level=3, tree_id=page.tree_id, ), ] Stop hard-coding the navigation level in the extensionfrom django.utils.text import capfirst from django.utils.translation import ugettext_lazy as _ from feincms.module.page.extensions.navigation import NavigationExtension, PagePretender class ZivinetzNavigationExtension(NavigationExtension): name = _('Zivinetz navigation extension') def children(self, page, **kwargs): url = page.get_navigation_url() return [ PagePretender( title=capfirst(_('drudges')), url='%sdrudges/' % url, level=page.level+1, tree_id=page.tree_id, ), PagePretender( title=capfirst(_('regional offices')), url='%sregional_offices/' % url, level=page.level+1, tree_id=page.tree_id, ), PagePretender( title=capfirst(_('scope statements')), url='%sscope_statements/' % url, level=page.level+1, tree_id=page.tree_id, ), PagePretender( title=capfirst(_('assignments')), url='%sassignments/' % url, level=page.level+1, tree_id=page.tree_id, ), PagePretender( title=capfirst(_('expense reports')), url='%sexpense_reports/' % url, level=page.level+1, tree_id=page.tree_id, ), ]
<commit_before>from django.utils.text import capfirst from django.utils.translation import ugettext_lazy as _ from feincms.module.page.extensions.navigation import NavigationExtension, PagePretender class ZivinetzNavigationExtension(NavigationExtension): name = _('Zivinetz navigation extension') def children(self, page, **kwargs): url = page.get_navigation_url() return [ PagePretender( title=capfirst(_('drudges')), url='%sdrudges/' % url, level=3, tree_id=page.tree_id, ), PagePretender( title=capfirst(_('regional offices')), url='%sregional_offices/' % url, level=3, tree_id=page.tree_id, ), PagePretender( title=capfirst(_('scope statements')), url='%sscope_statements/' % url, level=3, tree_id=page.tree_id, ), PagePretender( title=capfirst(_('assignments')), url='%sassignments/' % url, level=3, tree_id=page.tree_id, ), PagePretender( title=capfirst(_('expense reports')), url='%sexpense_reports/' % url, level=3, tree_id=page.tree_id, ), ] <commit_msg>Stop hard-coding the navigation level in the extension<commit_after>from django.utils.text import capfirst from django.utils.translation import ugettext_lazy as _ from feincms.module.page.extensions.navigation import NavigationExtension, PagePretender class ZivinetzNavigationExtension(NavigationExtension): name = _('Zivinetz navigation extension') def children(self, page, **kwargs): url = page.get_navigation_url() return [ PagePretender( title=capfirst(_('drudges')), url='%sdrudges/' % url, level=page.level+1, tree_id=page.tree_id, ), PagePretender( title=capfirst(_('regional offices')), url='%sregional_offices/' % url, level=page.level+1, tree_id=page.tree_id, ), PagePretender( title=capfirst(_('scope statements')), url='%sscope_statements/' % url, level=page.level+1, tree_id=page.tree_id, ), PagePretender( title=capfirst(_('assignments')), url='%sassignments/' % url, level=page.level+1, tree_id=page.tree_id, ), PagePretender( title=capfirst(_('expense reports')), url='%sexpense_reports/' % url, level=page.level+1, tree_id=page.tree_id, ), ]
8608283592338960c80113ff4d68f42936ddb969
linter.py
linter.py
# # linter.py # Linter for SublimeLinter3, a code checking framework for Sublime Text 3 # # Written by Gregory Oschwald # Copyright (c) 2013 Gregory Oschwald # # License: MIT # """This module exports the Perl plugin class.""" import shlex from SublimeLinter.lint import Linter, util class Perl(Linter): """Provides an interface to perl -c.""" syntax = ('modernperl', 'perl') executable = 'perl' base_cmd = ('perl -c') regex = r'(?P<message>.+?) at .+? line (?P<line>\d+)(, near "(?P<near>.+?)")?' error_stream = util.STREAM_STDERR def cmd(self): """ Return the command line to execute. Overridden so we can add include paths based on the 'include_dirs' settings. """ full_cmd = self.base_cmd settings = self.get_view_settings() include_dirs = settings.get('include_dirs', []) if include_dirs: full_cmd += ' ' . join([' -I ' + shlex.quote(include) for include in include_dirs]) return full_cmd
# # linter.py # Linter for SublimeLinter3, a code checking framework for Sublime Text 3 # # Written by Gregory Oschwald # Copyright (c) 2013 Gregory Oschwald # # License: MIT # """This module exports the Perl plugin class.""" import shlex from SublimeLinter.lint import Linter, util class Perl(Linter): """Provides an interface to perl -c.""" syntax = ('modernperl', 'perl') executable = 'perl' regex = r'(?P<message>.+?) at .+? line (?P<line>\d+)(, near "(?P<near>.+?)")?' error_stream = util.STREAM_STDERR def cmd(self): """ Return the command line to execute. Overridden so we can add include paths based on the 'include_dirs' settings. """ command = [self.executable_path, '-c'] include_dirs = self.get_view_settings().get('include_dirs', []) for e in include_dirs: command.append('-I') command.append(shlex.quote(e)) return command
Clean up include dir code
Clean up include dir code
Python
mit
oschwald/SublimeLinter-perl
# # linter.py # Linter for SublimeLinter3, a code checking framework for Sublime Text 3 # # Written by Gregory Oschwald # Copyright (c) 2013 Gregory Oschwald # # License: MIT # """This module exports the Perl plugin class.""" import shlex from SublimeLinter.lint import Linter, util class Perl(Linter): """Provides an interface to perl -c.""" syntax = ('modernperl', 'perl') executable = 'perl' base_cmd = ('perl -c') regex = r'(?P<message>.+?) at .+? line (?P<line>\d+)(, near "(?P<near>.+?)")?' error_stream = util.STREAM_STDERR def cmd(self): """ Return the command line to execute. Overridden so we can add include paths based on the 'include_dirs' settings. """ full_cmd = self.base_cmd settings = self.get_view_settings() include_dirs = settings.get('include_dirs', []) if include_dirs: full_cmd += ' ' . join([' -I ' + shlex.quote(include) for include in include_dirs]) return full_cmd Clean up include dir code
# # linter.py # Linter for SublimeLinter3, a code checking framework for Sublime Text 3 # # Written by Gregory Oschwald # Copyright (c) 2013 Gregory Oschwald # # License: MIT # """This module exports the Perl plugin class.""" import shlex from SublimeLinter.lint import Linter, util class Perl(Linter): """Provides an interface to perl -c.""" syntax = ('modernperl', 'perl') executable = 'perl' regex = r'(?P<message>.+?) at .+? line (?P<line>\d+)(, near "(?P<near>.+?)")?' error_stream = util.STREAM_STDERR def cmd(self): """ Return the command line to execute. Overridden so we can add include paths based on the 'include_dirs' settings. """ command = [self.executable_path, '-c'] include_dirs = self.get_view_settings().get('include_dirs', []) for e in include_dirs: command.append('-I') command.append(shlex.quote(e)) return command
<commit_before># # linter.py # Linter for SublimeLinter3, a code checking framework for Sublime Text 3 # # Written by Gregory Oschwald # Copyright (c) 2013 Gregory Oschwald # # License: MIT # """This module exports the Perl plugin class.""" import shlex from SublimeLinter.lint import Linter, util class Perl(Linter): """Provides an interface to perl -c.""" syntax = ('modernperl', 'perl') executable = 'perl' base_cmd = ('perl -c') regex = r'(?P<message>.+?) at .+? line (?P<line>\d+)(, near "(?P<near>.+?)")?' error_stream = util.STREAM_STDERR def cmd(self): """ Return the command line to execute. Overridden so we can add include paths based on the 'include_dirs' settings. """ full_cmd = self.base_cmd settings = self.get_view_settings() include_dirs = settings.get('include_dirs', []) if include_dirs: full_cmd += ' ' . join([' -I ' + shlex.quote(include) for include in include_dirs]) return full_cmd <commit_msg>Clean up include dir code<commit_after>
# # linter.py # Linter for SublimeLinter3, a code checking framework for Sublime Text 3 # # Written by Gregory Oschwald # Copyright (c) 2013 Gregory Oschwald # # License: MIT # """This module exports the Perl plugin class.""" import shlex from SublimeLinter.lint import Linter, util class Perl(Linter): """Provides an interface to perl -c.""" syntax = ('modernperl', 'perl') executable = 'perl' regex = r'(?P<message>.+?) at .+? line (?P<line>\d+)(, near "(?P<near>.+?)")?' error_stream = util.STREAM_STDERR def cmd(self): """ Return the command line to execute. Overridden so we can add include paths based on the 'include_dirs' settings. """ command = [self.executable_path, '-c'] include_dirs = self.get_view_settings().get('include_dirs', []) for e in include_dirs: command.append('-I') command.append(shlex.quote(e)) return command
# # linter.py # Linter for SublimeLinter3, a code checking framework for Sublime Text 3 # # Written by Gregory Oschwald # Copyright (c) 2013 Gregory Oschwald # # License: MIT # """This module exports the Perl plugin class.""" import shlex from SublimeLinter.lint import Linter, util class Perl(Linter): """Provides an interface to perl -c.""" syntax = ('modernperl', 'perl') executable = 'perl' base_cmd = ('perl -c') regex = r'(?P<message>.+?) at .+? line (?P<line>\d+)(, near "(?P<near>.+?)")?' error_stream = util.STREAM_STDERR def cmd(self): """ Return the command line to execute. Overridden so we can add include paths based on the 'include_dirs' settings. """ full_cmd = self.base_cmd settings = self.get_view_settings() include_dirs = settings.get('include_dirs', []) if include_dirs: full_cmd += ' ' . join([' -I ' + shlex.quote(include) for include in include_dirs]) return full_cmd Clean up include dir code# # linter.py # Linter for SublimeLinter3, a code checking framework for Sublime Text 3 # # Written by Gregory Oschwald # Copyright (c) 2013 Gregory Oschwald # # License: MIT # """This module exports the Perl plugin class.""" import shlex from SublimeLinter.lint import Linter, util class Perl(Linter): """Provides an interface to perl -c.""" syntax = ('modernperl', 'perl') executable = 'perl' regex = r'(?P<message>.+?) at .+? line (?P<line>\d+)(, near "(?P<near>.+?)")?' error_stream = util.STREAM_STDERR def cmd(self): """ Return the command line to execute. Overridden so we can add include paths based on the 'include_dirs' settings. """ command = [self.executable_path, '-c'] include_dirs = self.get_view_settings().get('include_dirs', []) for e in include_dirs: command.append('-I') command.append(shlex.quote(e)) return command
<commit_before># # linter.py # Linter for SublimeLinter3, a code checking framework for Sublime Text 3 # # Written by Gregory Oschwald # Copyright (c) 2013 Gregory Oschwald # # License: MIT # """This module exports the Perl plugin class.""" import shlex from SublimeLinter.lint import Linter, util class Perl(Linter): """Provides an interface to perl -c.""" syntax = ('modernperl', 'perl') executable = 'perl' base_cmd = ('perl -c') regex = r'(?P<message>.+?) at .+? line (?P<line>\d+)(, near "(?P<near>.+?)")?' error_stream = util.STREAM_STDERR def cmd(self): """ Return the command line to execute. Overridden so we can add include paths based on the 'include_dirs' settings. """ full_cmd = self.base_cmd settings = self.get_view_settings() include_dirs = settings.get('include_dirs', []) if include_dirs: full_cmd += ' ' . join([' -I ' + shlex.quote(include) for include in include_dirs]) return full_cmd <commit_msg>Clean up include dir code<commit_after># # linter.py # Linter for SublimeLinter3, a code checking framework for Sublime Text 3 # # Written by Gregory Oschwald # Copyright (c) 2013 Gregory Oschwald # # License: MIT # """This module exports the Perl plugin class.""" import shlex from SublimeLinter.lint import Linter, util class Perl(Linter): """Provides an interface to perl -c.""" syntax = ('modernperl', 'perl') executable = 'perl' regex = r'(?P<message>.+?) at .+? line (?P<line>\d+)(, near "(?P<near>.+?)")?' error_stream = util.STREAM_STDERR def cmd(self): """ Return the command line to execute. Overridden so we can add include paths based on the 'include_dirs' settings. """ command = [self.executable_path, '-c'] include_dirs = self.get_view_settings().get('include_dirs', []) for e in include_dirs: command.append('-I') command.append(shlex.quote(e)) return command
6d91ed53a15672b1e70d74691158e9afb7162e74
src/etcd/__init__.py
src/etcd/__init__.py
import collections from client import Client class EtcdResult(collections.namedtuple( 'EtcdResult', [ 'action', 'index', 'key', 'prevValue', 'value', 'expiration', 'ttl', 'newKey'])): def __new__( cls, action=None, index=None, key=None, prevValue=None, value=None, expiration=None, ttl=None, newKey=None): return super(EtcdResult, cls).__new__( cls, action, index, key, prevValue, value, expiration, ttl, newKey) class EtcdException(Exception): """ Generic Etcd Exception. """ pass
import collections from client import Client class EtcdResult(collections.namedtuple( 'EtcdResult', [ 'action', 'index', 'key', 'prevValue', 'value', 'expiration', 'ttl', 'newKey'])): def __new__( cls, action=None, index=None, key=None, prevValue=None, value=None, expiration=None, ttl=None, newKey=None): return super(EtcdResult, cls).__new__( cls, action, index, key, prevValue, value, expiration, ttl, newKey) class EtcdException(Exception): """ Generic Etcd Exception. """ pass # Attempt to enable urllib3's SNI support, if possible # Blatantly copied from requests. try: from urllib3.contrib import pyopenssl pyopenssl.inject_into_urllib3() except ImportError: pass
Add optional support for SSL SNI
Add optional support for SSL SNI
Python
mit
dmonroy/python-etcd,ocadotechnology/python-etcd,thepwagner/python-etcd,dmonroy/python-etcd,jlamillan/python-etcd,ocadotechnology/python-etcd,aziontech/python-etcd,jplana/python-etcd,sentinelleader/python-etcd,vodik/python-etcd,projectcalico/python-etcd,j-mcnally/python-etcd,j-mcnally/python-etcd,mbrukman/python-etcd,aziontech/python-etcd,jlamillan/python-etcd,kireal/python-etcd,jplana/python-etcd,kireal/python-etcd,mbrukman/python-etcd,vodik/python-etcd,thepwagner/python-etcd,projectcalico/python-etcd,sentinelleader/python-etcd
import collections from client import Client class EtcdResult(collections.namedtuple( 'EtcdResult', [ 'action', 'index', 'key', 'prevValue', 'value', 'expiration', 'ttl', 'newKey'])): def __new__( cls, action=None, index=None, key=None, prevValue=None, value=None, expiration=None, ttl=None, newKey=None): return super(EtcdResult, cls).__new__( cls, action, index, key, prevValue, value, expiration, ttl, newKey) class EtcdException(Exception): """ Generic Etcd Exception. """ pass Add optional support for SSL SNI
import collections from client import Client class EtcdResult(collections.namedtuple( 'EtcdResult', [ 'action', 'index', 'key', 'prevValue', 'value', 'expiration', 'ttl', 'newKey'])): def __new__( cls, action=None, index=None, key=None, prevValue=None, value=None, expiration=None, ttl=None, newKey=None): return super(EtcdResult, cls).__new__( cls, action, index, key, prevValue, value, expiration, ttl, newKey) class EtcdException(Exception): """ Generic Etcd Exception. """ pass # Attempt to enable urllib3's SNI support, if possible # Blatantly copied from requests. try: from urllib3.contrib import pyopenssl pyopenssl.inject_into_urllib3() except ImportError: pass
<commit_before>import collections from client import Client class EtcdResult(collections.namedtuple( 'EtcdResult', [ 'action', 'index', 'key', 'prevValue', 'value', 'expiration', 'ttl', 'newKey'])): def __new__( cls, action=None, index=None, key=None, prevValue=None, value=None, expiration=None, ttl=None, newKey=None): return super(EtcdResult, cls).__new__( cls, action, index, key, prevValue, value, expiration, ttl, newKey) class EtcdException(Exception): """ Generic Etcd Exception. """ pass <commit_msg>Add optional support for SSL SNI<commit_after>
import collections from client import Client class EtcdResult(collections.namedtuple( 'EtcdResult', [ 'action', 'index', 'key', 'prevValue', 'value', 'expiration', 'ttl', 'newKey'])): def __new__( cls, action=None, index=None, key=None, prevValue=None, value=None, expiration=None, ttl=None, newKey=None): return super(EtcdResult, cls).__new__( cls, action, index, key, prevValue, value, expiration, ttl, newKey) class EtcdException(Exception): """ Generic Etcd Exception. """ pass # Attempt to enable urllib3's SNI support, if possible # Blatantly copied from requests. try: from urllib3.contrib import pyopenssl pyopenssl.inject_into_urllib3() except ImportError: pass
import collections from client import Client class EtcdResult(collections.namedtuple( 'EtcdResult', [ 'action', 'index', 'key', 'prevValue', 'value', 'expiration', 'ttl', 'newKey'])): def __new__( cls, action=None, index=None, key=None, prevValue=None, value=None, expiration=None, ttl=None, newKey=None): return super(EtcdResult, cls).__new__( cls, action, index, key, prevValue, value, expiration, ttl, newKey) class EtcdException(Exception): """ Generic Etcd Exception. """ pass Add optional support for SSL SNIimport collections from client import Client class EtcdResult(collections.namedtuple( 'EtcdResult', [ 'action', 'index', 'key', 'prevValue', 'value', 'expiration', 'ttl', 'newKey'])): def __new__( cls, action=None, index=None, key=None, prevValue=None, value=None, expiration=None, ttl=None, newKey=None): return super(EtcdResult, cls).__new__( cls, action, index, key, prevValue, value, expiration, ttl, newKey) class EtcdException(Exception): """ Generic Etcd Exception. """ pass # Attempt to enable urllib3's SNI support, if possible # Blatantly copied from requests. try: from urllib3.contrib import pyopenssl pyopenssl.inject_into_urllib3() except ImportError: pass
<commit_before>import collections from client import Client class EtcdResult(collections.namedtuple( 'EtcdResult', [ 'action', 'index', 'key', 'prevValue', 'value', 'expiration', 'ttl', 'newKey'])): def __new__( cls, action=None, index=None, key=None, prevValue=None, value=None, expiration=None, ttl=None, newKey=None): return super(EtcdResult, cls).__new__( cls, action, index, key, prevValue, value, expiration, ttl, newKey) class EtcdException(Exception): """ Generic Etcd Exception. """ pass <commit_msg>Add optional support for SSL SNI<commit_after>import collections from client import Client class EtcdResult(collections.namedtuple( 'EtcdResult', [ 'action', 'index', 'key', 'prevValue', 'value', 'expiration', 'ttl', 'newKey'])): def __new__( cls, action=None, index=None, key=None, prevValue=None, value=None, expiration=None, ttl=None, newKey=None): return super(EtcdResult, cls).__new__( cls, action, index, key, prevValue, value, expiration, ttl, newKey) class EtcdException(Exception): """ Generic Etcd Exception. """ pass # Attempt to enable urllib3's SNI support, if possible # Blatantly copied from requests. try: from urllib3.contrib import pyopenssl pyopenssl.inject_into_urllib3() except ImportError: pass
84deb31cc2bdbf30d8b6f30725a3eaaccd1dd903
ade25/assetmanager/browser/repository.py
ade25/assetmanager/browser/repository.py
# -*- coding: utf-8 -*- """Module providing views for asset storage folder""" from Products.Five.browser import BrowserView from plone import api from plone.app.blob.interfaces import IATBlobImage class AssetRepositoryView(BrowserView): """ Folderish content page default view """ def contained_items(self, uid): stack = api.content.get(UID=uid) return stack.restrictedTraverse('@@folderListing')() def item_index(self, uid): return len(self.contained_items(uid)) def preview_image(self, uid): images = self.contained_items(uid) preview = None if len(images): first_item = images[0].getObject() if IATBlobImage.providedBy(first_item): preview = first_item return preview
# -*- coding: utf-8 -*- """Module providing views for asset storage folder""" from Products.Five.browser import BrowserView from plone import api from plone.app.contenttypes.interfaces import IImage class AssetRepositoryView(BrowserView): """ Folderish content page default view """ def contained_items(self, uid): stack = api.content.get(UID=uid) return stack.restrictedTraverse('@@folderListing')() def item_index(self, uid): return len(self.contained_items(uid)) def preview_image(self, uid): images = self.contained_items(uid) preview = None if len(images): first_item = images[0].getObject() if IImage.providedBy(first_item): preview = first_item return preview
Use newer interface for filtering
Use newer interface for filtering The image provided by `plone.app.contenttypes` does reintroduce the global IImage identifier lost during the transition to dexterity based types
Python
mit
ade25/ade25.assetmanager
# -*- coding: utf-8 -*- """Module providing views for asset storage folder""" from Products.Five.browser import BrowserView from plone import api from plone.app.blob.interfaces import IATBlobImage class AssetRepositoryView(BrowserView): """ Folderish content page default view """ def contained_items(self, uid): stack = api.content.get(UID=uid) return stack.restrictedTraverse('@@folderListing')() def item_index(self, uid): return len(self.contained_items(uid)) def preview_image(self, uid): images = self.contained_items(uid) preview = None if len(images): first_item = images[0].getObject() if IATBlobImage.providedBy(first_item): preview = first_item return preview Use newer interface for filtering The image provided by `plone.app.contenttypes` does reintroduce the global IImage identifier lost during the transition to dexterity based types
# -*- coding: utf-8 -*- """Module providing views for asset storage folder""" from Products.Five.browser import BrowserView from plone import api from plone.app.contenttypes.interfaces import IImage class AssetRepositoryView(BrowserView): """ Folderish content page default view """ def contained_items(self, uid): stack = api.content.get(UID=uid) return stack.restrictedTraverse('@@folderListing')() def item_index(self, uid): return len(self.contained_items(uid)) def preview_image(self, uid): images = self.contained_items(uid) preview = None if len(images): first_item = images[0].getObject() if IImage.providedBy(first_item): preview = first_item return preview
<commit_before># -*- coding: utf-8 -*- """Module providing views for asset storage folder""" from Products.Five.browser import BrowserView from plone import api from plone.app.blob.interfaces import IATBlobImage class AssetRepositoryView(BrowserView): """ Folderish content page default view """ def contained_items(self, uid): stack = api.content.get(UID=uid) return stack.restrictedTraverse('@@folderListing')() def item_index(self, uid): return len(self.contained_items(uid)) def preview_image(self, uid): images = self.contained_items(uid) preview = None if len(images): first_item = images[0].getObject() if IATBlobImage.providedBy(first_item): preview = first_item return preview <commit_msg>Use newer interface for filtering The image provided by `plone.app.contenttypes` does reintroduce the global IImage identifier lost during the transition to dexterity based types<commit_after>
# -*- coding: utf-8 -*- """Module providing views for asset storage folder""" from Products.Five.browser import BrowserView from plone import api from plone.app.contenttypes.interfaces import IImage class AssetRepositoryView(BrowserView): """ Folderish content page default view """ def contained_items(self, uid): stack = api.content.get(UID=uid) return stack.restrictedTraverse('@@folderListing')() def item_index(self, uid): return len(self.contained_items(uid)) def preview_image(self, uid): images = self.contained_items(uid) preview = None if len(images): first_item = images[0].getObject() if IImage.providedBy(first_item): preview = first_item return preview
# -*- coding: utf-8 -*- """Module providing views for asset storage folder""" from Products.Five.browser import BrowserView from plone import api from plone.app.blob.interfaces import IATBlobImage class AssetRepositoryView(BrowserView): """ Folderish content page default view """ def contained_items(self, uid): stack = api.content.get(UID=uid) return stack.restrictedTraverse('@@folderListing')() def item_index(self, uid): return len(self.contained_items(uid)) def preview_image(self, uid): images = self.contained_items(uid) preview = None if len(images): first_item = images[0].getObject() if IATBlobImage.providedBy(first_item): preview = first_item return preview Use newer interface for filtering The image provided by `plone.app.contenttypes` does reintroduce the global IImage identifier lost during the transition to dexterity based types# -*- coding: utf-8 -*- """Module providing views for asset storage folder""" from Products.Five.browser import BrowserView from plone import api from plone.app.contenttypes.interfaces import IImage class AssetRepositoryView(BrowserView): """ Folderish content page default view """ def contained_items(self, uid): stack = api.content.get(UID=uid) return stack.restrictedTraverse('@@folderListing')() def item_index(self, uid): return len(self.contained_items(uid)) def preview_image(self, uid): images = self.contained_items(uid) preview = None if len(images): first_item = images[0].getObject() if IImage.providedBy(first_item): preview = first_item return preview
<commit_before># -*- coding: utf-8 -*- """Module providing views for asset storage folder""" from Products.Five.browser import BrowserView from plone import api from plone.app.blob.interfaces import IATBlobImage class AssetRepositoryView(BrowserView): """ Folderish content page default view """ def contained_items(self, uid): stack = api.content.get(UID=uid) return stack.restrictedTraverse('@@folderListing')() def item_index(self, uid): return len(self.contained_items(uid)) def preview_image(self, uid): images = self.contained_items(uid) preview = None if len(images): first_item = images[0].getObject() if IATBlobImage.providedBy(first_item): preview = first_item return preview <commit_msg>Use newer interface for filtering The image provided by `plone.app.contenttypes` does reintroduce the global IImage identifier lost during the transition to dexterity based types<commit_after># -*- coding: utf-8 -*- """Module providing views for asset storage folder""" from Products.Five.browser import BrowserView from plone import api from plone.app.contenttypes.interfaces import IImage class AssetRepositoryView(BrowserView): """ Folderish content page default view """ def contained_items(self, uid): stack = api.content.get(UID=uid) return stack.restrictedTraverse('@@folderListing')() def item_index(self, uid): return len(self.contained_items(uid)) def preview_image(self, uid): images = self.contained_items(uid) preview = None if len(images): first_item = images[0].getObject() if IImage.providedBy(first_item): preview = first_item return preview
1dd257b157cfcb13a13a9c97ff6580045026118c
__openerp__.py
__openerp__.py
# -*- coding: utf-8 -*- ############################################################################## # ############################################################################## { "name": "Account Streamline", "version": "0.1", "author": "XCG Consulting", "category": 'Accounting', "description": """Enhancements to the account module to streamline its usage. """, 'website': 'http://www.openerp-experts.com', 'init_xml': [], "depends": [ 'base', 'account_accountant', 'account_voucher', 'account_payment', 'account_sequence', 'analytic_structure', 'advanced_filter'], "data": [ 'data/partner_data.xml', 'wizard/account_reconcile_view.xml', 'account_move_line_search_unreconciled.xml', 'account_move_line_tree.xml', 'account_move_view.xml', 'account_view.xml', 'partner_view.xml', 'payment_selection.xml', 'account_move_line_journal_items.xml', 'account_menu_entries.xml', 'account_move_line_journal_view.xml', 'data/analytic.code.csv', 'data/analytic.dimension.csv', 'data/analytic.structure.csv' ], #'demo_xml': [], 'test': [], 'installable': True, 'active': False, } # vim:expandtab:smartindent:tabstop=4:softtabstop=4:shiftwidth=4:
# -*- coding: utf-8 -*- ############################################################################## # ############################################################################## { "name": "Account Streamline", "version": "0.1", "author": "XCG Consulting", "category": 'Accounting', "description": """Enhancements to the account module to streamline its usage. """, 'website': 'http://www.openerp-experts.com', 'init_xml': [], "depends": [ 'base', 'account_accountant', 'account_voucher', 'account_payment', 'account_sequence', 'analytic_structure', 'advanced_filter'], "data": [ 'data/partner_data.xml', 'wizard/account_reconcile_view.xml', 'account_move_line_search_unreconciled.xml', 'account_move_line_tree.xml', 'account_move_view.xml', 'account_view.xml', 'partner_view.xml', 'payment_selection.xml', 'account_move_line_journal_items.xml', 'account_move_line_journal_view.xml', 'account_menu_entries.xml', 'data/analytic.code.csv', 'data/analytic.dimension.csv', 'data/analytic.structure.csv' ], #'demo_xml': [], 'test': [], 'installable': True, 'active': False, } # vim:expandtab:smartindent:tabstop=4:softtabstop=4:shiftwidth=4:
Change the order when loading xml data
Change the order when loading xml data
Python
agpl-3.0
xcgd/account_streamline
# -*- coding: utf-8 -*- ############################################################################## # ############################################################################## { "name": "Account Streamline", "version": "0.1", "author": "XCG Consulting", "category": 'Accounting', "description": """Enhancements to the account module to streamline its usage. """, 'website': 'http://www.openerp-experts.com', 'init_xml': [], "depends": [ 'base', 'account_accountant', 'account_voucher', 'account_payment', 'account_sequence', 'analytic_structure', 'advanced_filter'], "data": [ 'data/partner_data.xml', 'wizard/account_reconcile_view.xml', 'account_move_line_search_unreconciled.xml', 'account_move_line_tree.xml', 'account_move_view.xml', 'account_view.xml', 'partner_view.xml', 'payment_selection.xml', 'account_move_line_journal_items.xml', 'account_menu_entries.xml', 'account_move_line_journal_view.xml', 'data/analytic.code.csv', 'data/analytic.dimension.csv', 'data/analytic.structure.csv' ], #'demo_xml': [], 'test': [], 'installable': True, 'active': False, } # vim:expandtab:smartindent:tabstop=4:softtabstop=4:shiftwidth=4: Change the order when loading xml data
# -*- coding: utf-8 -*- ############################################################################## # ############################################################################## { "name": "Account Streamline", "version": "0.1", "author": "XCG Consulting", "category": 'Accounting', "description": """Enhancements to the account module to streamline its usage. """, 'website': 'http://www.openerp-experts.com', 'init_xml': [], "depends": [ 'base', 'account_accountant', 'account_voucher', 'account_payment', 'account_sequence', 'analytic_structure', 'advanced_filter'], "data": [ 'data/partner_data.xml', 'wizard/account_reconcile_view.xml', 'account_move_line_search_unreconciled.xml', 'account_move_line_tree.xml', 'account_move_view.xml', 'account_view.xml', 'partner_view.xml', 'payment_selection.xml', 'account_move_line_journal_items.xml', 'account_move_line_journal_view.xml', 'account_menu_entries.xml', 'data/analytic.code.csv', 'data/analytic.dimension.csv', 'data/analytic.structure.csv' ], #'demo_xml': [], 'test': [], 'installable': True, 'active': False, } # vim:expandtab:smartindent:tabstop=4:softtabstop=4:shiftwidth=4:
<commit_before># -*- coding: utf-8 -*- ############################################################################## # ############################################################################## { "name": "Account Streamline", "version": "0.1", "author": "XCG Consulting", "category": 'Accounting', "description": """Enhancements to the account module to streamline its usage. """, 'website': 'http://www.openerp-experts.com', 'init_xml': [], "depends": [ 'base', 'account_accountant', 'account_voucher', 'account_payment', 'account_sequence', 'analytic_structure', 'advanced_filter'], "data": [ 'data/partner_data.xml', 'wizard/account_reconcile_view.xml', 'account_move_line_search_unreconciled.xml', 'account_move_line_tree.xml', 'account_move_view.xml', 'account_view.xml', 'partner_view.xml', 'payment_selection.xml', 'account_move_line_journal_items.xml', 'account_menu_entries.xml', 'account_move_line_journal_view.xml', 'data/analytic.code.csv', 'data/analytic.dimension.csv', 'data/analytic.structure.csv' ], #'demo_xml': [], 'test': [], 'installable': True, 'active': False, } # vim:expandtab:smartindent:tabstop=4:softtabstop=4:shiftwidth=4: <commit_msg>Change the order when loading xml data<commit_after>
# -*- coding: utf-8 -*- ############################################################################## # ############################################################################## { "name": "Account Streamline", "version": "0.1", "author": "XCG Consulting", "category": 'Accounting', "description": """Enhancements to the account module to streamline its usage. """, 'website': 'http://www.openerp-experts.com', 'init_xml': [], "depends": [ 'base', 'account_accountant', 'account_voucher', 'account_payment', 'account_sequence', 'analytic_structure', 'advanced_filter'], "data": [ 'data/partner_data.xml', 'wizard/account_reconcile_view.xml', 'account_move_line_search_unreconciled.xml', 'account_move_line_tree.xml', 'account_move_view.xml', 'account_view.xml', 'partner_view.xml', 'payment_selection.xml', 'account_move_line_journal_items.xml', 'account_move_line_journal_view.xml', 'account_menu_entries.xml', 'data/analytic.code.csv', 'data/analytic.dimension.csv', 'data/analytic.structure.csv' ], #'demo_xml': [], 'test': [], 'installable': True, 'active': False, } # vim:expandtab:smartindent:tabstop=4:softtabstop=4:shiftwidth=4:
# -*- coding: utf-8 -*- ############################################################################## # ############################################################################## { "name": "Account Streamline", "version": "0.1", "author": "XCG Consulting", "category": 'Accounting', "description": """Enhancements to the account module to streamline its usage. """, 'website': 'http://www.openerp-experts.com', 'init_xml': [], "depends": [ 'base', 'account_accountant', 'account_voucher', 'account_payment', 'account_sequence', 'analytic_structure', 'advanced_filter'], "data": [ 'data/partner_data.xml', 'wizard/account_reconcile_view.xml', 'account_move_line_search_unreconciled.xml', 'account_move_line_tree.xml', 'account_move_view.xml', 'account_view.xml', 'partner_view.xml', 'payment_selection.xml', 'account_move_line_journal_items.xml', 'account_menu_entries.xml', 'account_move_line_journal_view.xml', 'data/analytic.code.csv', 'data/analytic.dimension.csv', 'data/analytic.structure.csv' ], #'demo_xml': [], 'test': [], 'installable': True, 'active': False, } # vim:expandtab:smartindent:tabstop=4:softtabstop=4:shiftwidth=4: Change the order when loading xml data# -*- coding: utf-8 -*- ############################################################################## # ############################################################################## { "name": "Account Streamline", "version": "0.1", "author": "XCG Consulting", "category": 'Accounting', "description": """Enhancements to the account module to streamline its usage. """, 'website': 'http://www.openerp-experts.com', 'init_xml': [], "depends": [ 'base', 'account_accountant', 'account_voucher', 'account_payment', 'account_sequence', 'analytic_structure', 'advanced_filter'], "data": [ 'data/partner_data.xml', 'wizard/account_reconcile_view.xml', 'account_move_line_search_unreconciled.xml', 'account_move_line_tree.xml', 'account_move_view.xml', 'account_view.xml', 'partner_view.xml', 'payment_selection.xml', 'account_move_line_journal_items.xml', 'account_move_line_journal_view.xml', 'account_menu_entries.xml', 'data/analytic.code.csv', 'data/analytic.dimension.csv', 'data/analytic.structure.csv' ], #'demo_xml': [], 'test': [], 'installable': True, 'active': False, } # vim:expandtab:smartindent:tabstop=4:softtabstop=4:shiftwidth=4:
<commit_before># -*- coding: utf-8 -*- ############################################################################## # ############################################################################## { "name": "Account Streamline", "version": "0.1", "author": "XCG Consulting", "category": 'Accounting', "description": """Enhancements to the account module to streamline its usage. """, 'website': 'http://www.openerp-experts.com', 'init_xml': [], "depends": [ 'base', 'account_accountant', 'account_voucher', 'account_payment', 'account_sequence', 'analytic_structure', 'advanced_filter'], "data": [ 'data/partner_data.xml', 'wizard/account_reconcile_view.xml', 'account_move_line_search_unreconciled.xml', 'account_move_line_tree.xml', 'account_move_view.xml', 'account_view.xml', 'partner_view.xml', 'payment_selection.xml', 'account_move_line_journal_items.xml', 'account_menu_entries.xml', 'account_move_line_journal_view.xml', 'data/analytic.code.csv', 'data/analytic.dimension.csv', 'data/analytic.structure.csv' ], #'demo_xml': [], 'test': [], 'installable': True, 'active': False, } # vim:expandtab:smartindent:tabstop=4:softtabstop=4:shiftwidth=4: <commit_msg>Change the order when loading xml data<commit_after># -*- coding: utf-8 -*- ############################################################################## # ############################################################################## { "name": "Account Streamline", "version": "0.1", "author": "XCG Consulting", "category": 'Accounting', "description": """Enhancements to the account module to streamline its usage. """, 'website': 'http://www.openerp-experts.com', 'init_xml': [], "depends": [ 'base', 'account_accountant', 'account_voucher', 'account_payment', 'account_sequence', 'analytic_structure', 'advanced_filter'], "data": [ 'data/partner_data.xml', 'wizard/account_reconcile_view.xml', 'account_move_line_search_unreconciled.xml', 'account_move_line_tree.xml', 'account_move_view.xml', 'account_view.xml', 'partner_view.xml', 'payment_selection.xml', 'account_move_line_journal_items.xml', 'account_move_line_journal_view.xml', 'account_menu_entries.xml', 'data/analytic.code.csv', 'data/analytic.dimension.csv', 'data/analytic.structure.csv' ], #'demo_xml': [], 'test': [], 'installable': True, 'active': False, } # vim:expandtab:smartindent:tabstop=4:softtabstop=4:shiftwidth=4:
ed491860864c363be36d99c09ff0131a5fe00aaf
test/Driver/Dependencies/Inputs/touch.py
test/Driver/Dependencies/Inputs/touch.py
#!/usr/bin/env python # touch.py - /bin/touch that writes the LLVM epoch -*- python -*- # # This source file is part of the Swift.org open source project # # Copyright (c) 2014 - 2016 Apple Inc. and the Swift project authors # Licensed under Apache License v2.0 with Runtime Library Exception # # See http://swift.org/LICENSE.txt for license information # See http://swift.org/CONTRIBUTORS.txt for the list of Swift project authors # # ---------------------------------------------------------------------------- # # Like /bin/touch, but takes a time using the LLVM epoch. # # ---------------------------------------------------------------------------- import os import sys assert len(sys.argv) >= 2 timeVal = int(sys.argv[1]) # offset between Unix and LLVM epochs timeVal += 946684800 # Update the output file mtime, or create it if necessary. # From http://stackoverflow.com/a/1160227. for outputFile in sys.argv[1:]: with open(outputFile, 'a'): os.utime(outputFile, (timeVal, timeVal))
#!/usr/bin/env python # touch.py - /bin/touch that writes the LLVM epoch -*- python -*- # # This source file is part of the Swift.org open source project # # Copyright (c) 2014 - 2016 Apple Inc. and the Swift project authors # Licensed under Apache License v2.0 with Runtime Library Exception # # See http://swift.org/LICENSE.txt for license information # See http://swift.org/CONTRIBUTORS.txt for the list of Swift project authors # # ---------------------------------------------------------------------------- # # Like /bin/touch, but takes a time using the system_clock epoch. # # ---------------------------------------------------------------------------- import os import sys assert len(sys.argv) >= 2 timeVal = int(sys.argv[1]) # Update the output file mtime, or create it if necessary. # From http://stackoverflow.com/a/1160227. for outputFile in sys.argv[1:]: with open(outputFile, 'a'): os.utime(outputFile, (timeVal, timeVal))
Fix tests for file timestamps to drop the LLVM epoch offset.
Fix tests for file timestamps to drop the LLVM epoch offset. Now that Swift is not using LLVM's TimeValue (564fc6f2 and previous commit) there is no offset from the system_clock epoch. The offset could be added into the tests that use touch.py (so the times would not be back in 1984) but I decided not to do that to avoid merge conflicts in the test files.
Python
apache-2.0
aschwaighofer/swift,tinysun212/swift-windows,arvedviehweger/swift,xwu/swift,airspeedswift/swift,parkera/swift,tinysun212/swift-windows,JGiola/swift,JaSpa/swift,JGiola/swift,parkera/swift,zisko/swift,hughbe/swift,codestergit/swift,CodaFi/swift,rudkx/swift,huonw/swift,tkremenek/swift,jtbandes/swift,codestergit/swift,practicalswift/swift,frootloops/swift,tardieu/swift,jckarter/swift,return/swift,benlangmuir/swift,xwu/swift,danielmartin/swift,djwbrown/swift,gribozavr/swift,return/swift,atrick/swift,arvedviehweger/swift,danielmartin/swift,nathawes/swift,tjw/swift,OscarSwanros/swift,shahmishal/swift,benlangmuir/swift,benlangmuir/swift,brentdax/swift,JGiola/swift,amraboelela/swift,roambotics/swift,tardieu/swift,tinysun212/swift-windows,jmgc/swift,parkera/swift,tardieu/swift,danielmartin/swift,gribozavr/swift,sschiau/swift,natecook1000/swift,huonw/swift,karwa/swift,gregomni/swift,harlanhaskins/swift,aschwaighofer/swift,practicalswift/swift,manavgabhawala/swift,glessard/swift,uasys/swift,gregomni/swift,allevato/swift,jckarter/swift,alblue/swift,natecook1000/swift,amraboelela/swift,xedin/swift,gottesmm/swift,hughbe/swift,harlanhaskins/swift,felix91gr/swift,return/swift,tkremenek/swift,shajrawi/swift,allevato/swift,sschiau/swift,sschiau/swift,stephentyrone/swift,tkremenek/swift,uasys/swift,JaSpa/swift,stephentyrone/swift,apple/swift,milseman/swift,lorentey/swift,shahmishal/swift,shajrawi/swift,atrick/swift,practicalswift/swift,xwu/swift,gribozavr/swift,return/swift,xedin/swift,hooman/swift,CodaFi/swift,gregomni/swift,apple/swift,austinzheng/swift,zisko/swift,milseman/swift,huonw/swift,jopamer/swift,jckarter/swift,benlangmuir/swift,stephentyrone/swift,xwu/swift,tardieu/swift,amraboelela/swift,swiftix/swift,CodaFi/swift,bitjammer/swift,danielmartin/swift,huonw/swift,JaSpa/swift,xedin/swift,austinzheng/swift,lorentey/swift,huonw/swift,deyton/swift,nathawes/swift,lorentey/swift,apple/swift,austinzheng/swift,harlanhaskins/swift,milseman/swift,glessard/swift,devincoughlin/swift,airspeedswift/swift,OscarSwanros/swift,glessard/swift,felix91gr/swift,OscarSwanros/swift,nathawes/swift,karwa/swift,gregomni/swift,tardieu/swift,felix91gr/swift,hooman/swift,apple/swift,atrick/swift,parkera/swift,gottesmm/swift,deyton/swift,aschwaighofer/swift,devincoughlin/swift,arvedviehweger/swift,bitjammer/swift,harlanhaskins/swift,gregomni/swift,felix91gr/swift,alblue/swift,shajrawi/swift,zisko/swift,jmgc/swift,tjw/swift,sschiau/swift,frootloops/swift,jtbandes/swift,xedin/swift,ahoppen/swift,xedin/swift,austinzheng/swift,JaSpa/swift,shahmishal/swift,atrick/swift,brentdax/swift,arvedviehweger/swift,brentdax/swift,tardieu/swift,JGiola/swift,xwu/swift,devincoughlin/swift,xwu/swift,calebd/swift,lorentey/swift,xedin/swift,jopamer/swift,atrick/swift,rudkx/swift,aschwaighofer/swift,jtbandes/swift,natecook1000/swift,hooman/swift,natecook1000/swift,tinysun212/swift-windows,uasys/swift,danielmartin/swift,airspeedswift/swift,brentdax/swift,swiftix/swift,devincoughlin/swift,natecook1000/swift,hooman/swift,codestergit/swift,tjw/swift,Jnosh/swift,swiftix/swift,gribozavr/swift,JGiola/swift,Jnosh/swift,harlanhaskins/swift,gottesmm/swift,hughbe/swift,tkremenek/swift,CodaFi/swift,hooman/swift,alblue/swift,karwa/swift,brentdax/swift,sschiau/swift,bitjammer/swift,uasys/swift,sschiau/swift,allevato/swift,return/swift,austinzheng/swift,gregomni/swift,arvedviehweger/swift,uasys/swift,zisko/swift,stephentyrone/swift,swiftix/swift,uasys/swift,shajrawi/swift,CodaFi/swift,lorentey/swift,hughbe/swift,bitjammer/swift,austinzheng/swift,tkremenek/swift,devincoughlin/swift,stephentyrone/swift,aschwaighofer/swift,frootloops/swift,xwu/swift,manavgabhawala/swift,manavgabhawala/swift,manavgabhawala/swift,Jnosh/swift,bitjammer/swift,bitjammer/swift,OscarSwanros/swift,karwa/swift,shahmishal/swift,hughbe/swift,milseman/swift,atrick/swift,hooman/swift,austinzheng/swift,glessard/swift,JGiola/swift,danielmartin/swift,parkera/swift,parkera/swift,return/swift,zisko/swift,glessard/swift,milseman/swift,alblue/swift,tkremenek/swift,tardieu/swift,rudkx/swift,jmgc/swift,shajrawi/swift,frootloops/swift,calebd/swift,tjw/swift,karwa/swift,ahoppen/swift,allevato/swift,tjw/swift,frootloops/swift,sschiau/swift,CodaFi/swift,felix91gr/swift,shahmishal/swift,aschwaighofer/swift,arvedviehweger/swift,codestergit/swift,JaSpa/swift,stephentyrone/swift,jopamer/swift,OscarSwanros/swift,apple/swift,apple/swift,jckarter/swift,djwbrown/swift,practicalswift/swift,deyton/swift,jopamer/swift,parkera/swift,harlanhaskins/swift,jmgc/swift,roambotics/swift,djwbrown/swift,gribozavr/swift,benlangmuir/swift,manavgabhawala/swift,parkera/swift,huonw/swift,amraboelela/swift,jtbandes/swift,tinysun212/swift-windows,shajrawi/swift,felix91gr/swift,roambotics/swift,ahoppen/swift,airspeedswift/swift,jtbandes/swift,djwbrown/swift,ahoppen/swift,tinysun212/swift-windows,devincoughlin/swift,tinysun212/swift-windows,JaSpa/swift,swiftix/swift,benlangmuir/swift,jtbandes/swift,amraboelela/swift,codestergit/swift,uasys/swift,milseman/swift,devincoughlin/swift,felix91gr/swift,lorentey/swift,OscarSwanros/swift,codestergit/swift,calebd/swift,Jnosh/swift,tjw/swift,rudkx/swift,manavgabhawala/swift,JaSpa/swift,zisko/swift,xedin/swift,jopamer/swift,allevato/swift,rudkx/swift,roambotics/swift,alblue/swift,rudkx/swift,lorentey/swift,shajrawi/swift,practicalswift/swift,jckarter/swift,stephentyrone/swift,natecook1000/swift,gribozavr/swift,calebd/swift,tjw/swift,practicalswift/swift,arvedviehweger/swift,gottesmm/swift,nathawes/swift,shahmishal/swift,Jnosh/swift,Jnosh/swift,hughbe/swift,nathawes/swift,practicalswift/swift,jmgc/swift,danielmartin/swift,jckarter/swift,devincoughlin/swift,shahmishal/swift,bitjammer/swift,shahmishal/swift,jtbandes/swift,jopamer/swift,airspeedswift/swift,gribozavr/swift,nathawes/swift,swiftix/swift,ahoppen/swift,djwbrown/swift,manavgabhawala/swift,hooman/swift,Jnosh/swift,codestergit/swift,gottesmm/swift,aschwaighofer/swift,sschiau/swift,calebd/swift,natecook1000/swift,roambotics/swift,allevato/swift,zisko/swift,karwa/swift,glessard/swift,airspeedswift/swift,gottesmm/swift,huonw/swift,djwbrown/swift,harlanhaskins/swift,hughbe/swift,practicalswift/swift,brentdax/swift,jopamer/swift,gottesmm/swift,return/swift,swiftix/swift,alblue/swift,frootloops/swift,amraboelela/swift,airspeedswift/swift,calebd/swift,karwa/swift,deyton/swift,allevato/swift,nathawes/swift,brentdax/swift,milseman/swift,tkremenek/swift,deyton/swift,gribozavr/swift,djwbrown/swift,OscarSwanros/swift,shajrawi/swift,lorentey/swift,jmgc/swift,xedin/swift,amraboelela/swift,jmgc/swift,frootloops/swift,CodaFi/swift,jckarter/swift,roambotics/swift,calebd/swift,alblue/swift,karwa/swift,deyton/swift,ahoppen/swift,deyton/swift
#!/usr/bin/env python # touch.py - /bin/touch that writes the LLVM epoch -*- python -*- # # This source file is part of the Swift.org open source project # # Copyright (c) 2014 - 2016 Apple Inc. and the Swift project authors # Licensed under Apache License v2.0 with Runtime Library Exception # # See http://swift.org/LICENSE.txt for license information # See http://swift.org/CONTRIBUTORS.txt for the list of Swift project authors # # ---------------------------------------------------------------------------- # # Like /bin/touch, but takes a time using the LLVM epoch. # # ---------------------------------------------------------------------------- import os import sys assert len(sys.argv) >= 2 timeVal = int(sys.argv[1]) # offset between Unix and LLVM epochs timeVal += 946684800 # Update the output file mtime, or create it if necessary. # From http://stackoverflow.com/a/1160227. for outputFile in sys.argv[1:]: with open(outputFile, 'a'): os.utime(outputFile, (timeVal, timeVal)) Fix tests for file timestamps to drop the LLVM epoch offset. Now that Swift is not using LLVM's TimeValue (564fc6f2 and previous commit) there is no offset from the system_clock epoch. The offset could be added into the tests that use touch.py (so the times would not be back in 1984) but I decided not to do that to avoid merge conflicts in the test files.
#!/usr/bin/env python # touch.py - /bin/touch that writes the LLVM epoch -*- python -*- # # This source file is part of the Swift.org open source project # # Copyright (c) 2014 - 2016 Apple Inc. and the Swift project authors # Licensed under Apache License v2.0 with Runtime Library Exception # # See http://swift.org/LICENSE.txt for license information # See http://swift.org/CONTRIBUTORS.txt for the list of Swift project authors # # ---------------------------------------------------------------------------- # # Like /bin/touch, but takes a time using the system_clock epoch. # # ---------------------------------------------------------------------------- import os import sys assert len(sys.argv) >= 2 timeVal = int(sys.argv[1]) # Update the output file mtime, or create it if necessary. # From http://stackoverflow.com/a/1160227. for outputFile in sys.argv[1:]: with open(outputFile, 'a'): os.utime(outputFile, (timeVal, timeVal))
<commit_before>#!/usr/bin/env python # touch.py - /bin/touch that writes the LLVM epoch -*- python -*- # # This source file is part of the Swift.org open source project # # Copyright (c) 2014 - 2016 Apple Inc. and the Swift project authors # Licensed under Apache License v2.0 with Runtime Library Exception # # See http://swift.org/LICENSE.txt for license information # See http://swift.org/CONTRIBUTORS.txt for the list of Swift project authors # # ---------------------------------------------------------------------------- # # Like /bin/touch, but takes a time using the LLVM epoch. # # ---------------------------------------------------------------------------- import os import sys assert len(sys.argv) >= 2 timeVal = int(sys.argv[1]) # offset between Unix and LLVM epochs timeVal += 946684800 # Update the output file mtime, or create it if necessary. # From http://stackoverflow.com/a/1160227. for outputFile in sys.argv[1:]: with open(outputFile, 'a'): os.utime(outputFile, (timeVal, timeVal)) <commit_msg>Fix tests for file timestamps to drop the LLVM epoch offset. Now that Swift is not using LLVM's TimeValue (564fc6f2 and previous commit) there is no offset from the system_clock epoch. The offset could be added into the tests that use touch.py (so the times would not be back in 1984) but I decided not to do that to avoid merge conflicts in the test files.<commit_after>
#!/usr/bin/env python # touch.py - /bin/touch that writes the LLVM epoch -*- python -*- # # This source file is part of the Swift.org open source project # # Copyright (c) 2014 - 2016 Apple Inc. and the Swift project authors # Licensed under Apache License v2.0 with Runtime Library Exception # # See http://swift.org/LICENSE.txt for license information # See http://swift.org/CONTRIBUTORS.txt for the list of Swift project authors # # ---------------------------------------------------------------------------- # # Like /bin/touch, but takes a time using the system_clock epoch. # # ---------------------------------------------------------------------------- import os import sys assert len(sys.argv) >= 2 timeVal = int(sys.argv[1]) # Update the output file mtime, or create it if necessary. # From http://stackoverflow.com/a/1160227. for outputFile in sys.argv[1:]: with open(outputFile, 'a'): os.utime(outputFile, (timeVal, timeVal))
#!/usr/bin/env python # touch.py - /bin/touch that writes the LLVM epoch -*- python -*- # # This source file is part of the Swift.org open source project # # Copyright (c) 2014 - 2016 Apple Inc. and the Swift project authors # Licensed under Apache License v2.0 with Runtime Library Exception # # See http://swift.org/LICENSE.txt for license information # See http://swift.org/CONTRIBUTORS.txt for the list of Swift project authors # # ---------------------------------------------------------------------------- # # Like /bin/touch, but takes a time using the LLVM epoch. # # ---------------------------------------------------------------------------- import os import sys assert len(sys.argv) >= 2 timeVal = int(sys.argv[1]) # offset between Unix and LLVM epochs timeVal += 946684800 # Update the output file mtime, or create it if necessary. # From http://stackoverflow.com/a/1160227. for outputFile in sys.argv[1:]: with open(outputFile, 'a'): os.utime(outputFile, (timeVal, timeVal)) Fix tests for file timestamps to drop the LLVM epoch offset. Now that Swift is not using LLVM's TimeValue (564fc6f2 and previous commit) there is no offset from the system_clock epoch. The offset could be added into the tests that use touch.py (so the times would not be back in 1984) but I decided not to do that to avoid merge conflicts in the test files.#!/usr/bin/env python # touch.py - /bin/touch that writes the LLVM epoch -*- python -*- # # This source file is part of the Swift.org open source project # # Copyright (c) 2014 - 2016 Apple Inc. and the Swift project authors # Licensed under Apache License v2.0 with Runtime Library Exception # # See http://swift.org/LICENSE.txt for license information # See http://swift.org/CONTRIBUTORS.txt for the list of Swift project authors # # ---------------------------------------------------------------------------- # # Like /bin/touch, but takes a time using the system_clock epoch. # # ---------------------------------------------------------------------------- import os import sys assert len(sys.argv) >= 2 timeVal = int(sys.argv[1]) # Update the output file mtime, or create it if necessary. # From http://stackoverflow.com/a/1160227. for outputFile in sys.argv[1:]: with open(outputFile, 'a'): os.utime(outputFile, (timeVal, timeVal))
<commit_before>#!/usr/bin/env python # touch.py - /bin/touch that writes the LLVM epoch -*- python -*- # # This source file is part of the Swift.org open source project # # Copyright (c) 2014 - 2016 Apple Inc. and the Swift project authors # Licensed under Apache License v2.0 with Runtime Library Exception # # See http://swift.org/LICENSE.txt for license information # See http://swift.org/CONTRIBUTORS.txt for the list of Swift project authors # # ---------------------------------------------------------------------------- # # Like /bin/touch, but takes a time using the LLVM epoch. # # ---------------------------------------------------------------------------- import os import sys assert len(sys.argv) >= 2 timeVal = int(sys.argv[1]) # offset between Unix and LLVM epochs timeVal += 946684800 # Update the output file mtime, or create it if necessary. # From http://stackoverflow.com/a/1160227. for outputFile in sys.argv[1:]: with open(outputFile, 'a'): os.utime(outputFile, (timeVal, timeVal)) <commit_msg>Fix tests for file timestamps to drop the LLVM epoch offset. Now that Swift is not using LLVM's TimeValue (564fc6f2 and previous commit) there is no offset from the system_clock epoch. The offset could be added into the tests that use touch.py (so the times would not be back in 1984) but I decided not to do that to avoid merge conflicts in the test files.<commit_after>#!/usr/bin/env python # touch.py - /bin/touch that writes the LLVM epoch -*- python -*- # # This source file is part of the Swift.org open source project # # Copyright (c) 2014 - 2016 Apple Inc. and the Swift project authors # Licensed under Apache License v2.0 with Runtime Library Exception # # See http://swift.org/LICENSE.txt for license information # See http://swift.org/CONTRIBUTORS.txt for the list of Swift project authors # # ---------------------------------------------------------------------------- # # Like /bin/touch, but takes a time using the system_clock epoch. # # ---------------------------------------------------------------------------- import os import sys assert len(sys.argv) >= 2 timeVal = int(sys.argv[1]) # Update the output file mtime, or create it if necessary. # From http://stackoverflow.com/a/1160227. for outputFile in sys.argv[1:]: with open(outputFile, 'a'): os.utime(outputFile, (timeVal, timeVal))
2da44025051e6a0c0bffe4dba25ed7084e7ae277
test/connect_remote/TestConnectRemote.py
test/connect_remote/TestConnectRemote.py
""" Test lldb 'process connect' command. """ import os import unittest2 import lldb import pexpect from lldbtest import * class ConnectRemoteTestCase(TestBase): mydir = "connect_remote" @unittest2.expectedFailure def test_connect_remote(self): """Test "process connect connect:://localhost:12345".""" # First, we'll start a fake debugserver (a simple echo server). fakeserver = pexpect.spawn('./EchoServer.py') # Turn on logging for what the child sends back. if self.TraceOn(): fakeserver.logfile_read = sys.stdout # Schedule the fake debugserver to be shutting down during teardown. def shutdown_fakeserver(): fakeserver.close() self.addTearDownHook(shutdown_fakeserver) # Wait until we receive the server ready message before continuing. fakeserver.expect_exact('Listening on localhost:12345') # Connect to the fake server.... self.runCmd("process connect connect://localhost:12345") if __name__ == '__main__': import atexit lldb.SBDebugger.Initialize() atexit.register(lambda: lldb.SBDebugger.Terminate()) unittest2.main()
""" Test lldb 'process connect' command. """ import os import unittest2 import lldb import pexpect from lldbtest import * class ConnectRemoteTestCase(TestBase): mydir = "connect_remote" def test_connect_remote(self): """Test "process connect connect:://localhost:12345".""" # First, we'll start a fake debugserver (a simple echo server). fakeserver = pexpect.spawn('./EchoServer.py') # Turn on logging for what the child sends back. if self.TraceOn(): fakeserver.logfile_read = sys.stdout # Schedule the fake debugserver to be shutting down during teardown. def shutdown_fakeserver(): fakeserver.close() self.addTearDownHook(shutdown_fakeserver) # Wait until we receive the server ready message before continuing. fakeserver.expect_exact('Listening on localhost:12345') # Connect to the fake server.... self.runCmd("process connect connect://localhost:12345") if __name__ == '__main__': import atexit lldb.SBDebugger.Initialize() atexit.register(lambda: lldb.SBDebugger.Terminate()) unittest2.main()
Test case test_connect_remote() has been passing consistently for some times. Let's remove the @expectedFailure marker from it.
Test case test_connect_remote() has been passing consistently for some times. Let's remove the @expectedFailure marker from it. git-svn-id: b33bab8abb5b18c12ee100cd7761ab452d00b2b0@133294 91177308-0d34-0410-b5e6-96231b3b80d8
Python
apache-2.0
apple/swift-lldb,apple/swift-lldb,apple/swift-lldb,apple/swift-lldb,apple/swift-lldb,llvm-mirror/lldb,llvm-mirror/lldb,llvm-mirror/lldb,llvm-mirror/lldb,apple/swift-lldb,llvm-mirror/lldb
""" Test lldb 'process connect' command. """ import os import unittest2 import lldb import pexpect from lldbtest import * class ConnectRemoteTestCase(TestBase): mydir = "connect_remote" @unittest2.expectedFailure def test_connect_remote(self): """Test "process connect connect:://localhost:12345".""" # First, we'll start a fake debugserver (a simple echo server). fakeserver = pexpect.spawn('./EchoServer.py') # Turn on logging for what the child sends back. if self.TraceOn(): fakeserver.logfile_read = sys.stdout # Schedule the fake debugserver to be shutting down during teardown. def shutdown_fakeserver(): fakeserver.close() self.addTearDownHook(shutdown_fakeserver) # Wait until we receive the server ready message before continuing. fakeserver.expect_exact('Listening on localhost:12345') # Connect to the fake server.... self.runCmd("process connect connect://localhost:12345") if __name__ == '__main__': import atexit lldb.SBDebugger.Initialize() atexit.register(lambda: lldb.SBDebugger.Terminate()) unittest2.main() Test case test_connect_remote() has been passing consistently for some times. Let's remove the @expectedFailure marker from it. git-svn-id: b33bab8abb5b18c12ee100cd7761ab452d00b2b0@133294 91177308-0d34-0410-b5e6-96231b3b80d8
""" Test lldb 'process connect' command. """ import os import unittest2 import lldb import pexpect from lldbtest import * class ConnectRemoteTestCase(TestBase): mydir = "connect_remote" def test_connect_remote(self): """Test "process connect connect:://localhost:12345".""" # First, we'll start a fake debugserver (a simple echo server). fakeserver = pexpect.spawn('./EchoServer.py') # Turn on logging for what the child sends back. if self.TraceOn(): fakeserver.logfile_read = sys.stdout # Schedule the fake debugserver to be shutting down during teardown. def shutdown_fakeserver(): fakeserver.close() self.addTearDownHook(shutdown_fakeserver) # Wait until we receive the server ready message before continuing. fakeserver.expect_exact('Listening on localhost:12345') # Connect to the fake server.... self.runCmd("process connect connect://localhost:12345") if __name__ == '__main__': import atexit lldb.SBDebugger.Initialize() atexit.register(lambda: lldb.SBDebugger.Terminate()) unittest2.main()
<commit_before>""" Test lldb 'process connect' command. """ import os import unittest2 import lldb import pexpect from lldbtest import * class ConnectRemoteTestCase(TestBase): mydir = "connect_remote" @unittest2.expectedFailure def test_connect_remote(self): """Test "process connect connect:://localhost:12345".""" # First, we'll start a fake debugserver (a simple echo server). fakeserver = pexpect.spawn('./EchoServer.py') # Turn on logging for what the child sends back. if self.TraceOn(): fakeserver.logfile_read = sys.stdout # Schedule the fake debugserver to be shutting down during teardown. def shutdown_fakeserver(): fakeserver.close() self.addTearDownHook(shutdown_fakeserver) # Wait until we receive the server ready message before continuing. fakeserver.expect_exact('Listening on localhost:12345') # Connect to the fake server.... self.runCmd("process connect connect://localhost:12345") if __name__ == '__main__': import atexit lldb.SBDebugger.Initialize() atexit.register(lambda: lldb.SBDebugger.Terminate()) unittest2.main() <commit_msg>Test case test_connect_remote() has been passing consistently for some times. Let's remove the @expectedFailure marker from it. git-svn-id: b33bab8abb5b18c12ee100cd7761ab452d00b2b0@133294 91177308-0d34-0410-b5e6-96231b3b80d8<commit_after>
""" Test lldb 'process connect' command. """ import os import unittest2 import lldb import pexpect from lldbtest import * class ConnectRemoteTestCase(TestBase): mydir = "connect_remote" def test_connect_remote(self): """Test "process connect connect:://localhost:12345".""" # First, we'll start a fake debugserver (a simple echo server). fakeserver = pexpect.spawn('./EchoServer.py') # Turn on logging for what the child sends back. if self.TraceOn(): fakeserver.logfile_read = sys.stdout # Schedule the fake debugserver to be shutting down during teardown. def shutdown_fakeserver(): fakeserver.close() self.addTearDownHook(shutdown_fakeserver) # Wait until we receive the server ready message before continuing. fakeserver.expect_exact('Listening on localhost:12345') # Connect to the fake server.... self.runCmd("process connect connect://localhost:12345") if __name__ == '__main__': import atexit lldb.SBDebugger.Initialize() atexit.register(lambda: lldb.SBDebugger.Terminate()) unittest2.main()
""" Test lldb 'process connect' command. """ import os import unittest2 import lldb import pexpect from lldbtest import * class ConnectRemoteTestCase(TestBase): mydir = "connect_remote" @unittest2.expectedFailure def test_connect_remote(self): """Test "process connect connect:://localhost:12345".""" # First, we'll start a fake debugserver (a simple echo server). fakeserver = pexpect.spawn('./EchoServer.py') # Turn on logging for what the child sends back. if self.TraceOn(): fakeserver.logfile_read = sys.stdout # Schedule the fake debugserver to be shutting down during teardown. def shutdown_fakeserver(): fakeserver.close() self.addTearDownHook(shutdown_fakeserver) # Wait until we receive the server ready message before continuing. fakeserver.expect_exact('Listening on localhost:12345') # Connect to the fake server.... self.runCmd("process connect connect://localhost:12345") if __name__ == '__main__': import atexit lldb.SBDebugger.Initialize() atexit.register(lambda: lldb.SBDebugger.Terminate()) unittest2.main() Test case test_connect_remote() has been passing consistently for some times. Let's remove the @expectedFailure marker from it. git-svn-id: b33bab8abb5b18c12ee100cd7761ab452d00b2b0@133294 91177308-0d34-0410-b5e6-96231b3b80d8""" Test lldb 'process connect' command. """ import os import unittest2 import lldb import pexpect from lldbtest import * class ConnectRemoteTestCase(TestBase): mydir = "connect_remote" def test_connect_remote(self): """Test "process connect connect:://localhost:12345".""" # First, we'll start a fake debugserver (a simple echo server). fakeserver = pexpect.spawn('./EchoServer.py') # Turn on logging for what the child sends back. if self.TraceOn(): fakeserver.logfile_read = sys.stdout # Schedule the fake debugserver to be shutting down during teardown. def shutdown_fakeserver(): fakeserver.close() self.addTearDownHook(shutdown_fakeserver) # Wait until we receive the server ready message before continuing. fakeserver.expect_exact('Listening on localhost:12345') # Connect to the fake server.... self.runCmd("process connect connect://localhost:12345") if __name__ == '__main__': import atexit lldb.SBDebugger.Initialize() atexit.register(lambda: lldb.SBDebugger.Terminate()) unittest2.main()
<commit_before>""" Test lldb 'process connect' command. """ import os import unittest2 import lldb import pexpect from lldbtest import * class ConnectRemoteTestCase(TestBase): mydir = "connect_remote" @unittest2.expectedFailure def test_connect_remote(self): """Test "process connect connect:://localhost:12345".""" # First, we'll start a fake debugserver (a simple echo server). fakeserver = pexpect.spawn('./EchoServer.py') # Turn on logging for what the child sends back. if self.TraceOn(): fakeserver.logfile_read = sys.stdout # Schedule the fake debugserver to be shutting down during teardown. def shutdown_fakeserver(): fakeserver.close() self.addTearDownHook(shutdown_fakeserver) # Wait until we receive the server ready message before continuing. fakeserver.expect_exact('Listening on localhost:12345') # Connect to the fake server.... self.runCmd("process connect connect://localhost:12345") if __name__ == '__main__': import atexit lldb.SBDebugger.Initialize() atexit.register(lambda: lldb.SBDebugger.Terminate()) unittest2.main() <commit_msg>Test case test_connect_remote() has been passing consistently for some times. Let's remove the @expectedFailure marker from it. git-svn-id: b33bab8abb5b18c12ee100cd7761ab452d00b2b0@133294 91177308-0d34-0410-b5e6-96231b3b80d8<commit_after>""" Test lldb 'process connect' command. """ import os import unittest2 import lldb import pexpect from lldbtest import * class ConnectRemoteTestCase(TestBase): mydir = "connect_remote" def test_connect_remote(self): """Test "process connect connect:://localhost:12345".""" # First, we'll start a fake debugserver (a simple echo server). fakeserver = pexpect.spawn('./EchoServer.py') # Turn on logging for what the child sends back. if self.TraceOn(): fakeserver.logfile_read = sys.stdout # Schedule the fake debugserver to be shutting down during teardown. def shutdown_fakeserver(): fakeserver.close() self.addTearDownHook(shutdown_fakeserver) # Wait until we receive the server ready message before continuing. fakeserver.expect_exact('Listening on localhost:12345') # Connect to the fake server.... self.runCmd("process connect connect://localhost:12345") if __name__ == '__main__': import atexit lldb.SBDebugger.Initialize() atexit.register(lambda: lldb.SBDebugger.Terminate()) unittest2.main()
642908032012baf200ab227803982730c6d4b083
stdnum/ca/__init__.py
stdnum/ca/__init__.py
# __init__.py - collection of Canadian numbers # coding: utf-8 # # Copyright (C) 2017 Arthur de Jong # # This library is free software; you can redistribute it and/or # modify it under the terms of the GNU Lesser General Public # License as published by the Free Software Foundation; either # version 2.1 of the License, or (at your option) any later version. # # This library is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # Lesser General Public License for more details. # # You should have received a copy of the GNU Lesser General Public # License along with this library; if not, write to the Free Software # Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA # 02110-1301 USA """Collection of Canadian numbers."""
# __init__.py - collection of Canadian numbers # coding: utf-8 # # Copyright (C) 2017 Arthur de Jong # # This library is free software; you can redistribute it and/or # modify it under the terms of the GNU Lesser General Public # License as published by the Free Software Foundation; either # version 2.1 of the License, or (at your option) any later version. # # This library is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # Lesser General Public License for more details. # # You should have received a copy of the GNU Lesser General Public # License along with this library; if not, write to the Free Software # Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA # 02110-1301 USA """Collection of Canadian numbers.""" from stdnum.ca import bn as vat # noqa: F401
Add missing vat alias for Canada
Add missing vat alias for Canada
Python
lgpl-2.1
arthurdejong/python-stdnum,arthurdejong/python-stdnum,arthurdejong/python-stdnum
# __init__.py - collection of Canadian numbers # coding: utf-8 # # Copyright (C) 2017 Arthur de Jong # # This library is free software; you can redistribute it and/or # modify it under the terms of the GNU Lesser General Public # License as published by the Free Software Foundation; either # version 2.1 of the License, or (at your option) any later version. # # This library is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # Lesser General Public License for more details. # # You should have received a copy of the GNU Lesser General Public # License along with this library; if not, write to the Free Software # Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA # 02110-1301 USA """Collection of Canadian numbers.""" Add missing vat alias for Canada
# __init__.py - collection of Canadian numbers # coding: utf-8 # # Copyright (C) 2017 Arthur de Jong # # This library is free software; you can redistribute it and/or # modify it under the terms of the GNU Lesser General Public # License as published by the Free Software Foundation; either # version 2.1 of the License, or (at your option) any later version. # # This library is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # Lesser General Public License for more details. # # You should have received a copy of the GNU Lesser General Public # License along with this library; if not, write to the Free Software # Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA # 02110-1301 USA """Collection of Canadian numbers.""" from stdnum.ca import bn as vat # noqa: F401
<commit_before># __init__.py - collection of Canadian numbers # coding: utf-8 # # Copyright (C) 2017 Arthur de Jong # # This library is free software; you can redistribute it and/or # modify it under the terms of the GNU Lesser General Public # License as published by the Free Software Foundation; either # version 2.1 of the License, or (at your option) any later version. # # This library is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # Lesser General Public License for more details. # # You should have received a copy of the GNU Lesser General Public # License along with this library; if not, write to the Free Software # Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA # 02110-1301 USA """Collection of Canadian numbers.""" <commit_msg>Add missing vat alias for Canada<commit_after>
# __init__.py - collection of Canadian numbers # coding: utf-8 # # Copyright (C) 2017 Arthur de Jong # # This library is free software; you can redistribute it and/or # modify it under the terms of the GNU Lesser General Public # License as published by the Free Software Foundation; either # version 2.1 of the License, or (at your option) any later version. # # This library is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # Lesser General Public License for more details. # # You should have received a copy of the GNU Lesser General Public # License along with this library; if not, write to the Free Software # Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA # 02110-1301 USA """Collection of Canadian numbers.""" from stdnum.ca import bn as vat # noqa: F401
# __init__.py - collection of Canadian numbers # coding: utf-8 # # Copyright (C) 2017 Arthur de Jong # # This library is free software; you can redistribute it and/or # modify it under the terms of the GNU Lesser General Public # License as published by the Free Software Foundation; either # version 2.1 of the License, or (at your option) any later version. # # This library is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # Lesser General Public License for more details. # # You should have received a copy of the GNU Lesser General Public # License along with this library; if not, write to the Free Software # Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA # 02110-1301 USA """Collection of Canadian numbers.""" Add missing vat alias for Canada# __init__.py - collection of Canadian numbers # coding: utf-8 # # Copyright (C) 2017 Arthur de Jong # # This library is free software; you can redistribute it and/or # modify it under the terms of the GNU Lesser General Public # License as published by the Free Software Foundation; either # version 2.1 of the License, or (at your option) any later version. # # This library is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # Lesser General Public License for more details. # # You should have received a copy of the GNU Lesser General Public # License along with this library; if not, write to the Free Software # Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA # 02110-1301 USA """Collection of Canadian numbers.""" from stdnum.ca import bn as vat # noqa: F401
<commit_before># __init__.py - collection of Canadian numbers # coding: utf-8 # # Copyright (C) 2017 Arthur de Jong # # This library is free software; you can redistribute it and/or # modify it under the terms of the GNU Lesser General Public # License as published by the Free Software Foundation; either # version 2.1 of the License, or (at your option) any later version. # # This library is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # Lesser General Public License for more details. # # You should have received a copy of the GNU Lesser General Public # License along with this library; if not, write to the Free Software # Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA # 02110-1301 USA """Collection of Canadian numbers.""" <commit_msg>Add missing vat alias for Canada<commit_after># __init__.py - collection of Canadian numbers # coding: utf-8 # # Copyright (C) 2017 Arthur de Jong # # This library is free software; you can redistribute it and/or # modify it under the terms of the GNU Lesser General Public # License as published by the Free Software Foundation; either # version 2.1 of the License, or (at your option) any later version. # # This library is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # Lesser General Public License for more details. # # You should have received a copy of the GNU Lesser General Public # License along with this library; if not, write to the Free Software # Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA # 02110-1301 USA """Collection of Canadian numbers.""" from stdnum.ca import bn as vat # noqa: F401
c2859bd8da741862ee01a276a1350fb4a5931dbc
data_access.py
data_access.py
#!/usr/bin/env python import sys import mysql.connector def insert(): cursor = connection.cursor() try: cursor.execute("drop table employees") except: pass cursor.execute("create table employees (id integer primary key, name text)") cursor.close() print("Inserting employees...") for n in xrange(0, 10000): cursor = connection.cursor() cursor.execute("insert into employees (id, name) values (%d, 'Employee_%d')" % (n, n)) connection.commit() cursor.close() def select(): print("Selecting employees...") while True: cursor = connection.cursor() cursor.execute("select * from employees where name like '%1417773'") for row in cursor: pass cursor.close() connection = mysql.connector.connect(host='localhost', database='test') if "insert" in sys.argv: while True: insert() elif "insert_once" in sys.argv: insert() elif "select" in sys.argv: select() else: print("USAGE: data_access.py <insert|insert_once|select>") connection.close()
#!/usr/bin/env python from random import randint import sys import mysql.connector NUM_EMPLOYEES = 10000 def insert(): cursor = connection.cursor() try: cursor.execute("drop table employees") except: pass cursor.execute("create table employees (id integer primary key, name text)") cursor.close() print("Inserting employees...") for n in xrange(0, NUM_EMPLOYEES): cursor = connection.cursor() cursor.execute("insert into employees (id, name) values (%d, 'Employee_%d')" % (n, n)) connection.commit() cursor.close() def select(): print("Selecting employees...") while True: cursor = connection.cursor() cursor.execute("select * from employees where name like '%%%d'" % randint(0, NUM_EMPLOYEES)) for row in cursor: pass cursor.close() connection = mysql.connector.connect(host='localhost', database='test') if "insert" in sys.argv: while True: insert() elif "insert_once" in sys.argv: insert() elif "select" in sys.argv: select() else: print("USAGE: data_access.py <insert|insert_once|select>") connection.close()
Change data access script to issue SELECTs that actually return a value
Change data access script to issue SELECTs that actually return a value This makes the part about tracing the SQL statements and tracing the number of rows returned a little more interesting.
Python
mit
goldshtn/linux-tracing-workshop,goldshtn/linux-tracing-workshop,goldshtn/linux-tracing-workshop,goldshtn/linux-tracing-workshop,goldshtn/linux-tracing-workshop,goldshtn/linux-tracing-workshop,goldshtn/linux-tracing-workshop,goldshtn/linux-tracing-workshop
#!/usr/bin/env python import sys import mysql.connector def insert(): cursor = connection.cursor() try: cursor.execute("drop table employees") except: pass cursor.execute("create table employees (id integer primary key, name text)") cursor.close() print("Inserting employees...") for n in xrange(0, 10000): cursor = connection.cursor() cursor.execute("insert into employees (id, name) values (%d, 'Employee_%d')" % (n, n)) connection.commit() cursor.close() def select(): print("Selecting employees...") while True: cursor = connection.cursor() cursor.execute("select * from employees where name like '%1417773'") for row in cursor: pass cursor.close() connection = mysql.connector.connect(host='localhost', database='test') if "insert" in sys.argv: while True: insert() elif "insert_once" in sys.argv: insert() elif "select" in sys.argv: select() else: print("USAGE: data_access.py <insert|insert_once|select>") connection.close() Change data access script to issue SELECTs that actually return a value This makes the part about tracing the SQL statements and tracing the number of rows returned a little more interesting.
#!/usr/bin/env python from random import randint import sys import mysql.connector NUM_EMPLOYEES = 10000 def insert(): cursor = connection.cursor() try: cursor.execute("drop table employees") except: pass cursor.execute("create table employees (id integer primary key, name text)") cursor.close() print("Inserting employees...") for n in xrange(0, NUM_EMPLOYEES): cursor = connection.cursor() cursor.execute("insert into employees (id, name) values (%d, 'Employee_%d')" % (n, n)) connection.commit() cursor.close() def select(): print("Selecting employees...") while True: cursor = connection.cursor() cursor.execute("select * from employees where name like '%%%d'" % randint(0, NUM_EMPLOYEES)) for row in cursor: pass cursor.close() connection = mysql.connector.connect(host='localhost', database='test') if "insert" in sys.argv: while True: insert() elif "insert_once" in sys.argv: insert() elif "select" in sys.argv: select() else: print("USAGE: data_access.py <insert|insert_once|select>") connection.close()
<commit_before>#!/usr/bin/env python import sys import mysql.connector def insert(): cursor = connection.cursor() try: cursor.execute("drop table employees") except: pass cursor.execute("create table employees (id integer primary key, name text)") cursor.close() print("Inserting employees...") for n in xrange(0, 10000): cursor = connection.cursor() cursor.execute("insert into employees (id, name) values (%d, 'Employee_%d')" % (n, n)) connection.commit() cursor.close() def select(): print("Selecting employees...") while True: cursor = connection.cursor() cursor.execute("select * from employees where name like '%1417773'") for row in cursor: pass cursor.close() connection = mysql.connector.connect(host='localhost', database='test') if "insert" in sys.argv: while True: insert() elif "insert_once" in sys.argv: insert() elif "select" in sys.argv: select() else: print("USAGE: data_access.py <insert|insert_once|select>") connection.close() <commit_msg>Change data access script to issue SELECTs that actually return a value This makes the part about tracing the SQL statements and tracing the number of rows returned a little more interesting.<commit_after>
#!/usr/bin/env python from random import randint import sys import mysql.connector NUM_EMPLOYEES = 10000 def insert(): cursor = connection.cursor() try: cursor.execute("drop table employees") except: pass cursor.execute("create table employees (id integer primary key, name text)") cursor.close() print("Inserting employees...") for n in xrange(0, NUM_EMPLOYEES): cursor = connection.cursor() cursor.execute("insert into employees (id, name) values (%d, 'Employee_%d')" % (n, n)) connection.commit() cursor.close() def select(): print("Selecting employees...") while True: cursor = connection.cursor() cursor.execute("select * from employees where name like '%%%d'" % randint(0, NUM_EMPLOYEES)) for row in cursor: pass cursor.close() connection = mysql.connector.connect(host='localhost', database='test') if "insert" in sys.argv: while True: insert() elif "insert_once" in sys.argv: insert() elif "select" in sys.argv: select() else: print("USAGE: data_access.py <insert|insert_once|select>") connection.close()
#!/usr/bin/env python import sys import mysql.connector def insert(): cursor = connection.cursor() try: cursor.execute("drop table employees") except: pass cursor.execute("create table employees (id integer primary key, name text)") cursor.close() print("Inserting employees...") for n in xrange(0, 10000): cursor = connection.cursor() cursor.execute("insert into employees (id, name) values (%d, 'Employee_%d')" % (n, n)) connection.commit() cursor.close() def select(): print("Selecting employees...") while True: cursor = connection.cursor() cursor.execute("select * from employees where name like '%1417773'") for row in cursor: pass cursor.close() connection = mysql.connector.connect(host='localhost', database='test') if "insert" in sys.argv: while True: insert() elif "insert_once" in sys.argv: insert() elif "select" in sys.argv: select() else: print("USAGE: data_access.py <insert|insert_once|select>") connection.close() Change data access script to issue SELECTs that actually return a value This makes the part about tracing the SQL statements and tracing the number of rows returned a little more interesting.#!/usr/bin/env python from random import randint import sys import mysql.connector NUM_EMPLOYEES = 10000 def insert(): cursor = connection.cursor() try: cursor.execute("drop table employees") except: pass cursor.execute("create table employees (id integer primary key, name text)") cursor.close() print("Inserting employees...") for n in xrange(0, NUM_EMPLOYEES): cursor = connection.cursor() cursor.execute("insert into employees (id, name) values (%d, 'Employee_%d')" % (n, n)) connection.commit() cursor.close() def select(): print("Selecting employees...") while True: cursor = connection.cursor() cursor.execute("select * from employees where name like '%%%d'" % randint(0, NUM_EMPLOYEES)) for row in cursor: pass cursor.close() connection = mysql.connector.connect(host='localhost', database='test') if "insert" in sys.argv: while True: insert() elif "insert_once" in sys.argv: insert() elif "select" in sys.argv: select() else: print("USAGE: data_access.py <insert|insert_once|select>") connection.close()
<commit_before>#!/usr/bin/env python import sys import mysql.connector def insert(): cursor = connection.cursor() try: cursor.execute("drop table employees") except: pass cursor.execute("create table employees (id integer primary key, name text)") cursor.close() print("Inserting employees...") for n in xrange(0, 10000): cursor = connection.cursor() cursor.execute("insert into employees (id, name) values (%d, 'Employee_%d')" % (n, n)) connection.commit() cursor.close() def select(): print("Selecting employees...") while True: cursor = connection.cursor() cursor.execute("select * from employees where name like '%1417773'") for row in cursor: pass cursor.close() connection = mysql.connector.connect(host='localhost', database='test') if "insert" in sys.argv: while True: insert() elif "insert_once" in sys.argv: insert() elif "select" in sys.argv: select() else: print("USAGE: data_access.py <insert|insert_once|select>") connection.close() <commit_msg>Change data access script to issue SELECTs that actually return a value This makes the part about tracing the SQL statements and tracing the number of rows returned a little more interesting.<commit_after>#!/usr/bin/env python from random import randint import sys import mysql.connector NUM_EMPLOYEES = 10000 def insert(): cursor = connection.cursor() try: cursor.execute("drop table employees") except: pass cursor.execute("create table employees (id integer primary key, name text)") cursor.close() print("Inserting employees...") for n in xrange(0, NUM_EMPLOYEES): cursor = connection.cursor() cursor.execute("insert into employees (id, name) values (%d, 'Employee_%d')" % (n, n)) connection.commit() cursor.close() def select(): print("Selecting employees...") while True: cursor = connection.cursor() cursor.execute("select * from employees where name like '%%%d'" % randint(0, NUM_EMPLOYEES)) for row in cursor: pass cursor.close() connection = mysql.connector.connect(host='localhost', database='test') if "insert" in sys.argv: while True: insert() elif "insert_once" in sys.argv: insert() elif "select" in sys.argv: select() else: print("USAGE: data_access.py <insert|insert_once|select>") connection.close()
b0d8280927bdd33cfb16da2782ca54100a5ece09
py/dynesty/__init__.py
py/dynesty/__init__.py
#!/usr/bin/env python # -*- coding: utf-8 -*- """ dynesty is nested sampling package. The main functionality of dynesty is performed by the dynesty.NestedSampler and dynesty.DynamicNestedSampler classes """ from .dynesty import NestedSampler, DynamicNestedSampler from . import bounding from . import utils __version__ = "1.2"
#!/usr/bin/env python # -*- coding: utf-8 -*- """ dynesty is nested sampling package. The main functionality of dynesty is performed by the dynesty.NestedSampler and dynesty.DynamicNestedSampler classes """ from .dynesty import NestedSampler, DynamicNestedSampler from . import bounding from . import utils __version__ = "1.2.0"
Change the version to 1.2.0
Change the version to 1.2.0
Python
mit
joshspeagle/dynesty
#!/usr/bin/env python # -*- coding: utf-8 -*- """ dynesty is nested sampling package. The main functionality of dynesty is performed by the dynesty.NestedSampler and dynesty.DynamicNestedSampler classes """ from .dynesty import NestedSampler, DynamicNestedSampler from . import bounding from . import utils __version__ = "1.2" Change the version to 1.2.0
#!/usr/bin/env python # -*- coding: utf-8 -*- """ dynesty is nested sampling package. The main functionality of dynesty is performed by the dynesty.NestedSampler and dynesty.DynamicNestedSampler classes """ from .dynesty import NestedSampler, DynamicNestedSampler from . import bounding from . import utils __version__ = "1.2.0"
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- """ dynesty is nested sampling package. The main functionality of dynesty is performed by the dynesty.NestedSampler and dynesty.DynamicNestedSampler classes """ from .dynesty import NestedSampler, DynamicNestedSampler from . import bounding from . import utils __version__ = "1.2" <commit_msg>Change the version to 1.2.0<commit_after>
#!/usr/bin/env python # -*- coding: utf-8 -*- """ dynesty is nested sampling package. The main functionality of dynesty is performed by the dynesty.NestedSampler and dynesty.DynamicNestedSampler classes """ from .dynesty import NestedSampler, DynamicNestedSampler from . import bounding from . import utils __version__ = "1.2.0"
#!/usr/bin/env python # -*- coding: utf-8 -*- """ dynesty is nested sampling package. The main functionality of dynesty is performed by the dynesty.NestedSampler and dynesty.DynamicNestedSampler classes """ from .dynesty import NestedSampler, DynamicNestedSampler from . import bounding from . import utils __version__ = "1.2" Change the version to 1.2.0#!/usr/bin/env python # -*- coding: utf-8 -*- """ dynesty is nested sampling package. The main functionality of dynesty is performed by the dynesty.NestedSampler and dynesty.DynamicNestedSampler classes """ from .dynesty import NestedSampler, DynamicNestedSampler from . import bounding from . import utils __version__ = "1.2.0"
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- """ dynesty is nested sampling package. The main functionality of dynesty is performed by the dynesty.NestedSampler and dynesty.DynamicNestedSampler classes """ from .dynesty import NestedSampler, DynamicNestedSampler from . import bounding from . import utils __version__ = "1.2" <commit_msg>Change the version to 1.2.0<commit_after>#!/usr/bin/env python # -*- coding: utf-8 -*- """ dynesty is nested sampling package. The main functionality of dynesty is performed by the dynesty.NestedSampler and dynesty.DynamicNestedSampler classes """ from .dynesty import NestedSampler, DynamicNestedSampler from . import bounding from . import utils __version__ = "1.2.0"
807d7efe7de00950df675e78249dcada298b6cd1
systemrdl/__init__.py
systemrdl/__init__.py
from .__about__ import __version__ from .compiler import RDLCompiler from .walker import RDLListener, RDLWalker from .messages import RDLCompileError
from .__about__ import __version__ from .compiler import RDLCompiler from .walker import RDLListener, RDLWalker from .messages import RDLCompileError from .node import AddressableNode, VectorNode, SignalNode from .node import FieldNode, RegNode, RegfileNode, AddrmapNode, MemNode from .component import AddressableComponent, VectorComponent, Signal from .component import Field, Reg, Regfile, Addrmap, Mem
Bring forward more contents into top namespace
Bring forward more contents into top namespace
Python
mit
SystemRDL/systemrdl-compiler,SystemRDL/systemrdl-compiler,SystemRDL/systemrdl-compiler,SystemRDL/systemrdl-compiler
from .__about__ import __version__ from .compiler import RDLCompiler from .walker import RDLListener, RDLWalker from .messages import RDLCompileError Bring forward more contents into top namespace
from .__about__ import __version__ from .compiler import RDLCompiler from .walker import RDLListener, RDLWalker from .messages import RDLCompileError from .node import AddressableNode, VectorNode, SignalNode from .node import FieldNode, RegNode, RegfileNode, AddrmapNode, MemNode from .component import AddressableComponent, VectorComponent, Signal from .component import Field, Reg, Regfile, Addrmap, Mem
<commit_before>from .__about__ import __version__ from .compiler import RDLCompiler from .walker import RDLListener, RDLWalker from .messages import RDLCompileError <commit_msg>Bring forward more contents into top namespace<commit_after>
from .__about__ import __version__ from .compiler import RDLCompiler from .walker import RDLListener, RDLWalker from .messages import RDLCompileError from .node import AddressableNode, VectorNode, SignalNode from .node import FieldNode, RegNode, RegfileNode, AddrmapNode, MemNode from .component import AddressableComponent, VectorComponent, Signal from .component import Field, Reg, Regfile, Addrmap, Mem
from .__about__ import __version__ from .compiler import RDLCompiler from .walker import RDLListener, RDLWalker from .messages import RDLCompileError Bring forward more contents into top namespacefrom .__about__ import __version__ from .compiler import RDLCompiler from .walker import RDLListener, RDLWalker from .messages import RDLCompileError from .node import AddressableNode, VectorNode, SignalNode from .node import FieldNode, RegNode, RegfileNode, AddrmapNode, MemNode from .component import AddressableComponent, VectorComponent, Signal from .component import Field, Reg, Regfile, Addrmap, Mem
<commit_before>from .__about__ import __version__ from .compiler import RDLCompiler from .walker import RDLListener, RDLWalker from .messages import RDLCompileError <commit_msg>Bring forward more contents into top namespace<commit_after>from .__about__ import __version__ from .compiler import RDLCompiler from .walker import RDLListener, RDLWalker from .messages import RDLCompileError from .node import AddressableNode, VectorNode, SignalNode from .node import FieldNode, RegNode, RegfileNode, AddrmapNode, MemNode from .component import AddressableComponent, VectorComponent, Signal from .component import Field, Reg, Regfile, Addrmap, Mem
ccb90932cf967190029b3ce9494a1fd9e6cb889a
gaphor/UML/classes/tests/test_propertypages.py
gaphor/UML/classes/tests/test_propertypages.py
from gi.repository import Gtk from gaphor import UML from gaphor.UML.classes import ClassItem from gaphor.UML.classes.classespropertypages import ClassAttributes class TestClassPropertyPages: def test_attribute_editing(self, case): class_item = case.create(ClassItem, UML.Class) model = ClassAttributes(class_item, (str, bool, object)) model.append([None, False, None]) path = Gtk.TreePath.new_first() iter = model.get_iter(path) model.update(iter, col=0, value="attr") assert model[iter][-1] is class_item.subject.ownedAttribute[0]
from gi.repository import Gtk from gaphor import UML from gaphor.UML.classes import ClassItem, EnumerationItem from gaphor.UML.classes.classespropertypages import ( ClassAttributes, ClassEnumerationLiterals, ) def test_attribute_editing(case): class_item = case.create(ClassItem, UML.Class) model = ClassAttributes(class_item, (str, bool, object)) model.append([None, False, None]) path = Gtk.TreePath.new_first() iter = model.get_iter(path) model.update(iter, col=0, value="attr") assert model[iter][-1] is class_item.subject.ownedAttribute[0] def test_enumeration_editing(case): enum_item = case.create(EnumerationItem, UML.Enumeration) model = ClassEnumerationLiterals(enum_item, (str, object)) model.append([None, None]) path = Gtk.TreePath.new_first() iter = model.get_iter(path) model.update(iter, col=0, value="enum") assert model[iter][-1] is enum_item.subject.ownedLiteral[0]
Add test for enumeration editing
Add test for enumeration editing Signed-off-by: Dan Yeaw <2591e5f46f28d303f9dc027d475a5c60d8dea17a@yeaw.me>
Python
lgpl-2.1
amolenaar/gaphor,amolenaar/gaphor
from gi.repository import Gtk from gaphor import UML from gaphor.UML.classes import ClassItem from gaphor.UML.classes.classespropertypages import ClassAttributes class TestClassPropertyPages: def test_attribute_editing(self, case): class_item = case.create(ClassItem, UML.Class) model = ClassAttributes(class_item, (str, bool, object)) model.append([None, False, None]) path = Gtk.TreePath.new_first() iter = model.get_iter(path) model.update(iter, col=0, value="attr") assert model[iter][-1] is class_item.subject.ownedAttribute[0] Add test for enumeration editing Signed-off-by: Dan Yeaw <2591e5f46f28d303f9dc027d475a5c60d8dea17a@yeaw.me>
from gi.repository import Gtk from gaphor import UML from gaphor.UML.classes import ClassItem, EnumerationItem from gaphor.UML.classes.classespropertypages import ( ClassAttributes, ClassEnumerationLiterals, ) def test_attribute_editing(case): class_item = case.create(ClassItem, UML.Class) model = ClassAttributes(class_item, (str, bool, object)) model.append([None, False, None]) path = Gtk.TreePath.new_first() iter = model.get_iter(path) model.update(iter, col=0, value="attr") assert model[iter][-1] is class_item.subject.ownedAttribute[0] def test_enumeration_editing(case): enum_item = case.create(EnumerationItem, UML.Enumeration) model = ClassEnumerationLiterals(enum_item, (str, object)) model.append([None, None]) path = Gtk.TreePath.new_first() iter = model.get_iter(path) model.update(iter, col=0, value="enum") assert model[iter][-1] is enum_item.subject.ownedLiteral[0]
<commit_before>from gi.repository import Gtk from gaphor import UML from gaphor.UML.classes import ClassItem from gaphor.UML.classes.classespropertypages import ClassAttributes class TestClassPropertyPages: def test_attribute_editing(self, case): class_item = case.create(ClassItem, UML.Class) model = ClassAttributes(class_item, (str, bool, object)) model.append([None, False, None]) path = Gtk.TreePath.new_first() iter = model.get_iter(path) model.update(iter, col=0, value="attr") assert model[iter][-1] is class_item.subject.ownedAttribute[0] <commit_msg>Add test for enumeration editing Signed-off-by: Dan Yeaw <2591e5f46f28d303f9dc027d475a5c60d8dea17a@yeaw.me><commit_after>
from gi.repository import Gtk from gaphor import UML from gaphor.UML.classes import ClassItem, EnumerationItem from gaphor.UML.classes.classespropertypages import ( ClassAttributes, ClassEnumerationLiterals, ) def test_attribute_editing(case): class_item = case.create(ClassItem, UML.Class) model = ClassAttributes(class_item, (str, bool, object)) model.append([None, False, None]) path = Gtk.TreePath.new_first() iter = model.get_iter(path) model.update(iter, col=0, value="attr") assert model[iter][-1] is class_item.subject.ownedAttribute[0] def test_enumeration_editing(case): enum_item = case.create(EnumerationItem, UML.Enumeration) model = ClassEnumerationLiterals(enum_item, (str, object)) model.append([None, None]) path = Gtk.TreePath.new_first() iter = model.get_iter(path) model.update(iter, col=0, value="enum") assert model[iter][-1] is enum_item.subject.ownedLiteral[0]
from gi.repository import Gtk from gaphor import UML from gaphor.UML.classes import ClassItem from gaphor.UML.classes.classespropertypages import ClassAttributes class TestClassPropertyPages: def test_attribute_editing(self, case): class_item = case.create(ClassItem, UML.Class) model = ClassAttributes(class_item, (str, bool, object)) model.append([None, False, None]) path = Gtk.TreePath.new_first() iter = model.get_iter(path) model.update(iter, col=0, value="attr") assert model[iter][-1] is class_item.subject.ownedAttribute[0] Add test for enumeration editing Signed-off-by: Dan Yeaw <2591e5f46f28d303f9dc027d475a5c60d8dea17a@yeaw.me>from gi.repository import Gtk from gaphor import UML from gaphor.UML.classes import ClassItem, EnumerationItem from gaphor.UML.classes.classespropertypages import ( ClassAttributes, ClassEnumerationLiterals, ) def test_attribute_editing(case): class_item = case.create(ClassItem, UML.Class) model = ClassAttributes(class_item, (str, bool, object)) model.append([None, False, None]) path = Gtk.TreePath.new_first() iter = model.get_iter(path) model.update(iter, col=0, value="attr") assert model[iter][-1] is class_item.subject.ownedAttribute[0] def test_enumeration_editing(case): enum_item = case.create(EnumerationItem, UML.Enumeration) model = ClassEnumerationLiterals(enum_item, (str, object)) model.append([None, None]) path = Gtk.TreePath.new_first() iter = model.get_iter(path) model.update(iter, col=0, value="enum") assert model[iter][-1] is enum_item.subject.ownedLiteral[0]
<commit_before>from gi.repository import Gtk from gaphor import UML from gaphor.UML.classes import ClassItem from gaphor.UML.classes.classespropertypages import ClassAttributes class TestClassPropertyPages: def test_attribute_editing(self, case): class_item = case.create(ClassItem, UML.Class) model = ClassAttributes(class_item, (str, bool, object)) model.append([None, False, None]) path = Gtk.TreePath.new_first() iter = model.get_iter(path) model.update(iter, col=0, value="attr") assert model[iter][-1] is class_item.subject.ownedAttribute[0] <commit_msg>Add test for enumeration editing Signed-off-by: Dan Yeaw <2591e5f46f28d303f9dc027d475a5c60d8dea17a@yeaw.me><commit_after>from gi.repository import Gtk from gaphor import UML from gaphor.UML.classes import ClassItem, EnumerationItem from gaphor.UML.classes.classespropertypages import ( ClassAttributes, ClassEnumerationLiterals, ) def test_attribute_editing(case): class_item = case.create(ClassItem, UML.Class) model = ClassAttributes(class_item, (str, bool, object)) model.append([None, False, None]) path = Gtk.TreePath.new_first() iter = model.get_iter(path) model.update(iter, col=0, value="attr") assert model[iter][-1] is class_item.subject.ownedAttribute[0] def test_enumeration_editing(case): enum_item = case.create(EnumerationItem, UML.Enumeration) model = ClassEnumerationLiterals(enum_item, (str, object)) model.append([None, None]) path = Gtk.TreePath.new_first() iter = model.get_iter(path) model.update(iter, col=0, value="enum") assert model[iter][-1] is enum_item.subject.ownedLiteral[0]
5747284df86016958ab1ae9dcf437b375e79beba
xbob/core/__init__.py
xbob/core/__init__.py
from ._convert import convert from . import log from . import random from . import version from .version import module as __version__ from .version import api as __api_version__ def get_include(): """Returns the directory containing the C/C++ API include directives""" return __import__('pkg_resources').resource_filename(__name__, 'include') # gets sphinx autodoc done right - don't remove it __all__ = [_ for _ in dir() if not _.startswith('_')]
from ._convert import convert from . import log from . import random from . import version from .version import module as __version__ from .version import api as __api_version__ def get_include(): """Returns the directory containing the C/C++ API include directives""" return __import__('pkg_resources').resource_filename(__name__, 'include') def get_config(): """Returns a string containing the configuration information. """ import pkg_resources from .version import externals packages = pkg_resources.require(__name__) this = packages[0] deps = packages[1:] retval = "%s: %s (%s)\n" % (this.key, this.version, this.location) retval += " - c/c++ dependencies:\n" for k in sorted(externals): retval += " - %s: %s\n" % (k, externals[k]) retval += " - python dependencies:\n" for d in deps: retval += " - %s: %s (%s)\n" % (d.key, d.version, d.location) return retval.strip() # gets sphinx autodoc done right - don't remove it __all__ = [_ for _ in dir() if not _.startswith('_')]
Add function to print comprehensive dependence list
Add function to print comprehensive dependence list
Python
bsd-3-clause
tiagofrepereira2012/bob.core,tiagofrepereira2012/bob.core,tiagofrepereira2012/bob.core
from ._convert import convert from . import log from . import random from . import version from .version import module as __version__ from .version import api as __api_version__ def get_include(): """Returns the directory containing the C/C++ API include directives""" return __import__('pkg_resources').resource_filename(__name__, 'include') # gets sphinx autodoc done right - don't remove it __all__ = [_ for _ in dir() if not _.startswith('_')] Add function to print comprehensive dependence list
from ._convert import convert from . import log from . import random from . import version from .version import module as __version__ from .version import api as __api_version__ def get_include(): """Returns the directory containing the C/C++ API include directives""" return __import__('pkg_resources').resource_filename(__name__, 'include') def get_config(): """Returns a string containing the configuration information. """ import pkg_resources from .version import externals packages = pkg_resources.require(__name__) this = packages[0] deps = packages[1:] retval = "%s: %s (%s)\n" % (this.key, this.version, this.location) retval += " - c/c++ dependencies:\n" for k in sorted(externals): retval += " - %s: %s\n" % (k, externals[k]) retval += " - python dependencies:\n" for d in deps: retval += " - %s: %s (%s)\n" % (d.key, d.version, d.location) return retval.strip() # gets sphinx autodoc done right - don't remove it __all__ = [_ for _ in dir() if not _.startswith('_')]
<commit_before>from ._convert import convert from . import log from . import random from . import version from .version import module as __version__ from .version import api as __api_version__ def get_include(): """Returns the directory containing the C/C++ API include directives""" return __import__('pkg_resources').resource_filename(__name__, 'include') # gets sphinx autodoc done right - don't remove it __all__ = [_ for _ in dir() if not _.startswith('_')] <commit_msg>Add function to print comprehensive dependence list<commit_after>
from ._convert import convert from . import log from . import random from . import version from .version import module as __version__ from .version import api as __api_version__ def get_include(): """Returns the directory containing the C/C++ API include directives""" return __import__('pkg_resources').resource_filename(__name__, 'include') def get_config(): """Returns a string containing the configuration information. """ import pkg_resources from .version import externals packages = pkg_resources.require(__name__) this = packages[0] deps = packages[1:] retval = "%s: %s (%s)\n" % (this.key, this.version, this.location) retval += " - c/c++ dependencies:\n" for k in sorted(externals): retval += " - %s: %s\n" % (k, externals[k]) retval += " - python dependencies:\n" for d in deps: retval += " - %s: %s (%s)\n" % (d.key, d.version, d.location) return retval.strip() # gets sphinx autodoc done right - don't remove it __all__ = [_ for _ in dir() if not _.startswith('_')]
from ._convert import convert from . import log from . import random from . import version from .version import module as __version__ from .version import api as __api_version__ def get_include(): """Returns the directory containing the C/C++ API include directives""" return __import__('pkg_resources').resource_filename(__name__, 'include') # gets sphinx autodoc done right - don't remove it __all__ = [_ for _ in dir() if not _.startswith('_')] Add function to print comprehensive dependence listfrom ._convert import convert from . import log from . import random from . import version from .version import module as __version__ from .version import api as __api_version__ def get_include(): """Returns the directory containing the C/C++ API include directives""" return __import__('pkg_resources').resource_filename(__name__, 'include') def get_config(): """Returns a string containing the configuration information. """ import pkg_resources from .version import externals packages = pkg_resources.require(__name__) this = packages[0] deps = packages[1:] retval = "%s: %s (%s)\n" % (this.key, this.version, this.location) retval += " - c/c++ dependencies:\n" for k in sorted(externals): retval += " - %s: %s\n" % (k, externals[k]) retval += " - python dependencies:\n" for d in deps: retval += " - %s: %s (%s)\n" % (d.key, d.version, d.location) return retval.strip() # gets sphinx autodoc done right - don't remove it __all__ = [_ for _ in dir() if not _.startswith('_')]
<commit_before>from ._convert import convert from . import log from . import random from . import version from .version import module as __version__ from .version import api as __api_version__ def get_include(): """Returns the directory containing the C/C++ API include directives""" return __import__('pkg_resources').resource_filename(__name__, 'include') # gets sphinx autodoc done right - don't remove it __all__ = [_ for _ in dir() if not _.startswith('_')] <commit_msg>Add function to print comprehensive dependence list<commit_after>from ._convert import convert from . import log from . import random from . import version from .version import module as __version__ from .version import api as __api_version__ def get_include(): """Returns the directory containing the C/C++ API include directives""" return __import__('pkg_resources').resource_filename(__name__, 'include') def get_config(): """Returns a string containing the configuration information. """ import pkg_resources from .version import externals packages = pkg_resources.require(__name__) this = packages[0] deps = packages[1:] retval = "%s: %s (%s)\n" % (this.key, this.version, this.location) retval += " - c/c++ dependencies:\n" for k in sorted(externals): retval += " - %s: %s\n" % (k, externals[k]) retval += " - python dependencies:\n" for d in deps: retval += " - %s: %s (%s)\n" % (d.key, d.version, d.location) return retval.strip() # gets sphinx autodoc done right - don't remove it __all__ = [_ for _ in dir() if not _.startswith('_')]
f68e8cb9751a32cc4d8bdc97c6f753395381e1e1
python/dnest4/utils.py
python/dnest4/utils.py
# -*- coding: utf-8 -*- __all__ = ["randh", "wrap"] import numpy as np import numpy.random as rng def randh(): """ Generate from the heavy-tailed distribution. """ return 10.0**(1.5 - 3*np.abs(rng.randn()/np.sqrt(-np.log(rng.rand()))))*rng.randn() def wrap(x, a, b): assert b > a return (x - a)%(b - a) + a
# -*- coding: utf-8 -*- __all__ = ["randh", "wrap"] import numpy as np import numpy.random as rng def randh(N=1): """ Generate from the heavy-tailed distribution. """ if N==1: return 10.0**(1.5 - 3*np.abs(rng.randn()/np.sqrt(-np.log(rng.rand()))))*rng.randn() return 10.0**(1.5 - 3*np.abs(rng.randn(N)/np.sqrt(-np.log(rng.rand(N)))))*rng.randn(N) def wrap(x, a, b): assert b > a return (x - a)%(b - a) + a
Allow N > 1 randhs to be generated
Allow N > 1 randhs to be generated
Python
mit
eggplantbren/DNest4,eggplantbren/DNest4,eggplantbren/DNest4,eggplantbren/DNest4,eggplantbren/DNest4
# -*- coding: utf-8 -*- __all__ = ["randh", "wrap"] import numpy as np import numpy.random as rng def randh(): """ Generate from the heavy-tailed distribution. """ return 10.0**(1.5 - 3*np.abs(rng.randn()/np.sqrt(-np.log(rng.rand()))))*rng.randn() def wrap(x, a, b): assert b > a return (x - a)%(b - a) + a Allow N > 1 randhs to be generated
# -*- coding: utf-8 -*- __all__ = ["randh", "wrap"] import numpy as np import numpy.random as rng def randh(N=1): """ Generate from the heavy-tailed distribution. """ if N==1: return 10.0**(1.5 - 3*np.abs(rng.randn()/np.sqrt(-np.log(rng.rand()))))*rng.randn() return 10.0**(1.5 - 3*np.abs(rng.randn(N)/np.sqrt(-np.log(rng.rand(N)))))*rng.randn(N) def wrap(x, a, b): assert b > a return (x - a)%(b - a) + a
<commit_before># -*- coding: utf-8 -*- __all__ = ["randh", "wrap"] import numpy as np import numpy.random as rng def randh(): """ Generate from the heavy-tailed distribution. """ return 10.0**(1.5 - 3*np.abs(rng.randn()/np.sqrt(-np.log(rng.rand()))))*rng.randn() def wrap(x, a, b): assert b > a return (x - a)%(b - a) + a <commit_msg>Allow N > 1 randhs to be generated<commit_after>
# -*- coding: utf-8 -*- __all__ = ["randh", "wrap"] import numpy as np import numpy.random as rng def randh(N=1): """ Generate from the heavy-tailed distribution. """ if N==1: return 10.0**(1.5 - 3*np.abs(rng.randn()/np.sqrt(-np.log(rng.rand()))))*rng.randn() return 10.0**(1.5 - 3*np.abs(rng.randn(N)/np.sqrt(-np.log(rng.rand(N)))))*rng.randn(N) def wrap(x, a, b): assert b > a return (x - a)%(b - a) + a
# -*- coding: utf-8 -*- __all__ = ["randh", "wrap"] import numpy as np import numpy.random as rng def randh(): """ Generate from the heavy-tailed distribution. """ return 10.0**(1.5 - 3*np.abs(rng.randn()/np.sqrt(-np.log(rng.rand()))))*rng.randn() def wrap(x, a, b): assert b > a return (x - a)%(b - a) + a Allow N > 1 randhs to be generated# -*- coding: utf-8 -*- __all__ = ["randh", "wrap"] import numpy as np import numpy.random as rng def randh(N=1): """ Generate from the heavy-tailed distribution. """ if N==1: return 10.0**(1.5 - 3*np.abs(rng.randn()/np.sqrt(-np.log(rng.rand()))))*rng.randn() return 10.0**(1.5 - 3*np.abs(rng.randn(N)/np.sqrt(-np.log(rng.rand(N)))))*rng.randn(N) def wrap(x, a, b): assert b > a return (x - a)%(b - a) + a
<commit_before># -*- coding: utf-8 -*- __all__ = ["randh", "wrap"] import numpy as np import numpy.random as rng def randh(): """ Generate from the heavy-tailed distribution. """ return 10.0**(1.5 - 3*np.abs(rng.randn()/np.sqrt(-np.log(rng.rand()))))*rng.randn() def wrap(x, a, b): assert b > a return (x - a)%(b - a) + a <commit_msg>Allow N > 1 randhs to be generated<commit_after># -*- coding: utf-8 -*- __all__ = ["randh", "wrap"] import numpy as np import numpy.random as rng def randh(N=1): """ Generate from the heavy-tailed distribution. """ if N==1: return 10.0**(1.5 - 3*np.abs(rng.randn()/np.sqrt(-np.log(rng.rand()))))*rng.randn() return 10.0**(1.5 - 3*np.abs(rng.randn(N)/np.sqrt(-np.log(rng.rand(N)))))*rng.randn(N) def wrap(x, a, b): assert b > a return (x - a)%(b - a) + a
6d7d04b095eacb413b07ebcb3fed9684dc40fc80
utils/gyb_syntax_support/protocolsMap.py
utils/gyb_syntax_support/protocolsMap.py
SYNTAX_BUILDABLE_EXPRESSIBLE_AS_CONFORMANCES = { 'DeclBuildable': [ 'CodeBlockItem', 'MemberDeclListItem', 'SyntaxBuildable' ], 'ExprList': [ 'ConditionElement', 'SyntaxBuildable' ], 'IdentifierPattern': [ 'PatternBuildable' ], 'MemberDeclList': [ 'MemberDeclBlock' ], 'SimpleTypeIdentifier': [ 'TypeAnnotation', 'TypeBuildable', 'TypeExpr' ], 'StmtBuildable': [ 'CodeBlockItem', 'SyntaxBuildable' ] }
SYNTAX_BUILDABLE_EXPRESSIBLE_AS_CONFORMANCES = { 'DeclBuildable': [ 'CodeBlockItem', 'MemberDeclListItem', 'SyntaxBuildable' ], 'ExprList': [ 'ConditionElement', 'SyntaxBuildable' ], 'IdentifierPattern': [ 'PatternBuildable' ], 'MemberDeclList': [ 'MemberDeclBlock' ], 'SimpleTypeIdentifier': [ 'TypeAnnotation', 'TypeBuildable', 'TypeExpr' ], 'StmtBuildable': [ 'CodeBlockItem', 'SyntaxBuildable' ], 'TokenSyntax': [ 'BinaryOperatorExpr' ] }
Add convenience initializer for `BinaryOperatorExpr`
[SwiftSyntax] Add convenience initializer for `BinaryOperatorExpr`
Python
apache-2.0
atrick/swift,atrick/swift,apple/swift,rudkx/swift,benlangmuir/swift,roambotics/swift,rudkx/swift,glessard/swift,atrick/swift,ahoppen/swift,glessard/swift,JGiola/swift,atrick/swift,ahoppen/swift,ahoppen/swift,apple/swift,glessard/swift,JGiola/swift,rudkx/swift,rudkx/swift,atrick/swift,roambotics/swift,apple/swift,roambotics/swift,glessard/swift,benlangmuir/swift,atrick/swift,benlangmuir/swift,apple/swift,benlangmuir/swift,ahoppen/swift,benlangmuir/swift,roambotics/swift,apple/swift,gregomni/swift,JGiola/swift,gregomni/swift,gregomni/swift,apple/swift,gregomni/swift,gregomni/swift,JGiola/swift,ahoppen/swift,glessard/swift,glessard/swift,roambotics/swift,JGiola/swift,benlangmuir/swift,ahoppen/swift,roambotics/swift,rudkx/swift,rudkx/swift,gregomni/swift,JGiola/swift
SYNTAX_BUILDABLE_EXPRESSIBLE_AS_CONFORMANCES = { 'DeclBuildable': [ 'CodeBlockItem', 'MemberDeclListItem', 'SyntaxBuildable' ], 'ExprList': [ 'ConditionElement', 'SyntaxBuildable' ], 'IdentifierPattern': [ 'PatternBuildable' ], 'MemberDeclList': [ 'MemberDeclBlock' ], 'SimpleTypeIdentifier': [ 'TypeAnnotation', 'TypeBuildable', 'TypeExpr' ], 'StmtBuildable': [ 'CodeBlockItem', 'SyntaxBuildable' ] } [SwiftSyntax] Add convenience initializer for `BinaryOperatorExpr`
SYNTAX_BUILDABLE_EXPRESSIBLE_AS_CONFORMANCES = { 'DeclBuildable': [ 'CodeBlockItem', 'MemberDeclListItem', 'SyntaxBuildable' ], 'ExprList': [ 'ConditionElement', 'SyntaxBuildable' ], 'IdentifierPattern': [ 'PatternBuildable' ], 'MemberDeclList': [ 'MemberDeclBlock' ], 'SimpleTypeIdentifier': [ 'TypeAnnotation', 'TypeBuildable', 'TypeExpr' ], 'StmtBuildable': [ 'CodeBlockItem', 'SyntaxBuildable' ], 'TokenSyntax': [ 'BinaryOperatorExpr' ] }
<commit_before>SYNTAX_BUILDABLE_EXPRESSIBLE_AS_CONFORMANCES = { 'DeclBuildable': [ 'CodeBlockItem', 'MemberDeclListItem', 'SyntaxBuildable' ], 'ExprList': [ 'ConditionElement', 'SyntaxBuildable' ], 'IdentifierPattern': [ 'PatternBuildable' ], 'MemberDeclList': [ 'MemberDeclBlock' ], 'SimpleTypeIdentifier': [ 'TypeAnnotation', 'TypeBuildable', 'TypeExpr' ], 'StmtBuildable': [ 'CodeBlockItem', 'SyntaxBuildable' ] } <commit_msg>[SwiftSyntax] Add convenience initializer for `BinaryOperatorExpr`<commit_after>
SYNTAX_BUILDABLE_EXPRESSIBLE_AS_CONFORMANCES = { 'DeclBuildable': [ 'CodeBlockItem', 'MemberDeclListItem', 'SyntaxBuildable' ], 'ExprList': [ 'ConditionElement', 'SyntaxBuildable' ], 'IdentifierPattern': [ 'PatternBuildable' ], 'MemberDeclList': [ 'MemberDeclBlock' ], 'SimpleTypeIdentifier': [ 'TypeAnnotation', 'TypeBuildable', 'TypeExpr' ], 'StmtBuildable': [ 'CodeBlockItem', 'SyntaxBuildable' ], 'TokenSyntax': [ 'BinaryOperatorExpr' ] }
SYNTAX_BUILDABLE_EXPRESSIBLE_AS_CONFORMANCES = { 'DeclBuildable': [ 'CodeBlockItem', 'MemberDeclListItem', 'SyntaxBuildable' ], 'ExprList': [ 'ConditionElement', 'SyntaxBuildable' ], 'IdentifierPattern': [ 'PatternBuildable' ], 'MemberDeclList': [ 'MemberDeclBlock' ], 'SimpleTypeIdentifier': [ 'TypeAnnotation', 'TypeBuildable', 'TypeExpr' ], 'StmtBuildable': [ 'CodeBlockItem', 'SyntaxBuildable' ] } [SwiftSyntax] Add convenience initializer for `BinaryOperatorExpr`SYNTAX_BUILDABLE_EXPRESSIBLE_AS_CONFORMANCES = { 'DeclBuildable': [ 'CodeBlockItem', 'MemberDeclListItem', 'SyntaxBuildable' ], 'ExprList': [ 'ConditionElement', 'SyntaxBuildable' ], 'IdentifierPattern': [ 'PatternBuildable' ], 'MemberDeclList': [ 'MemberDeclBlock' ], 'SimpleTypeIdentifier': [ 'TypeAnnotation', 'TypeBuildable', 'TypeExpr' ], 'StmtBuildable': [ 'CodeBlockItem', 'SyntaxBuildable' ], 'TokenSyntax': [ 'BinaryOperatorExpr' ] }
<commit_before>SYNTAX_BUILDABLE_EXPRESSIBLE_AS_CONFORMANCES = { 'DeclBuildable': [ 'CodeBlockItem', 'MemberDeclListItem', 'SyntaxBuildable' ], 'ExprList': [ 'ConditionElement', 'SyntaxBuildable' ], 'IdentifierPattern': [ 'PatternBuildable' ], 'MemberDeclList': [ 'MemberDeclBlock' ], 'SimpleTypeIdentifier': [ 'TypeAnnotation', 'TypeBuildable', 'TypeExpr' ], 'StmtBuildable': [ 'CodeBlockItem', 'SyntaxBuildable' ] } <commit_msg>[SwiftSyntax] Add convenience initializer for `BinaryOperatorExpr`<commit_after>SYNTAX_BUILDABLE_EXPRESSIBLE_AS_CONFORMANCES = { 'DeclBuildable': [ 'CodeBlockItem', 'MemberDeclListItem', 'SyntaxBuildable' ], 'ExprList': [ 'ConditionElement', 'SyntaxBuildable' ], 'IdentifierPattern': [ 'PatternBuildable' ], 'MemberDeclList': [ 'MemberDeclBlock' ], 'SimpleTypeIdentifier': [ 'TypeAnnotation', 'TypeBuildable', 'TypeExpr' ], 'StmtBuildable': [ 'CodeBlockItem', 'SyntaxBuildable' ], 'TokenSyntax': [ 'BinaryOperatorExpr' ] }
93b4fd9c6d2c7b113551d1f6f565c7fffc66b5e2
astral/conf/global_settings.py
astral/conf/global_settings.py
import logging import logging.handlers DEBUG = True LOG_FORMAT = '[%(asctime)s: %(levelname)s] %(message)s' if DEBUG: LOG_LEVEL = logging.DEBUG else: LOG_LEVEL = logging.WARN LOG_COLOR = True PORT = 8000 TORNADO_SETTINGS = {} TORNADO_SETTINGS['debug'] = DEBUG TORNADO_SETTINGS['xsrf_cookies'] = False TORNADO_SETTINGS['port'] = PORT if DEBUG: LOG_LEVEL = logging.DEBUG else: LOG_LEVEL = logging.INFO USE_SYSLOG = False LOGGING_CONFIG = 'astral.conf.logconfig.initialize_logging' LOGGING = { 'loggers': { 'astral': {}, }, 'syslog_facility': logging.handlers.SysLogHandler.LOG_LOCAL0, 'syslog_tag': "astral", 'log_level': LOG_LEVEL, 'use_syslog': USE_SYSLOG, } ASTRAL_WEBSERVER = "http://localhost:4567" BOOTSTRAP_NODES = [ ] DOWNSTREAM_CHECK_LIMIT = 1024 * 1024 * 2 UPSTREAM_CHECK_LIMIT = 1024 * 256 OUTGOING_STREAM_LIMIT = 2
import logging import logging.handlers DEBUG = True LOG_FORMAT = '[%(asctime)s: %(levelname)s] %(message)s' if DEBUG: LOG_LEVEL = logging.DEBUG else: LOG_LEVEL = logging.WARN LOG_COLOR = True PORT = 8000 TORNADO_SETTINGS = {} TORNADO_SETTINGS['debug'] = DEBUG TORNADO_SETTINGS['xsrf_cookies'] = False TORNADO_SETTINGS['port'] = PORT if DEBUG: LOG_LEVEL = logging.DEBUG else: LOG_LEVEL = logging.INFO USE_SYSLOG = False LOGGING_CONFIG = 'astral.conf.logconfig.initialize_logging' LOGGING = { 'loggers': { 'astral': {}, }, 'syslog_facility': logging.handlers.SysLogHandler.LOG_LOCAL0, 'syslog_tag': "astral", 'log_level': LOG_LEVEL, 'use_syslog': USE_SYSLOG, } if DEBUG: ASTRAL_WEBSERVER = "http://localhost:4567" else: ASTRAL_WEBSERVER = "http://astral-video.heroku.com" BOOTSTRAP_NODES = [ ] DOWNSTREAM_CHECK_LIMIT = 1024 * 1024 * 2 UPSTREAM_CHECK_LIMIT = 1024 * 256 OUTGOING_STREAM_LIMIT = 2
Use Heroku instance of webapp if not in DEBUG mode.
Use Heroku instance of webapp if not in DEBUG mode.
Python
mit
peplin/astral
import logging import logging.handlers DEBUG = True LOG_FORMAT = '[%(asctime)s: %(levelname)s] %(message)s' if DEBUG: LOG_LEVEL = logging.DEBUG else: LOG_LEVEL = logging.WARN LOG_COLOR = True PORT = 8000 TORNADO_SETTINGS = {} TORNADO_SETTINGS['debug'] = DEBUG TORNADO_SETTINGS['xsrf_cookies'] = False TORNADO_SETTINGS['port'] = PORT if DEBUG: LOG_LEVEL = logging.DEBUG else: LOG_LEVEL = logging.INFO USE_SYSLOG = False LOGGING_CONFIG = 'astral.conf.logconfig.initialize_logging' LOGGING = { 'loggers': { 'astral': {}, }, 'syslog_facility': logging.handlers.SysLogHandler.LOG_LOCAL0, 'syslog_tag': "astral", 'log_level': LOG_LEVEL, 'use_syslog': USE_SYSLOG, } ASTRAL_WEBSERVER = "http://localhost:4567" BOOTSTRAP_NODES = [ ] DOWNSTREAM_CHECK_LIMIT = 1024 * 1024 * 2 UPSTREAM_CHECK_LIMIT = 1024 * 256 OUTGOING_STREAM_LIMIT = 2 Use Heroku instance of webapp if not in DEBUG mode.
import logging import logging.handlers DEBUG = True LOG_FORMAT = '[%(asctime)s: %(levelname)s] %(message)s' if DEBUG: LOG_LEVEL = logging.DEBUG else: LOG_LEVEL = logging.WARN LOG_COLOR = True PORT = 8000 TORNADO_SETTINGS = {} TORNADO_SETTINGS['debug'] = DEBUG TORNADO_SETTINGS['xsrf_cookies'] = False TORNADO_SETTINGS['port'] = PORT if DEBUG: LOG_LEVEL = logging.DEBUG else: LOG_LEVEL = logging.INFO USE_SYSLOG = False LOGGING_CONFIG = 'astral.conf.logconfig.initialize_logging' LOGGING = { 'loggers': { 'astral': {}, }, 'syslog_facility': logging.handlers.SysLogHandler.LOG_LOCAL0, 'syslog_tag': "astral", 'log_level': LOG_LEVEL, 'use_syslog': USE_SYSLOG, } if DEBUG: ASTRAL_WEBSERVER = "http://localhost:4567" else: ASTRAL_WEBSERVER = "http://astral-video.heroku.com" BOOTSTRAP_NODES = [ ] DOWNSTREAM_CHECK_LIMIT = 1024 * 1024 * 2 UPSTREAM_CHECK_LIMIT = 1024 * 256 OUTGOING_STREAM_LIMIT = 2
<commit_before>import logging import logging.handlers DEBUG = True LOG_FORMAT = '[%(asctime)s: %(levelname)s] %(message)s' if DEBUG: LOG_LEVEL = logging.DEBUG else: LOG_LEVEL = logging.WARN LOG_COLOR = True PORT = 8000 TORNADO_SETTINGS = {} TORNADO_SETTINGS['debug'] = DEBUG TORNADO_SETTINGS['xsrf_cookies'] = False TORNADO_SETTINGS['port'] = PORT if DEBUG: LOG_LEVEL = logging.DEBUG else: LOG_LEVEL = logging.INFO USE_SYSLOG = False LOGGING_CONFIG = 'astral.conf.logconfig.initialize_logging' LOGGING = { 'loggers': { 'astral': {}, }, 'syslog_facility': logging.handlers.SysLogHandler.LOG_LOCAL0, 'syslog_tag': "astral", 'log_level': LOG_LEVEL, 'use_syslog': USE_SYSLOG, } ASTRAL_WEBSERVER = "http://localhost:4567" BOOTSTRAP_NODES = [ ] DOWNSTREAM_CHECK_LIMIT = 1024 * 1024 * 2 UPSTREAM_CHECK_LIMIT = 1024 * 256 OUTGOING_STREAM_LIMIT = 2 <commit_msg>Use Heroku instance of webapp if not in DEBUG mode.<commit_after>
import logging import logging.handlers DEBUG = True LOG_FORMAT = '[%(asctime)s: %(levelname)s] %(message)s' if DEBUG: LOG_LEVEL = logging.DEBUG else: LOG_LEVEL = logging.WARN LOG_COLOR = True PORT = 8000 TORNADO_SETTINGS = {} TORNADO_SETTINGS['debug'] = DEBUG TORNADO_SETTINGS['xsrf_cookies'] = False TORNADO_SETTINGS['port'] = PORT if DEBUG: LOG_LEVEL = logging.DEBUG else: LOG_LEVEL = logging.INFO USE_SYSLOG = False LOGGING_CONFIG = 'astral.conf.logconfig.initialize_logging' LOGGING = { 'loggers': { 'astral': {}, }, 'syslog_facility': logging.handlers.SysLogHandler.LOG_LOCAL0, 'syslog_tag': "astral", 'log_level': LOG_LEVEL, 'use_syslog': USE_SYSLOG, } if DEBUG: ASTRAL_WEBSERVER = "http://localhost:4567" else: ASTRAL_WEBSERVER = "http://astral-video.heroku.com" BOOTSTRAP_NODES = [ ] DOWNSTREAM_CHECK_LIMIT = 1024 * 1024 * 2 UPSTREAM_CHECK_LIMIT = 1024 * 256 OUTGOING_STREAM_LIMIT = 2
import logging import logging.handlers DEBUG = True LOG_FORMAT = '[%(asctime)s: %(levelname)s] %(message)s' if DEBUG: LOG_LEVEL = logging.DEBUG else: LOG_LEVEL = logging.WARN LOG_COLOR = True PORT = 8000 TORNADO_SETTINGS = {} TORNADO_SETTINGS['debug'] = DEBUG TORNADO_SETTINGS['xsrf_cookies'] = False TORNADO_SETTINGS['port'] = PORT if DEBUG: LOG_LEVEL = logging.DEBUG else: LOG_LEVEL = logging.INFO USE_SYSLOG = False LOGGING_CONFIG = 'astral.conf.logconfig.initialize_logging' LOGGING = { 'loggers': { 'astral': {}, }, 'syslog_facility': logging.handlers.SysLogHandler.LOG_LOCAL0, 'syslog_tag': "astral", 'log_level': LOG_LEVEL, 'use_syslog': USE_SYSLOG, } ASTRAL_WEBSERVER = "http://localhost:4567" BOOTSTRAP_NODES = [ ] DOWNSTREAM_CHECK_LIMIT = 1024 * 1024 * 2 UPSTREAM_CHECK_LIMIT = 1024 * 256 OUTGOING_STREAM_LIMIT = 2 Use Heroku instance of webapp if not in DEBUG mode.import logging import logging.handlers DEBUG = True LOG_FORMAT = '[%(asctime)s: %(levelname)s] %(message)s' if DEBUG: LOG_LEVEL = logging.DEBUG else: LOG_LEVEL = logging.WARN LOG_COLOR = True PORT = 8000 TORNADO_SETTINGS = {} TORNADO_SETTINGS['debug'] = DEBUG TORNADO_SETTINGS['xsrf_cookies'] = False TORNADO_SETTINGS['port'] = PORT if DEBUG: LOG_LEVEL = logging.DEBUG else: LOG_LEVEL = logging.INFO USE_SYSLOG = False LOGGING_CONFIG = 'astral.conf.logconfig.initialize_logging' LOGGING = { 'loggers': { 'astral': {}, }, 'syslog_facility': logging.handlers.SysLogHandler.LOG_LOCAL0, 'syslog_tag': "astral", 'log_level': LOG_LEVEL, 'use_syslog': USE_SYSLOG, } if DEBUG: ASTRAL_WEBSERVER = "http://localhost:4567" else: ASTRAL_WEBSERVER = "http://astral-video.heroku.com" BOOTSTRAP_NODES = [ ] DOWNSTREAM_CHECK_LIMIT = 1024 * 1024 * 2 UPSTREAM_CHECK_LIMIT = 1024 * 256 OUTGOING_STREAM_LIMIT = 2
<commit_before>import logging import logging.handlers DEBUG = True LOG_FORMAT = '[%(asctime)s: %(levelname)s] %(message)s' if DEBUG: LOG_LEVEL = logging.DEBUG else: LOG_LEVEL = logging.WARN LOG_COLOR = True PORT = 8000 TORNADO_SETTINGS = {} TORNADO_SETTINGS['debug'] = DEBUG TORNADO_SETTINGS['xsrf_cookies'] = False TORNADO_SETTINGS['port'] = PORT if DEBUG: LOG_LEVEL = logging.DEBUG else: LOG_LEVEL = logging.INFO USE_SYSLOG = False LOGGING_CONFIG = 'astral.conf.logconfig.initialize_logging' LOGGING = { 'loggers': { 'astral': {}, }, 'syslog_facility': logging.handlers.SysLogHandler.LOG_LOCAL0, 'syslog_tag': "astral", 'log_level': LOG_LEVEL, 'use_syslog': USE_SYSLOG, } ASTRAL_WEBSERVER = "http://localhost:4567" BOOTSTRAP_NODES = [ ] DOWNSTREAM_CHECK_LIMIT = 1024 * 1024 * 2 UPSTREAM_CHECK_LIMIT = 1024 * 256 OUTGOING_STREAM_LIMIT = 2 <commit_msg>Use Heroku instance of webapp if not in DEBUG mode.<commit_after>import logging import logging.handlers DEBUG = True LOG_FORMAT = '[%(asctime)s: %(levelname)s] %(message)s' if DEBUG: LOG_LEVEL = logging.DEBUG else: LOG_LEVEL = logging.WARN LOG_COLOR = True PORT = 8000 TORNADO_SETTINGS = {} TORNADO_SETTINGS['debug'] = DEBUG TORNADO_SETTINGS['xsrf_cookies'] = False TORNADO_SETTINGS['port'] = PORT if DEBUG: LOG_LEVEL = logging.DEBUG else: LOG_LEVEL = logging.INFO USE_SYSLOG = False LOGGING_CONFIG = 'astral.conf.logconfig.initialize_logging' LOGGING = { 'loggers': { 'astral': {}, }, 'syslog_facility': logging.handlers.SysLogHandler.LOG_LOCAL0, 'syslog_tag': "astral", 'log_level': LOG_LEVEL, 'use_syslog': USE_SYSLOG, } if DEBUG: ASTRAL_WEBSERVER = "http://localhost:4567" else: ASTRAL_WEBSERVER = "http://astral-video.heroku.com" BOOTSTRAP_NODES = [ ] DOWNSTREAM_CHECK_LIMIT = 1024 * 1024 * 2 UPSTREAM_CHECK_LIMIT = 1024 * 256 OUTGOING_STREAM_LIMIT = 2
4fa21d91ada4df904fc8fdddff608fc40c49aaee
reviewboard/accounts/urls.py
reviewboard/accounts/urls.py
from __future__ import unicode_literals from django.conf.urls import patterns, url from reviewboard.accounts.views import MyAccountView urlpatterns = patterns( "reviewboard.accounts.views", url(r'^register/$', 'account_register', {'next_url': 'dashboard'}, name="register"), url(r'^preferences/$', MyAccountView.as_view(), name="user-preferences"), ) urlpatterns += patterns( "django.contrib.auth.views", url(r'^login/$', 'login', {'template_name': 'accounts/login.html'}, name='login'), url(r'^logout/$', 'logout_then_login', name='logout'), url(r'^recover/$', 'password_reset', { 'template_name': 'accounts/password_reset.html', 'email_template_name': 'accounts/password_reset_email.txt' }, name='recover'), url(r'^recover/done/$', 'password_reset_done', {'template_name': 'accounts/password_reset_done.html'}, name='password_reset_done'), url(r'^reset/(?P<uidb64>[0-9A-Za-z_\-]+)-(?P<token>.+)/$', 'password_reset_confirm', {'template_name': 'accounts/password_reset_confirm.html'}, name='password-reset-confirm'), url(r'^reset/done/$', 'password_reset_complete', {'template_name': 'accounts/password_reset_complete.html'}), )
from __future__ import unicode_literals from django.conf.urls import patterns, url from reviewboard.accounts.views import MyAccountView urlpatterns = patterns( "reviewboard.accounts.views", url(r'^register/$', 'account_register', {'next_url': 'dashboard'}, name="register"), url(r'^preferences/$', MyAccountView.as_view(), name="user-preferences"), ) urlpatterns += patterns( "django.contrib.auth.views", url(r'^login/$', 'login', {'template_name': 'accounts/login.html'}, name='login'), url(r'^logout/$', 'logout_then_login', name='logout'), url(r'^recover/$', 'password_reset', { 'template_name': 'accounts/password_reset.html', 'email_template_name': 'accounts/password_reset_email.txt' }, name='recover'), url(r'^recover/done/$', 'password_reset_complete', {'template_name': 'accounts/password_reset_done.html'}, name='password_reset_done'), url(r'^reset/(?P<uidb64>[0-9A-Za-z_\-]+)-(?P<token>.+)/$', 'password_reset_confirm', {'template_name': 'accounts/password_reset_confirm.html'}, name='password-reset-confirm'), url(r'^reset/done/$', 'password_reset_complete', {'template_name': 'accounts/password_reset_complete.html'}), )
Change 'password_reset_done' URL name to 'password_reset_complete'
Change 'password_reset_done' URL name to 'password_reset_complete' The built-in password reset views use a different name for the last page in the flow than we did before, and somehow we never noticed this until recently. Trivial fix. Fixes bug 3345.
Python
mit
reviewboard/reviewboard,custode/reviewboard,1tush/reviewboard,brennie/reviewboard,1tush/reviewboard,bkochendorfer/reviewboard,reviewboard/reviewboard,1tush/reviewboard,custode/reviewboard,chipx86/reviewboard,beol/reviewboard,reviewboard/reviewboard,1tush/reviewboard,custode/reviewboard,1tush/reviewboard,davidt/reviewboard,bkochendorfer/reviewboard,sgallagher/reviewboard,sgallagher/reviewboard,beol/reviewboard,sgallagher/reviewboard,chipx86/reviewboard,KnowNo/reviewboard,1tush/reviewboard,brennie/reviewboard,sgallagher/reviewboard,1tush/reviewboard,davidt/reviewboard,bkochendorfer/reviewboard,KnowNo/reviewboard,custode/reviewboard,beol/reviewboard,chipx86/reviewboard,beol/reviewboard,davidt/reviewboard,1tush/reviewboard,davidt/reviewboard,brennie/reviewboard,KnowNo/reviewboard,1tush/reviewboard,chipx86/reviewboard,bkochendorfer/reviewboard,KnowNo/reviewboard,brennie/reviewboard,reviewboard/reviewboard
from __future__ import unicode_literals from django.conf.urls import patterns, url from reviewboard.accounts.views import MyAccountView urlpatterns = patterns( "reviewboard.accounts.views", url(r'^register/$', 'account_register', {'next_url': 'dashboard'}, name="register"), url(r'^preferences/$', MyAccountView.as_view(), name="user-preferences"), ) urlpatterns += patterns( "django.contrib.auth.views", url(r'^login/$', 'login', {'template_name': 'accounts/login.html'}, name='login'), url(r'^logout/$', 'logout_then_login', name='logout'), url(r'^recover/$', 'password_reset', { 'template_name': 'accounts/password_reset.html', 'email_template_name': 'accounts/password_reset_email.txt' }, name='recover'), url(r'^recover/done/$', 'password_reset_done', {'template_name': 'accounts/password_reset_done.html'}, name='password_reset_done'), url(r'^reset/(?P<uidb64>[0-9A-Za-z_\-]+)-(?P<token>.+)/$', 'password_reset_confirm', {'template_name': 'accounts/password_reset_confirm.html'}, name='password-reset-confirm'), url(r'^reset/done/$', 'password_reset_complete', {'template_name': 'accounts/password_reset_complete.html'}), ) Change 'password_reset_done' URL name to 'password_reset_complete' The built-in password reset views use a different name for the last page in the flow than we did before, and somehow we never noticed this until recently. Trivial fix. Fixes bug 3345.
from __future__ import unicode_literals from django.conf.urls import patterns, url from reviewboard.accounts.views import MyAccountView urlpatterns = patterns( "reviewboard.accounts.views", url(r'^register/$', 'account_register', {'next_url': 'dashboard'}, name="register"), url(r'^preferences/$', MyAccountView.as_view(), name="user-preferences"), ) urlpatterns += patterns( "django.contrib.auth.views", url(r'^login/$', 'login', {'template_name': 'accounts/login.html'}, name='login'), url(r'^logout/$', 'logout_then_login', name='logout'), url(r'^recover/$', 'password_reset', { 'template_name': 'accounts/password_reset.html', 'email_template_name': 'accounts/password_reset_email.txt' }, name='recover'), url(r'^recover/done/$', 'password_reset_complete', {'template_name': 'accounts/password_reset_done.html'}, name='password_reset_done'), url(r'^reset/(?P<uidb64>[0-9A-Za-z_\-]+)-(?P<token>.+)/$', 'password_reset_confirm', {'template_name': 'accounts/password_reset_confirm.html'}, name='password-reset-confirm'), url(r'^reset/done/$', 'password_reset_complete', {'template_name': 'accounts/password_reset_complete.html'}), )
<commit_before>from __future__ import unicode_literals from django.conf.urls import patterns, url from reviewboard.accounts.views import MyAccountView urlpatterns = patterns( "reviewboard.accounts.views", url(r'^register/$', 'account_register', {'next_url': 'dashboard'}, name="register"), url(r'^preferences/$', MyAccountView.as_view(), name="user-preferences"), ) urlpatterns += patterns( "django.contrib.auth.views", url(r'^login/$', 'login', {'template_name': 'accounts/login.html'}, name='login'), url(r'^logout/$', 'logout_then_login', name='logout'), url(r'^recover/$', 'password_reset', { 'template_name': 'accounts/password_reset.html', 'email_template_name': 'accounts/password_reset_email.txt' }, name='recover'), url(r'^recover/done/$', 'password_reset_done', {'template_name': 'accounts/password_reset_done.html'}, name='password_reset_done'), url(r'^reset/(?P<uidb64>[0-9A-Za-z_\-]+)-(?P<token>.+)/$', 'password_reset_confirm', {'template_name': 'accounts/password_reset_confirm.html'}, name='password-reset-confirm'), url(r'^reset/done/$', 'password_reset_complete', {'template_name': 'accounts/password_reset_complete.html'}), ) <commit_msg>Change 'password_reset_done' URL name to 'password_reset_complete' The built-in password reset views use a different name for the last page in the flow than we did before, and somehow we never noticed this until recently. Trivial fix. Fixes bug 3345.<commit_after>
from __future__ import unicode_literals from django.conf.urls import patterns, url from reviewboard.accounts.views import MyAccountView urlpatterns = patterns( "reviewboard.accounts.views", url(r'^register/$', 'account_register', {'next_url': 'dashboard'}, name="register"), url(r'^preferences/$', MyAccountView.as_view(), name="user-preferences"), ) urlpatterns += patterns( "django.contrib.auth.views", url(r'^login/$', 'login', {'template_name': 'accounts/login.html'}, name='login'), url(r'^logout/$', 'logout_then_login', name='logout'), url(r'^recover/$', 'password_reset', { 'template_name': 'accounts/password_reset.html', 'email_template_name': 'accounts/password_reset_email.txt' }, name='recover'), url(r'^recover/done/$', 'password_reset_complete', {'template_name': 'accounts/password_reset_done.html'}, name='password_reset_done'), url(r'^reset/(?P<uidb64>[0-9A-Za-z_\-]+)-(?P<token>.+)/$', 'password_reset_confirm', {'template_name': 'accounts/password_reset_confirm.html'}, name='password-reset-confirm'), url(r'^reset/done/$', 'password_reset_complete', {'template_name': 'accounts/password_reset_complete.html'}), )
from __future__ import unicode_literals from django.conf.urls import patterns, url from reviewboard.accounts.views import MyAccountView urlpatterns = patterns( "reviewboard.accounts.views", url(r'^register/$', 'account_register', {'next_url': 'dashboard'}, name="register"), url(r'^preferences/$', MyAccountView.as_view(), name="user-preferences"), ) urlpatterns += patterns( "django.contrib.auth.views", url(r'^login/$', 'login', {'template_name': 'accounts/login.html'}, name='login'), url(r'^logout/$', 'logout_then_login', name='logout'), url(r'^recover/$', 'password_reset', { 'template_name': 'accounts/password_reset.html', 'email_template_name': 'accounts/password_reset_email.txt' }, name='recover'), url(r'^recover/done/$', 'password_reset_done', {'template_name': 'accounts/password_reset_done.html'}, name='password_reset_done'), url(r'^reset/(?P<uidb64>[0-9A-Za-z_\-]+)-(?P<token>.+)/$', 'password_reset_confirm', {'template_name': 'accounts/password_reset_confirm.html'}, name='password-reset-confirm'), url(r'^reset/done/$', 'password_reset_complete', {'template_name': 'accounts/password_reset_complete.html'}), ) Change 'password_reset_done' URL name to 'password_reset_complete' The built-in password reset views use a different name for the last page in the flow than we did before, and somehow we never noticed this until recently. Trivial fix. Fixes bug 3345.from __future__ import unicode_literals from django.conf.urls import patterns, url from reviewboard.accounts.views import MyAccountView urlpatterns = patterns( "reviewboard.accounts.views", url(r'^register/$', 'account_register', {'next_url': 'dashboard'}, name="register"), url(r'^preferences/$', MyAccountView.as_view(), name="user-preferences"), ) urlpatterns += patterns( "django.contrib.auth.views", url(r'^login/$', 'login', {'template_name': 'accounts/login.html'}, name='login'), url(r'^logout/$', 'logout_then_login', name='logout'), url(r'^recover/$', 'password_reset', { 'template_name': 'accounts/password_reset.html', 'email_template_name': 'accounts/password_reset_email.txt' }, name='recover'), url(r'^recover/done/$', 'password_reset_complete', {'template_name': 'accounts/password_reset_done.html'}, name='password_reset_done'), url(r'^reset/(?P<uidb64>[0-9A-Za-z_\-]+)-(?P<token>.+)/$', 'password_reset_confirm', {'template_name': 'accounts/password_reset_confirm.html'}, name='password-reset-confirm'), url(r'^reset/done/$', 'password_reset_complete', {'template_name': 'accounts/password_reset_complete.html'}), )
<commit_before>from __future__ import unicode_literals from django.conf.urls import patterns, url from reviewboard.accounts.views import MyAccountView urlpatterns = patterns( "reviewboard.accounts.views", url(r'^register/$', 'account_register', {'next_url': 'dashboard'}, name="register"), url(r'^preferences/$', MyAccountView.as_view(), name="user-preferences"), ) urlpatterns += patterns( "django.contrib.auth.views", url(r'^login/$', 'login', {'template_name': 'accounts/login.html'}, name='login'), url(r'^logout/$', 'logout_then_login', name='logout'), url(r'^recover/$', 'password_reset', { 'template_name': 'accounts/password_reset.html', 'email_template_name': 'accounts/password_reset_email.txt' }, name='recover'), url(r'^recover/done/$', 'password_reset_done', {'template_name': 'accounts/password_reset_done.html'}, name='password_reset_done'), url(r'^reset/(?P<uidb64>[0-9A-Za-z_\-]+)-(?P<token>.+)/$', 'password_reset_confirm', {'template_name': 'accounts/password_reset_confirm.html'}, name='password-reset-confirm'), url(r'^reset/done/$', 'password_reset_complete', {'template_name': 'accounts/password_reset_complete.html'}), ) <commit_msg>Change 'password_reset_done' URL name to 'password_reset_complete' The built-in password reset views use a different name for the last page in the flow than we did before, and somehow we never noticed this until recently. Trivial fix. Fixes bug 3345.<commit_after>from __future__ import unicode_literals from django.conf.urls import patterns, url from reviewboard.accounts.views import MyAccountView urlpatterns = patterns( "reviewboard.accounts.views", url(r'^register/$', 'account_register', {'next_url': 'dashboard'}, name="register"), url(r'^preferences/$', MyAccountView.as_view(), name="user-preferences"), ) urlpatterns += patterns( "django.contrib.auth.views", url(r'^login/$', 'login', {'template_name': 'accounts/login.html'}, name='login'), url(r'^logout/$', 'logout_then_login', name='logout'), url(r'^recover/$', 'password_reset', { 'template_name': 'accounts/password_reset.html', 'email_template_name': 'accounts/password_reset_email.txt' }, name='recover'), url(r'^recover/done/$', 'password_reset_complete', {'template_name': 'accounts/password_reset_done.html'}, name='password_reset_done'), url(r'^reset/(?P<uidb64>[0-9A-Za-z_\-]+)-(?P<token>.+)/$', 'password_reset_confirm', {'template_name': 'accounts/password_reset_confirm.html'}, name='password-reset-confirm'), url(r'^reset/done/$', 'password_reset_complete', {'template_name': 'accounts/password_reset_complete.html'}), )
62bdb6f2a6df565119661cf2ba1517d19126e26a
member.py
member.py
#!/usr/bin/python3 import os import sys from sqlalchemy import Column, ForeignKey, Integer, String from sqlalchemy.ext.declarative import declarative_base from sqlalchemy.orm import relationship from sqlalchemy import create_engine Base = declarative_base() class Member(Base): __tablename__ = 'member' id = Column(Integer, primary_key=True) globalName = Column(String(250), nullable=False) nickname = Column(String(250), nullable=True) role = Column(String(250), nullable=True) points = Column(Integer, nullable=True) def __init__(self,id,globalAccountName,serverNickname,role,points): self.id = id self.globalAccountName = globalAccountName self.serverNickname = serverNickname self.role = role self.points = points #Display member method. def displayMember(self): print ("ID:", self.id, "GlobalAccount:", self.globalName, "Nickname:", self.Nickname, "Role:", self.role, "Points:", self.points) #Example of creating one object, changing value of an object and displaying an object using a method. '''mem1 = Member(43344454,"larry","doessuck","officer",150) mem1.globalAccountName="Jack" mem1.displayMember()''' engine = create_engine('sqlite:///bayohwoolph.db') Base.metadata.create_all(engine)
#!/usr/bin/python3 import os import sys from sqlalchemy import Column, ForeignKey, Integer, String from sqlalchemy.ext.declarative import declarative_base from sqlalchemy.orm import relationship from sqlalchemy import create_engine Base = declarative_base() class Member(Base): __tablename__ = 'member' id = Column(Integer, primary_key=True) globalName = Column(String(250), nullable=True) nickname = Column(String(250), nullable=True) role = Column(String(250), nullable=True) points = Column(Integer, nullable=True) def __init__(self,id,globalAccountName,serverNickname,role,points): self.id = id self.globalAccountName = globalAccountName self.serverNickname = serverNickname self.role = role self.points = points #Display member method. def displayMember(self): print ("ID:", self.id, "GlobalAccount:", self.globalName, "Nickname:", self.Nickname, "Role:", self.role, "Points:", self.points) #Example of creating one object, changing value of an object and displaying an object using a method. '''mem1 = Member(43344454,"larry","doessuck","officer",150) mem1.globalAccountName="Jack" mem1.displayMember()''' engine = create_engine('sqlite:///bayohwoolph.db') Base.metadata.create_all(engine)
Make globalName nullable for now.
Make globalName nullable for now.
Python
agpl-3.0
dark-echo/Bay-Oh-Woolph,freiheit/Bay-Oh-Woolph
#!/usr/bin/python3 import os import sys from sqlalchemy import Column, ForeignKey, Integer, String from sqlalchemy.ext.declarative import declarative_base from sqlalchemy.orm import relationship from sqlalchemy import create_engine Base = declarative_base() class Member(Base): __tablename__ = 'member' id = Column(Integer, primary_key=True) globalName = Column(String(250), nullable=False) nickname = Column(String(250), nullable=True) role = Column(String(250), nullable=True) points = Column(Integer, nullable=True) def __init__(self,id,globalAccountName,serverNickname,role,points): self.id = id self.globalAccountName = globalAccountName self.serverNickname = serverNickname self.role = role self.points = points #Display member method. def displayMember(self): print ("ID:", self.id, "GlobalAccount:", self.globalName, "Nickname:", self.Nickname, "Role:", self.role, "Points:", self.points) #Example of creating one object, changing value of an object and displaying an object using a method. '''mem1 = Member(43344454,"larry","doessuck","officer",150) mem1.globalAccountName="Jack" mem1.displayMember()''' engine = create_engine('sqlite:///bayohwoolph.db') Base.metadata.create_all(engine)Make globalName nullable for now.
#!/usr/bin/python3 import os import sys from sqlalchemy import Column, ForeignKey, Integer, String from sqlalchemy.ext.declarative import declarative_base from sqlalchemy.orm import relationship from sqlalchemy import create_engine Base = declarative_base() class Member(Base): __tablename__ = 'member' id = Column(Integer, primary_key=True) globalName = Column(String(250), nullable=True) nickname = Column(String(250), nullable=True) role = Column(String(250), nullable=True) points = Column(Integer, nullable=True) def __init__(self,id,globalAccountName,serverNickname,role,points): self.id = id self.globalAccountName = globalAccountName self.serverNickname = serverNickname self.role = role self.points = points #Display member method. def displayMember(self): print ("ID:", self.id, "GlobalAccount:", self.globalName, "Nickname:", self.Nickname, "Role:", self.role, "Points:", self.points) #Example of creating one object, changing value of an object and displaying an object using a method. '''mem1 = Member(43344454,"larry","doessuck","officer",150) mem1.globalAccountName="Jack" mem1.displayMember()''' engine = create_engine('sqlite:///bayohwoolph.db') Base.metadata.create_all(engine)
<commit_before>#!/usr/bin/python3 import os import sys from sqlalchemy import Column, ForeignKey, Integer, String from sqlalchemy.ext.declarative import declarative_base from sqlalchemy.orm import relationship from sqlalchemy import create_engine Base = declarative_base() class Member(Base): __tablename__ = 'member' id = Column(Integer, primary_key=True) globalName = Column(String(250), nullable=False) nickname = Column(String(250), nullable=True) role = Column(String(250), nullable=True) points = Column(Integer, nullable=True) def __init__(self,id,globalAccountName,serverNickname,role,points): self.id = id self.globalAccountName = globalAccountName self.serverNickname = serverNickname self.role = role self.points = points #Display member method. def displayMember(self): print ("ID:", self.id, "GlobalAccount:", self.globalName, "Nickname:", self.Nickname, "Role:", self.role, "Points:", self.points) #Example of creating one object, changing value of an object and displaying an object using a method. '''mem1 = Member(43344454,"larry","doessuck","officer",150) mem1.globalAccountName="Jack" mem1.displayMember()''' engine = create_engine('sqlite:///bayohwoolph.db') Base.metadata.create_all(engine)<commit_msg>Make globalName nullable for now.<commit_after>
#!/usr/bin/python3 import os import sys from sqlalchemy import Column, ForeignKey, Integer, String from sqlalchemy.ext.declarative import declarative_base from sqlalchemy.orm import relationship from sqlalchemy import create_engine Base = declarative_base() class Member(Base): __tablename__ = 'member' id = Column(Integer, primary_key=True) globalName = Column(String(250), nullable=True) nickname = Column(String(250), nullable=True) role = Column(String(250), nullable=True) points = Column(Integer, nullable=True) def __init__(self,id,globalAccountName,serverNickname,role,points): self.id = id self.globalAccountName = globalAccountName self.serverNickname = serverNickname self.role = role self.points = points #Display member method. def displayMember(self): print ("ID:", self.id, "GlobalAccount:", self.globalName, "Nickname:", self.Nickname, "Role:", self.role, "Points:", self.points) #Example of creating one object, changing value of an object and displaying an object using a method. '''mem1 = Member(43344454,"larry","doessuck","officer",150) mem1.globalAccountName="Jack" mem1.displayMember()''' engine = create_engine('sqlite:///bayohwoolph.db') Base.metadata.create_all(engine)
#!/usr/bin/python3 import os import sys from sqlalchemy import Column, ForeignKey, Integer, String from sqlalchemy.ext.declarative import declarative_base from sqlalchemy.orm import relationship from sqlalchemy import create_engine Base = declarative_base() class Member(Base): __tablename__ = 'member' id = Column(Integer, primary_key=True) globalName = Column(String(250), nullable=False) nickname = Column(String(250), nullable=True) role = Column(String(250), nullable=True) points = Column(Integer, nullable=True) def __init__(self,id,globalAccountName,serverNickname,role,points): self.id = id self.globalAccountName = globalAccountName self.serverNickname = serverNickname self.role = role self.points = points #Display member method. def displayMember(self): print ("ID:", self.id, "GlobalAccount:", self.globalName, "Nickname:", self.Nickname, "Role:", self.role, "Points:", self.points) #Example of creating one object, changing value of an object and displaying an object using a method. '''mem1 = Member(43344454,"larry","doessuck","officer",150) mem1.globalAccountName="Jack" mem1.displayMember()''' engine = create_engine('sqlite:///bayohwoolph.db') Base.metadata.create_all(engine)Make globalName nullable for now.#!/usr/bin/python3 import os import sys from sqlalchemy import Column, ForeignKey, Integer, String from sqlalchemy.ext.declarative import declarative_base from sqlalchemy.orm import relationship from sqlalchemy import create_engine Base = declarative_base() class Member(Base): __tablename__ = 'member' id = Column(Integer, primary_key=True) globalName = Column(String(250), nullable=True) nickname = Column(String(250), nullable=True) role = Column(String(250), nullable=True) points = Column(Integer, nullable=True) def __init__(self,id,globalAccountName,serverNickname,role,points): self.id = id self.globalAccountName = globalAccountName self.serverNickname = serverNickname self.role = role self.points = points #Display member method. def displayMember(self): print ("ID:", self.id, "GlobalAccount:", self.globalName, "Nickname:", self.Nickname, "Role:", self.role, "Points:", self.points) #Example of creating one object, changing value of an object and displaying an object using a method. '''mem1 = Member(43344454,"larry","doessuck","officer",150) mem1.globalAccountName="Jack" mem1.displayMember()''' engine = create_engine('sqlite:///bayohwoolph.db') Base.metadata.create_all(engine)
<commit_before>#!/usr/bin/python3 import os import sys from sqlalchemy import Column, ForeignKey, Integer, String from sqlalchemy.ext.declarative import declarative_base from sqlalchemy.orm import relationship from sqlalchemy import create_engine Base = declarative_base() class Member(Base): __tablename__ = 'member' id = Column(Integer, primary_key=True) globalName = Column(String(250), nullable=False) nickname = Column(String(250), nullable=True) role = Column(String(250), nullable=True) points = Column(Integer, nullable=True) def __init__(self,id,globalAccountName,serverNickname,role,points): self.id = id self.globalAccountName = globalAccountName self.serverNickname = serverNickname self.role = role self.points = points #Display member method. def displayMember(self): print ("ID:", self.id, "GlobalAccount:", self.globalName, "Nickname:", self.Nickname, "Role:", self.role, "Points:", self.points) #Example of creating one object, changing value of an object and displaying an object using a method. '''mem1 = Member(43344454,"larry","doessuck","officer",150) mem1.globalAccountName="Jack" mem1.displayMember()''' engine = create_engine('sqlite:///bayohwoolph.db') Base.metadata.create_all(engine)<commit_msg>Make globalName nullable for now.<commit_after>#!/usr/bin/python3 import os import sys from sqlalchemy import Column, ForeignKey, Integer, String from sqlalchemy.ext.declarative import declarative_base from sqlalchemy.orm import relationship from sqlalchemy import create_engine Base = declarative_base() class Member(Base): __tablename__ = 'member' id = Column(Integer, primary_key=True) globalName = Column(String(250), nullable=True) nickname = Column(String(250), nullable=True) role = Column(String(250), nullable=True) points = Column(Integer, nullable=True) def __init__(self,id,globalAccountName,serverNickname,role,points): self.id = id self.globalAccountName = globalAccountName self.serverNickname = serverNickname self.role = role self.points = points #Display member method. def displayMember(self): print ("ID:", self.id, "GlobalAccount:", self.globalName, "Nickname:", self.Nickname, "Role:", self.role, "Points:", self.points) #Example of creating one object, changing value of an object and displaying an object using a method. '''mem1 = Member(43344454,"larry","doessuck","officer",150) mem1.globalAccountName="Jack" mem1.displayMember()''' engine = create_engine('sqlite:///bayohwoolph.db') Base.metadata.create_all(engine)
9fb89c7e76bd3c6db5f7283b91a2852225056b40
tests/test_analyse.py
tests/test_analyse.py
"""Test analysis page.""" def test_analysis(webapp): """Test we can analyse a page.""" response = webapp.get('/analyse/646e73747769737465722e7265706f7274') assert response.status_code == 200 assert 'Use these tools to safely analyse dnstwister.report' in response.body
"""Test analysis page.""" import pytest import webtest.app def test_analysis(webapp): """Test we can analyse a page.""" response = webapp.get('/analyse/646e73747769737465722e7265706f7274') assert response.status_code == 200 assert 'Use these tools to safely analyse dnstwister.report' in response.body def test_bad_domain_fails(webapp): """Test the analyse page checks domain validity.""" with pytest.raises(webtest.app.AppError) as err: webapp.get('/analyse/3234jskdnfsdf7y34') assert '400 BAD REQUEST' in err.value.message
Test analyse page checks domain validity
Test analyse page checks domain validity
Python
unlicense
thisismyrobot/dnstwister,thisismyrobot/dnstwister,thisismyrobot/dnstwister
"""Test analysis page.""" def test_analysis(webapp): """Test we can analyse a page.""" response = webapp.get('/analyse/646e73747769737465722e7265706f7274') assert response.status_code == 200 assert 'Use these tools to safely analyse dnstwister.report' in response.body Test analyse page checks domain validity
"""Test analysis page.""" import pytest import webtest.app def test_analysis(webapp): """Test we can analyse a page.""" response = webapp.get('/analyse/646e73747769737465722e7265706f7274') assert response.status_code == 200 assert 'Use these tools to safely analyse dnstwister.report' in response.body def test_bad_domain_fails(webapp): """Test the analyse page checks domain validity.""" with pytest.raises(webtest.app.AppError) as err: webapp.get('/analyse/3234jskdnfsdf7y34') assert '400 BAD REQUEST' in err.value.message
<commit_before>"""Test analysis page.""" def test_analysis(webapp): """Test we can analyse a page.""" response = webapp.get('/analyse/646e73747769737465722e7265706f7274') assert response.status_code == 200 assert 'Use these tools to safely analyse dnstwister.report' in response.body <commit_msg>Test analyse page checks domain validity<commit_after>
"""Test analysis page.""" import pytest import webtest.app def test_analysis(webapp): """Test we can analyse a page.""" response = webapp.get('/analyse/646e73747769737465722e7265706f7274') assert response.status_code == 200 assert 'Use these tools to safely analyse dnstwister.report' in response.body def test_bad_domain_fails(webapp): """Test the analyse page checks domain validity.""" with pytest.raises(webtest.app.AppError) as err: webapp.get('/analyse/3234jskdnfsdf7y34') assert '400 BAD REQUEST' in err.value.message
"""Test analysis page.""" def test_analysis(webapp): """Test we can analyse a page.""" response = webapp.get('/analyse/646e73747769737465722e7265706f7274') assert response.status_code == 200 assert 'Use these tools to safely analyse dnstwister.report' in response.body Test analyse page checks domain validity"""Test analysis page.""" import pytest import webtest.app def test_analysis(webapp): """Test we can analyse a page.""" response = webapp.get('/analyse/646e73747769737465722e7265706f7274') assert response.status_code == 200 assert 'Use these tools to safely analyse dnstwister.report' in response.body def test_bad_domain_fails(webapp): """Test the analyse page checks domain validity.""" with pytest.raises(webtest.app.AppError) as err: webapp.get('/analyse/3234jskdnfsdf7y34') assert '400 BAD REQUEST' in err.value.message
<commit_before>"""Test analysis page.""" def test_analysis(webapp): """Test we can analyse a page.""" response = webapp.get('/analyse/646e73747769737465722e7265706f7274') assert response.status_code == 200 assert 'Use these tools to safely analyse dnstwister.report' in response.body <commit_msg>Test analyse page checks domain validity<commit_after>"""Test analysis page.""" import pytest import webtest.app def test_analysis(webapp): """Test we can analyse a page.""" response = webapp.get('/analyse/646e73747769737465722e7265706f7274') assert response.status_code == 200 assert 'Use these tools to safely analyse dnstwister.report' in response.body def test_bad_domain_fails(webapp): """Test the analyse page checks domain validity.""" with pytest.raises(webtest.app.AppError) as err: webapp.get('/analyse/3234jskdnfsdf7y34') assert '400 BAD REQUEST' in err.value.message
323b201b8a498402d9f45cbc5cbac049299feea8
api/bots/incrementor/incrementor.py
api/bots/incrementor/incrementor.py
# See readme.md for instructions on running this code. class IncrementorHandler(object): def __init__(self): self.number = 0 self.message_id = None def usage(self): return ''' This is a boilerplate bot that makes use of the update_message function. For the first @-mention, it initially replies with one message containing a `1`. Every time the bot is @-mentioned, this number will be incremented in the same message. ''' def handle_message(self, message, bot_handler, state_handler): self.number += 1 if self.message_id is None: result = bot_handler.send_reply(message, str(self.number)) self.message_id = result['id'] else: bot_handler.update_message(dict( message_id=self.message_id, content=str(self.number), )) handler_class = IncrementorHandler
# See readme.md for instructions on running this code. class IncrementorHandler(object): def usage(self): return ''' This is a boilerplate bot that makes use of the update_message function. For the first @-mention, it initially replies with one message containing a `1`. Every time the bot is @-mentioned, this number will be incremented in the same message. ''' def handle_message(self, message, bot_handler, state_handler): state = state_handler.get_state() or {'number': 0, 'message_id': None} state['number'] += 1 state_handler.set_state(state) if state['message_id'] is None: result = bot_handler.send_reply(message, str(state['number'])) state['message_id'] = result['id'] state_handler.set_state(state) else: bot_handler.update_message(dict( message_id = state['message_id'], content = str(state['number']) )) handler_class = IncrementorHandler
Adjust Incrementor bot to use StateHandler
Bots: Adjust Incrementor bot to use StateHandler
Python
apache-2.0
jackrzhang/zulip,zulip/zulip,Galexrt/zulip,eeshangarg/zulip,kou/zulip,brainwane/zulip,timabbott/zulip,verma-varsha/zulip,vabs22/zulip,mahim97/zulip,verma-varsha/zulip,hackerkid/zulip,rishig/zulip,zulip/zulip,rht/zulip,timabbott/zulip,punchagan/zulip,jackrzhang/zulip,punchagan/zulip,timabbott/zulip,hackerkid/zulip,punchagan/zulip,vaidap/zulip,synicalsyntax/zulip,vaidap/zulip,verma-varsha/zulip,vabs22/zulip,tommyip/zulip,jackrzhang/zulip,punchagan/zulip,Galexrt/zulip,amanharitsh123/zulip,verma-varsha/zulip,hackerkid/zulip,brockwhittaker/zulip,jackrzhang/zulip,showell/zulip,jrowan/zulip,showell/zulip,Galexrt/zulip,vaidap/zulip,synicalsyntax/zulip,tommyip/zulip,synicalsyntax/zulip,brainwane/zulip,jackrzhang/zulip,mahim97/zulip,eeshangarg/zulip,hackerkid/zulip,kou/zulip,eeshangarg/zulip,dhcrzf/zulip,kou/zulip,dhcrzf/zulip,rishig/zulip,rht/zulip,showell/zulip,andersk/zulip,dhcrzf/zulip,Galexrt/zulip,punchagan/zulip,synicalsyntax/zulip,zulip/zulip,andersk/zulip,dhcrzf/zulip,tommyip/zulip,timabbott/zulip,tommyip/zulip,vabs22/zulip,timabbott/zulip,kou/zulip,jrowan/zulip,jackrzhang/zulip,jrowan/zulip,amanharitsh123/zulip,vaidap/zulip,rht/zulip,eeshangarg/zulip,rht/zulip,amanharitsh123/zulip,shubhamdhama/zulip,rht/zulip,brainwane/zulip,mahim97/zulip,mahim97/zulip,jrowan/zulip,brockwhittaker/zulip,brainwane/zulip,brainwane/zulip,punchagan/zulip,dhcrzf/zulip,rht/zulip,zulip/zulip,mahim97/zulip,kou/zulip,kou/zulip,shubhamdhama/zulip,Galexrt/zulip,shubhamdhama/zulip,vaidap/zulip,jrowan/zulip,brockwhittaker/zulip,timabbott/zulip,hackerkid/zulip,andersk/zulip,mahim97/zulip,eeshangarg/zulip,andersk/zulip,brockwhittaker/zulip,brockwhittaker/zulip,punchagan/zulip,rishig/zulip,timabbott/zulip,amanharitsh123/zulip,shubhamdhama/zulip,showell/zulip,eeshangarg/zulip,jackrzhang/zulip,andersk/zulip,eeshangarg/zulip,zulip/zulip,amanharitsh123/zulip,zulip/zulip,shubhamdhama/zulip,dhcrzf/zulip,verma-varsha/zulip,rht/zulip,tommyip/zulip,showell/zulip,dhcrzf/zulip,tommyip/zulip,showell/zulip,hackerkid/zulip,tommyip/zulip,vabs22/zulip,verma-varsha/zulip,shubhamdhama/zulip,synicalsyntax/zulip,brainwane/zulip,showell/zulip,rishig/zulip,synicalsyntax/zulip,shubhamdhama/zulip,rishig/zulip,vaidap/zulip,hackerkid/zulip,Galexrt/zulip,vabs22/zulip,brainwane/zulip,Galexrt/zulip,brockwhittaker/zulip,rishig/zulip,zulip/zulip,amanharitsh123/zulip,rishig/zulip,jrowan/zulip,kou/zulip,andersk/zulip,synicalsyntax/zulip,vabs22/zulip,andersk/zulip
# See readme.md for instructions on running this code. class IncrementorHandler(object): def __init__(self): self.number = 0 self.message_id = None def usage(self): return ''' This is a boilerplate bot that makes use of the update_message function. For the first @-mention, it initially replies with one message containing a `1`. Every time the bot is @-mentioned, this number will be incremented in the same message. ''' def handle_message(self, message, bot_handler, state_handler): self.number += 1 if self.message_id is None: result = bot_handler.send_reply(message, str(self.number)) self.message_id = result['id'] else: bot_handler.update_message(dict( message_id=self.message_id, content=str(self.number), )) handler_class = IncrementorHandler Bots: Adjust Incrementor bot to use StateHandler
# See readme.md for instructions on running this code. class IncrementorHandler(object): def usage(self): return ''' This is a boilerplate bot that makes use of the update_message function. For the first @-mention, it initially replies with one message containing a `1`. Every time the bot is @-mentioned, this number will be incremented in the same message. ''' def handle_message(self, message, bot_handler, state_handler): state = state_handler.get_state() or {'number': 0, 'message_id': None} state['number'] += 1 state_handler.set_state(state) if state['message_id'] is None: result = bot_handler.send_reply(message, str(state['number'])) state['message_id'] = result['id'] state_handler.set_state(state) else: bot_handler.update_message(dict( message_id = state['message_id'], content = str(state['number']) )) handler_class = IncrementorHandler
<commit_before># See readme.md for instructions on running this code. class IncrementorHandler(object): def __init__(self): self.number = 0 self.message_id = None def usage(self): return ''' This is a boilerplate bot that makes use of the update_message function. For the first @-mention, it initially replies with one message containing a `1`. Every time the bot is @-mentioned, this number will be incremented in the same message. ''' def handle_message(self, message, bot_handler, state_handler): self.number += 1 if self.message_id is None: result = bot_handler.send_reply(message, str(self.number)) self.message_id = result['id'] else: bot_handler.update_message(dict( message_id=self.message_id, content=str(self.number), )) handler_class = IncrementorHandler <commit_msg>Bots: Adjust Incrementor bot to use StateHandler<commit_after>
# See readme.md for instructions on running this code. class IncrementorHandler(object): def usage(self): return ''' This is a boilerplate bot that makes use of the update_message function. For the first @-mention, it initially replies with one message containing a `1`. Every time the bot is @-mentioned, this number will be incremented in the same message. ''' def handle_message(self, message, bot_handler, state_handler): state = state_handler.get_state() or {'number': 0, 'message_id': None} state['number'] += 1 state_handler.set_state(state) if state['message_id'] is None: result = bot_handler.send_reply(message, str(state['number'])) state['message_id'] = result['id'] state_handler.set_state(state) else: bot_handler.update_message(dict( message_id = state['message_id'], content = str(state['number']) )) handler_class = IncrementorHandler
# See readme.md for instructions on running this code. class IncrementorHandler(object): def __init__(self): self.number = 0 self.message_id = None def usage(self): return ''' This is a boilerplate bot that makes use of the update_message function. For the first @-mention, it initially replies with one message containing a `1`. Every time the bot is @-mentioned, this number will be incremented in the same message. ''' def handle_message(self, message, bot_handler, state_handler): self.number += 1 if self.message_id is None: result = bot_handler.send_reply(message, str(self.number)) self.message_id = result['id'] else: bot_handler.update_message(dict( message_id=self.message_id, content=str(self.number), )) handler_class = IncrementorHandler Bots: Adjust Incrementor bot to use StateHandler# See readme.md for instructions on running this code. class IncrementorHandler(object): def usage(self): return ''' This is a boilerplate bot that makes use of the update_message function. For the first @-mention, it initially replies with one message containing a `1`. Every time the bot is @-mentioned, this number will be incremented in the same message. ''' def handle_message(self, message, bot_handler, state_handler): state = state_handler.get_state() or {'number': 0, 'message_id': None} state['number'] += 1 state_handler.set_state(state) if state['message_id'] is None: result = bot_handler.send_reply(message, str(state['number'])) state['message_id'] = result['id'] state_handler.set_state(state) else: bot_handler.update_message(dict( message_id = state['message_id'], content = str(state['number']) )) handler_class = IncrementorHandler
<commit_before># See readme.md for instructions on running this code. class IncrementorHandler(object): def __init__(self): self.number = 0 self.message_id = None def usage(self): return ''' This is a boilerplate bot that makes use of the update_message function. For the first @-mention, it initially replies with one message containing a `1`. Every time the bot is @-mentioned, this number will be incremented in the same message. ''' def handle_message(self, message, bot_handler, state_handler): self.number += 1 if self.message_id is None: result = bot_handler.send_reply(message, str(self.number)) self.message_id = result['id'] else: bot_handler.update_message(dict( message_id=self.message_id, content=str(self.number), )) handler_class = IncrementorHandler <commit_msg>Bots: Adjust Incrementor bot to use StateHandler<commit_after># See readme.md for instructions on running this code. class IncrementorHandler(object): def usage(self): return ''' This is a boilerplate bot that makes use of the update_message function. For the first @-mention, it initially replies with one message containing a `1`. Every time the bot is @-mentioned, this number will be incremented in the same message. ''' def handle_message(self, message, bot_handler, state_handler): state = state_handler.get_state() or {'number': 0, 'message_id': None} state['number'] += 1 state_handler.set_state(state) if state['message_id'] is None: result = bot_handler.send_reply(message, str(state['number'])) state['message_id'] = result['id'] state_handler.set_state(state) else: bot_handler.update_message(dict( message_id = state['message_id'], content = str(state['number']) )) handler_class = IncrementorHandler
ffc7f4e87da72866ce391f78084385eac3485049
src/dicomweb_client/__init__.py
src/dicomweb_client/__init__.py
__version__ = '0.9.2' from dicomweb_client.api import DICOMwebClient
__version__ = '0.9.3' from dicomweb_client.api import DICOMwebClient
Increase version to 0.9.3 for release
Increase version to 0.9.3 for release
Python
mit
MGHComputationalPathology/dicomweb-client
__version__ = '0.9.2' from dicomweb_client.api import DICOMwebClient Increase version to 0.9.3 for release
__version__ = '0.9.3' from dicomweb_client.api import DICOMwebClient
<commit_before>__version__ = '0.9.2' from dicomweb_client.api import DICOMwebClient <commit_msg>Increase version to 0.9.3 for release<commit_after>
__version__ = '0.9.3' from dicomweb_client.api import DICOMwebClient
__version__ = '0.9.2' from dicomweb_client.api import DICOMwebClient Increase version to 0.9.3 for release__version__ = '0.9.3' from dicomweb_client.api import DICOMwebClient
<commit_before>__version__ = '0.9.2' from dicomweb_client.api import DICOMwebClient <commit_msg>Increase version to 0.9.3 for release<commit_after>__version__ = '0.9.3' from dicomweb_client.api import DICOMwebClient
58120c937e04357f6fbdcf1431f69fe7a38aacb2
app/mod_budget/model.py
app/mod_budget/model.py
from app import db from app.mod_auth.model import User class Category(db.Document): # The name of the category. name = db.StringField(required = True) class Entry(db.Document): # The amount of the entry. amount = db.DecimalField(precision = 2, required = True) # A short description for the entry. description = db.StringField(required = True) # The owner of the entry. # Should the owner be deleted, we also want to delete all of his entries. owner = db.ReferenceField(User, reverse_delete_rule = db.CASCADE, required = True) # The category of this entry. category = db.ReferenceField(Category, required = True) class CategoryBudget(db.Document): # The amount of the budget. amount = db.DecimalField(precision = 2, required = True) # The category. category = db.ReferenceField(Category, required = True) def sumEntries(): return sum([entry.amount for entry in Entry.objects if entry.amount > 0])
from app import db from app.mod_auth.model import User class Category(db.Document): # The name of the category. name = db.StringField(required = True) class Income(db.Document): # The amount of the entry. amount = db.DecimalField(precision = 2, required = True) # A short description for the entry. description = db.StringField(required = True) # The owner of the entry. # Should the owner be deleted, we also want to delete all of his entries. owner = db.ReferenceField(User, reverse_delete_rule = db.CASCADE, required = True) # The category of this entry. category = db.ReferenceField(Category, required = True) class Expense(db.Document): # The amount of the entry. amount = db.DecimalField(precision = 2, required = True) # A short description for the expense. description = db.StringField(required = True) # The owner of the expense. # Should the owner be deleted, we also want to delete all of his entries. owner = db.ReferenceField(User, reverse_delete_rule = db.CASCADE, required = True) # The category of this entry. category = db.ReferenceField(Category, required = True) class CategoryBudget(db.Document): # The amount of the budget. amount = db.DecimalField(precision = 2, required = True) # The category. category = db.ReferenceField(Category, required = True) def sumEntries(): return sum([entry.amount for entry in Entry.objects if entry.amount > 0])
Split Entry into Income and Expense schemes
Split Entry into Income and Expense schemes Splitting the Entry schema into two seperate schemes allows us to use different collections to store them, which in turn makes our work easier later on.
Python
mit
Zillolo/mana-vault,Zillolo/mana-vault,Zillolo/mana-vault
from app import db from app.mod_auth.model import User class Category(db.Document): # The name of the category. name = db.StringField(required = True) class Entry(db.Document): # The amount of the entry. amount = db.DecimalField(precision = 2, required = True) # A short description for the entry. description = db.StringField(required = True) # The owner of the entry. # Should the owner be deleted, we also want to delete all of his entries. owner = db.ReferenceField(User, reverse_delete_rule = db.CASCADE, required = True) # The category of this entry. category = db.ReferenceField(Category, required = True) class CategoryBudget(db.Document): # The amount of the budget. amount = db.DecimalField(precision = 2, required = True) # The category. category = db.ReferenceField(Category, required = True) def sumEntries(): return sum([entry.amount for entry in Entry.objects if entry.amount > 0]) Split Entry into Income and Expense schemes Splitting the Entry schema into two seperate schemes allows us to use different collections to store them, which in turn makes our work easier later on.
from app import db from app.mod_auth.model import User class Category(db.Document): # The name of the category. name = db.StringField(required = True) class Income(db.Document): # The amount of the entry. amount = db.DecimalField(precision = 2, required = True) # A short description for the entry. description = db.StringField(required = True) # The owner of the entry. # Should the owner be deleted, we also want to delete all of his entries. owner = db.ReferenceField(User, reverse_delete_rule = db.CASCADE, required = True) # The category of this entry. category = db.ReferenceField(Category, required = True) class Expense(db.Document): # The amount of the entry. amount = db.DecimalField(precision = 2, required = True) # A short description for the expense. description = db.StringField(required = True) # The owner of the expense. # Should the owner be deleted, we also want to delete all of his entries. owner = db.ReferenceField(User, reverse_delete_rule = db.CASCADE, required = True) # The category of this entry. category = db.ReferenceField(Category, required = True) class CategoryBudget(db.Document): # The amount of the budget. amount = db.DecimalField(precision = 2, required = True) # The category. category = db.ReferenceField(Category, required = True) def sumEntries(): return sum([entry.amount for entry in Entry.objects if entry.amount > 0])
<commit_before>from app import db from app.mod_auth.model import User class Category(db.Document): # The name of the category. name = db.StringField(required = True) class Entry(db.Document): # The amount of the entry. amount = db.DecimalField(precision = 2, required = True) # A short description for the entry. description = db.StringField(required = True) # The owner of the entry. # Should the owner be deleted, we also want to delete all of his entries. owner = db.ReferenceField(User, reverse_delete_rule = db.CASCADE, required = True) # The category of this entry. category = db.ReferenceField(Category, required = True) class CategoryBudget(db.Document): # The amount of the budget. amount = db.DecimalField(precision = 2, required = True) # The category. category = db.ReferenceField(Category, required = True) def sumEntries(): return sum([entry.amount for entry in Entry.objects if entry.amount > 0]) <commit_msg>Split Entry into Income and Expense schemes Splitting the Entry schema into two seperate schemes allows us to use different collections to store them, which in turn makes our work easier later on.<commit_after>
from app import db from app.mod_auth.model import User class Category(db.Document): # The name of the category. name = db.StringField(required = True) class Income(db.Document): # The amount of the entry. amount = db.DecimalField(precision = 2, required = True) # A short description for the entry. description = db.StringField(required = True) # The owner of the entry. # Should the owner be deleted, we also want to delete all of his entries. owner = db.ReferenceField(User, reverse_delete_rule = db.CASCADE, required = True) # The category of this entry. category = db.ReferenceField(Category, required = True) class Expense(db.Document): # The amount of the entry. amount = db.DecimalField(precision = 2, required = True) # A short description for the expense. description = db.StringField(required = True) # The owner of the expense. # Should the owner be deleted, we also want to delete all of his entries. owner = db.ReferenceField(User, reverse_delete_rule = db.CASCADE, required = True) # The category of this entry. category = db.ReferenceField(Category, required = True) class CategoryBudget(db.Document): # The amount of the budget. amount = db.DecimalField(precision = 2, required = True) # The category. category = db.ReferenceField(Category, required = True) def sumEntries(): return sum([entry.amount for entry in Entry.objects if entry.amount > 0])
from app import db from app.mod_auth.model import User class Category(db.Document): # The name of the category. name = db.StringField(required = True) class Entry(db.Document): # The amount of the entry. amount = db.DecimalField(precision = 2, required = True) # A short description for the entry. description = db.StringField(required = True) # The owner of the entry. # Should the owner be deleted, we also want to delete all of his entries. owner = db.ReferenceField(User, reverse_delete_rule = db.CASCADE, required = True) # The category of this entry. category = db.ReferenceField(Category, required = True) class CategoryBudget(db.Document): # The amount of the budget. amount = db.DecimalField(precision = 2, required = True) # The category. category = db.ReferenceField(Category, required = True) def sumEntries(): return sum([entry.amount for entry in Entry.objects if entry.amount > 0]) Split Entry into Income and Expense schemes Splitting the Entry schema into two seperate schemes allows us to use different collections to store them, which in turn makes our work easier later on.from app import db from app.mod_auth.model import User class Category(db.Document): # The name of the category. name = db.StringField(required = True) class Income(db.Document): # The amount of the entry. amount = db.DecimalField(precision = 2, required = True) # A short description for the entry. description = db.StringField(required = True) # The owner of the entry. # Should the owner be deleted, we also want to delete all of his entries. owner = db.ReferenceField(User, reverse_delete_rule = db.CASCADE, required = True) # The category of this entry. category = db.ReferenceField(Category, required = True) class Expense(db.Document): # The amount of the entry. amount = db.DecimalField(precision = 2, required = True) # A short description for the expense. description = db.StringField(required = True) # The owner of the expense. # Should the owner be deleted, we also want to delete all of his entries. owner = db.ReferenceField(User, reverse_delete_rule = db.CASCADE, required = True) # The category of this entry. category = db.ReferenceField(Category, required = True) class CategoryBudget(db.Document): # The amount of the budget. amount = db.DecimalField(precision = 2, required = True) # The category. category = db.ReferenceField(Category, required = True) def sumEntries(): return sum([entry.amount for entry in Entry.objects if entry.amount > 0])
<commit_before>from app import db from app.mod_auth.model import User class Category(db.Document): # The name of the category. name = db.StringField(required = True) class Entry(db.Document): # The amount of the entry. amount = db.DecimalField(precision = 2, required = True) # A short description for the entry. description = db.StringField(required = True) # The owner of the entry. # Should the owner be deleted, we also want to delete all of his entries. owner = db.ReferenceField(User, reverse_delete_rule = db.CASCADE, required = True) # The category of this entry. category = db.ReferenceField(Category, required = True) class CategoryBudget(db.Document): # The amount of the budget. amount = db.DecimalField(precision = 2, required = True) # The category. category = db.ReferenceField(Category, required = True) def sumEntries(): return sum([entry.amount for entry in Entry.objects if entry.amount > 0]) <commit_msg>Split Entry into Income and Expense schemes Splitting the Entry schema into two seperate schemes allows us to use different collections to store them, which in turn makes our work easier later on.<commit_after>from app import db from app.mod_auth.model import User class Category(db.Document): # The name of the category. name = db.StringField(required = True) class Income(db.Document): # The amount of the entry. amount = db.DecimalField(precision = 2, required = True) # A short description for the entry. description = db.StringField(required = True) # The owner of the entry. # Should the owner be deleted, we also want to delete all of his entries. owner = db.ReferenceField(User, reverse_delete_rule = db.CASCADE, required = True) # The category of this entry. category = db.ReferenceField(Category, required = True) class Expense(db.Document): # The amount of the entry. amount = db.DecimalField(precision = 2, required = True) # A short description for the expense. description = db.StringField(required = True) # The owner of the expense. # Should the owner be deleted, we also want to delete all of his entries. owner = db.ReferenceField(User, reverse_delete_rule = db.CASCADE, required = True) # The category of this entry. category = db.ReferenceField(Category, required = True) class CategoryBudget(db.Document): # The amount of the budget. amount = db.DecimalField(precision = 2, required = True) # The category. category = db.ReferenceField(Category, required = True) def sumEntries(): return sum([entry.amount for entry in Entry.objects if entry.amount > 0])
d4b1c89a8d365457e1162d5a39815e7fc47781a1
src/epiweb/apps/profile/urls.py
src/epiweb/apps/profile/urls.py
from django.conf.urls.defaults import * urlpatterns = patterns('', (r'^edit/$', 'epiweb.apps.profile.views.edit'), (r'^$', 'epiweb.apps.profile.views.index'), )
from django.conf.urls.defaults import * urlpatterns = patterns('', (r'^$', 'epiweb.apps.profile.views.index'), )
Remove profile 'show' page, only 'edit' page is provided.
Remove profile 'show' page, only 'edit' page is provided.
Python
agpl-3.0
ISIFoundation/influenzanet-website,ISIFoundation/influenzanet-website,ISIFoundation/influenzanet-website,ISIFoundation/influenzanet-website,ISIFoundation/influenzanet-website,ISIFoundation/influenzanet-website,ISIFoundation/influenzanet-website
from django.conf.urls.defaults import * urlpatterns = patterns('', (r'^edit/$', 'epiweb.apps.profile.views.edit'), (r'^$', 'epiweb.apps.profile.views.index'), ) Remove profile 'show' page, only 'edit' page is provided.
from django.conf.urls.defaults import * urlpatterns = patterns('', (r'^$', 'epiweb.apps.profile.views.index'), )
<commit_before>from django.conf.urls.defaults import * urlpatterns = patterns('', (r'^edit/$', 'epiweb.apps.profile.views.edit'), (r'^$', 'epiweb.apps.profile.views.index'), ) <commit_msg>Remove profile 'show' page, only 'edit' page is provided.<commit_after>
from django.conf.urls.defaults import * urlpatterns = patterns('', (r'^$', 'epiweb.apps.profile.views.index'), )
from django.conf.urls.defaults import * urlpatterns = patterns('', (r'^edit/$', 'epiweb.apps.profile.views.edit'), (r'^$', 'epiweb.apps.profile.views.index'), ) Remove profile 'show' page, only 'edit' page is provided.from django.conf.urls.defaults import * urlpatterns = patterns('', (r'^$', 'epiweb.apps.profile.views.index'), )
<commit_before>from django.conf.urls.defaults import * urlpatterns = patterns('', (r'^edit/$', 'epiweb.apps.profile.views.edit'), (r'^$', 'epiweb.apps.profile.views.index'), ) <commit_msg>Remove profile 'show' page, only 'edit' page is provided.<commit_after>from django.conf.urls.defaults import * urlpatterns = patterns('', (r'^$', 'epiweb.apps.profile.views.index'), )
5a0116378b6906fbfb3146bbf635d6d9c39ec714
saleor/dashboard/collection/forms.py
saleor/dashboard/collection/forms.py
from django import forms from ...product.models import Collection class CollectionForm(forms.ModelForm): class Meta: model = Collection exclude = []
from unidecode import unidecode from django import forms from django.utils.text import slugify from ...product.models import Collection class CollectionForm(forms.ModelForm): class Meta: model = Collection exclude = ['slug'] def save(self, commit=True): self.instance.slug = slugify(unidecode(self.instance.name)) super(CollectionForm, self).save(commit=commit) return self.instance
Update CollectionForm to handle slug field
Update CollectionForm to handle slug field
Python
bsd-3-clause
mociepka/saleor,mociepka/saleor,mociepka/saleor,UITools/saleor,maferelo/saleor,UITools/saleor,maferelo/saleor,UITools/saleor,UITools/saleor,UITools/saleor,maferelo/saleor
from django import forms from ...product.models import Collection class CollectionForm(forms.ModelForm): class Meta: model = Collection exclude = [] Update CollectionForm to handle slug field
from unidecode import unidecode from django import forms from django.utils.text import slugify from ...product.models import Collection class CollectionForm(forms.ModelForm): class Meta: model = Collection exclude = ['slug'] def save(self, commit=True): self.instance.slug = slugify(unidecode(self.instance.name)) super(CollectionForm, self).save(commit=commit) return self.instance
<commit_before>from django import forms from ...product.models import Collection class CollectionForm(forms.ModelForm): class Meta: model = Collection exclude = [] <commit_msg>Update CollectionForm to handle slug field<commit_after>
from unidecode import unidecode from django import forms from django.utils.text import slugify from ...product.models import Collection class CollectionForm(forms.ModelForm): class Meta: model = Collection exclude = ['slug'] def save(self, commit=True): self.instance.slug = slugify(unidecode(self.instance.name)) super(CollectionForm, self).save(commit=commit) return self.instance
from django import forms from ...product.models import Collection class CollectionForm(forms.ModelForm): class Meta: model = Collection exclude = [] Update CollectionForm to handle slug fieldfrom unidecode import unidecode from django import forms from django.utils.text import slugify from ...product.models import Collection class CollectionForm(forms.ModelForm): class Meta: model = Collection exclude = ['slug'] def save(self, commit=True): self.instance.slug = slugify(unidecode(self.instance.name)) super(CollectionForm, self).save(commit=commit) return self.instance
<commit_before>from django import forms from ...product.models import Collection class CollectionForm(forms.ModelForm): class Meta: model = Collection exclude = [] <commit_msg>Update CollectionForm to handle slug field<commit_after>from unidecode import unidecode from django import forms from django.utils.text import slugify from ...product.models import Collection class CollectionForm(forms.ModelForm): class Meta: model = Collection exclude = ['slug'] def save(self, commit=True): self.instance.slug = slugify(unidecode(self.instance.name)) super(CollectionForm, self).save(commit=commit) return self.instance
5329c48a6f0a36809d3088560f91b427f7a2bf0b
models.py
models.py
from datetime import datetime from app import db, bcrypt class User(db.Model): __tablename__ = "users" id = db.Column(db.Integer, primary_key=True) username = db.Column(db.String(80), unique=True) email = db.Column(db.String(120), unique=True) name = db.Column(db.String()) pw_hash = db.Column(db.String()) graphs = db.relationship("Graph", backref="user", lazy="dynamic") def __init__(self, username, email, name, password): self.username = username self.email = email self.name = name self.pw_hash = bcrypt.generate_password_hash(password).decode("utf-8") def __repr__(self): return self.username def check_password(self, password): return bcrypt.check_password_hash(self.pw_hash, password) class Graph(db.Model): __tablename__ = "graphs" id = db.Column(db.Integer, primary_key=True) created_at = db.Column(db.DateTime) title = db.Column(db.String()) serialized_string = db.Column(db.String()) user_id = db.Column(db.Integer, db.ForeignKey("users.id")) def __init__(self, serialized_string): self.created_at = datetime.utcnow() self.serialized_string = serialized_string def __repr__(self): return self.title
from datetime import datetime from app import db, bcrypt class User(db.Model): __tablename__ = "users" id = db.Column(db.Integer, primary_key=True) username = db.Column(db.String(80), unique=True) email = db.Column(db.String(120), unique=True) name = db.Column(db.String()) pw_hash = db.Column(db.String()) graphs = db.relationship("Graph", backref="user", lazy="dynamic") def __init__(self, username, email, name, password): self.username = username self.email = email self.name = name self.pw_hash = bcrypt.generate_password_hash(password).decode("utf-8") def __repr__(self): return self.username def check_password(self, password): return bcrypt.check_password_hash(self.pw_hash, password) class Graph(db.Model): __tablename__ = "graphs" id = db.Column(db.Integer, primary_key=True) created_at = db.Column(db.DateTime) title = db.Column(db.String()) serialized_string = db.Column(db.String()) user_id = db.Column(db.Integer, db.ForeignKey("users.id")) def __init__(self, title, serialized_string): self.created_at = datetime.utcnow() self.title = title self.serialized_string = serialized_string def __repr__(self): return self.title
Add title to Graph object constructor
Add title to Graph object constructor
Python
mit
ChristopherChudzicki/math3d,stardust66/math3d,stardust66/math3d,stardust66/math3d,stardust66/math3d,ChristopherChudzicki/math3d,ChristopherChudzicki/math3d,ChristopherChudzicki/math3d
from datetime import datetime from app import db, bcrypt class User(db.Model): __tablename__ = "users" id = db.Column(db.Integer, primary_key=True) username = db.Column(db.String(80), unique=True) email = db.Column(db.String(120), unique=True) name = db.Column(db.String()) pw_hash = db.Column(db.String()) graphs = db.relationship("Graph", backref="user", lazy="dynamic") def __init__(self, username, email, name, password): self.username = username self.email = email self.name = name self.pw_hash = bcrypt.generate_password_hash(password).decode("utf-8") def __repr__(self): return self.username def check_password(self, password): return bcrypt.check_password_hash(self.pw_hash, password) class Graph(db.Model): __tablename__ = "graphs" id = db.Column(db.Integer, primary_key=True) created_at = db.Column(db.DateTime) title = db.Column(db.String()) serialized_string = db.Column(db.String()) user_id = db.Column(db.Integer, db.ForeignKey("users.id")) def __init__(self, serialized_string): self.created_at = datetime.utcnow() self.serialized_string = serialized_string def __repr__(self): return self.title Add title to Graph object constructor
from datetime import datetime from app import db, bcrypt class User(db.Model): __tablename__ = "users" id = db.Column(db.Integer, primary_key=True) username = db.Column(db.String(80), unique=True) email = db.Column(db.String(120), unique=True) name = db.Column(db.String()) pw_hash = db.Column(db.String()) graphs = db.relationship("Graph", backref="user", lazy="dynamic") def __init__(self, username, email, name, password): self.username = username self.email = email self.name = name self.pw_hash = bcrypt.generate_password_hash(password).decode("utf-8") def __repr__(self): return self.username def check_password(self, password): return bcrypt.check_password_hash(self.pw_hash, password) class Graph(db.Model): __tablename__ = "graphs" id = db.Column(db.Integer, primary_key=True) created_at = db.Column(db.DateTime) title = db.Column(db.String()) serialized_string = db.Column(db.String()) user_id = db.Column(db.Integer, db.ForeignKey("users.id")) def __init__(self, title, serialized_string): self.created_at = datetime.utcnow() self.title = title self.serialized_string = serialized_string def __repr__(self): return self.title
<commit_before>from datetime import datetime from app import db, bcrypt class User(db.Model): __tablename__ = "users" id = db.Column(db.Integer, primary_key=True) username = db.Column(db.String(80), unique=True) email = db.Column(db.String(120), unique=True) name = db.Column(db.String()) pw_hash = db.Column(db.String()) graphs = db.relationship("Graph", backref="user", lazy="dynamic") def __init__(self, username, email, name, password): self.username = username self.email = email self.name = name self.pw_hash = bcrypt.generate_password_hash(password).decode("utf-8") def __repr__(self): return self.username def check_password(self, password): return bcrypt.check_password_hash(self.pw_hash, password) class Graph(db.Model): __tablename__ = "graphs" id = db.Column(db.Integer, primary_key=True) created_at = db.Column(db.DateTime) title = db.Column(db.String()) serialized_string = db.Column(db.String()) user_id = db.Column(db.Integer, db.ForeignKey("users.id")) def __init__(self, serialized_string): self.created_at = datetime.utcnow() self.serialized_string = serialized_string def __repr__(self): return self.title <commit_msg>Add title to Graph object constructor<commit_after>
from datetime import datetime from app import db, bcrypt class User(db.Model): __tablename__ = "users" id = db.Column(db.Integer, primary_key=True) username = db.Column(db.String(80), unique=True) email = db.Column(db.String(120), unique=True) name = db.Column(db.String()) pw_hash = db.Column(db.String()) graphs = db.relationship("Graph", backref="user", lazy="dynamic") def __init__(self, username, email, name, password): self.username = username self.email = email self.name = name self.pw_hash = bcrypt.generate_password_hash(password).decode("utf-8") def __repr__(self): return self.username def check_password(self, password): return bcrypt.check_password_hash(self.pw_hash, password) class Graph(db.Model): __tablename__ = "graphs" id = db.Column(db.Integer, primary_key=True) created_at = db.Column(db.DateTime) title = db.Column(db.String()) serialized_string = db.Column(db.String()) user_id = db.Column(db.Integer, db.ForeignKey("users.id")) def __init__(self, title, serialized_string): self.created_at = datetime.utcnow() self.title = title self.serialized_string = serialized_string def __repr__(self): return self.title
from datetime import datetime from app import db, bcrypt class User(db.Model): __tablename__ = "users" id = db.Column(db.Integer, primary_key=True) username = db.Column(db.String(80), unique=True) email = db.Column(db.String(120), unique=True) name = db.Column(db.String()) pw_hash = db.Column(db.String()) graphs = db.relationship("Graph", backref="user", lazy="dynamic") def __init__(self, username, email, name, password): self.username = username self.email = email self.name = name self.pw_hash = bcrypt.generate_password_hash(password).decode("utf-8") def __repr__(self): return self.username def check_password(self, password): return bcrypt.check_password_hash(self.pw_hash, password) class Graph(db.Model): __tablename__ = "graphs" id = db.Column(db.Integer, primary_key=True) created_at = db.Column(db.DateTime) title = db.Column(db.String()) serialized_string = db.Column(db.String()) user_id = db.Column(db.Integer, db.ForeignKey("users.id")) def __init__(self, serialized_string): self.created_at = datetime.utcnow() self.serialized_string = serialized_string def __repr__(self): return self.title Add title to Graph object constructorfrom datetime import datetime from app import db, bcrypt class User(db.Model): __tablename__ = "users" id = db.Column(db.Integer, primary_key=True) username = db.Column(db.String(80), unique=True) email = db.Column(db.String(120), unique=True) name = db.Column(db.String()) pw_hash = db.Column(db.String()) graphs = db.relationship("Graph", backref="user", lazy="dynamic") def __init__(self, username, email, name, password): self.username = username self.email = email self.name = name self.pw_hash = bcrypt.generate_password_hash(password).decode("utf-8") def __repr__(self): return self.username def check_password(self, password): return bcrypt.check_password_hash(self.pw_hash, password) class Graph(db.Model): __tablename__ = "graphs" id = db.Column(db.Integer, primary_key=True) created_at = db.Column(db.DateTime) title = db.Column(db.String()) serialized_string = db.Column(db.String()) user_id = db.Column(db.Integer, db.ForeignKey("users.id")) def __init__(self, title, serialized_string): self.created_at = datetime.utcnow() self.title = title self.serialized_string = serialized_string def __repr__(self): return self.title
<commit_before>from datetime import datetime from app import db, bcrypt class User(db.Model): __tablename__ = "users" id = db.Column(db.Integer, primary_key=True) username = db.Column(db.String(80), unique=True) email = db.Column(db.String(120), unique=True) name = db.Column(db.String()) pw_hash = db.Column(db.String()) graphs = db.relationship("Graph", backref="user", lazy="dynamic") def __init__(self, username, email, name, password): self.username = username self.email = email self.name = name self.pw_hash = bcrypt.generate_password_hash(password).decode("utf-8") def __repr__(self): return self.username def check_password(self, password): return bcrypt.check_password_hash(self.pw_hash, password) class Graph(db.Model): __tablename__ = "graphs" id = db.Column(db.Integer, primary_key=True) created_at = db.Column(db.DateTime) title = db.Column(db.String()) serialized_string = db.Column(db.String()) user_id = db.Column(db.Integer, db.ForeignKey("users.id")) def __init__(self, serialized_string): self.created_at = datetime.utcnow() self.serialized_string = serialized_string def __repr__(self): return self.title <commit_msg>Add title to Graph object constructor<commit_after>from datetime import datetime from app import db, bcrypt class User(db.Model): __tablename__ = "users" id = db.Column(db.Integer, primary_key=True) username = db.Column(db.String(80), unique=True) email = db.Column(db.String(120), unique=True) name = db.Column(db.String()) pw_hash = db.Column(db.String()) graphs = db.relationship("Graph", backref="user", lazy="dynamic") def __init__(self, username, email, name, password): self.username = username self.email = email self.name = name self.pw_hash = bcrypt.generate_password_hash(password).decode("utf-8") def __repr__(self): return self.username def check_password(self, password): return bcrypt.check_password_hash(self.pw_hash, password) class Graph(db.Model): __tablename__ = "graphs" id = db.Column(db.Integer, primary_key=True) created_at = db.Column(db.DateTime) title = db.Column(db.String()) serialized_string = db.Column(db.String()) user_id = db.Column(db.Integer, db.ForeignKey("users.id")) def __init__(self, title, serialized_string): self.created_at = datetime.utcnow() self.title = title self.serialized_string = serialized_string def __repr__(self): return self.title
44e3f48b8832dd14f8f9269c02929fa5d4a5c9af
src/librement/profile/models.py
src/librement/profile/models.py
from django.db import models from django_enumfield import EnumField from librement.utils.user_data import PerUserData from .enums import AccountEnum, CountryEnum class Profile(PerUserData('profile')): account_type = EnumField(AccountEnum) organisation = models.CharField(max_length=100, blank=True) address_1 = models.CharField(max_length=150, blank=True) address_2 = models.CharField(max_length=150, blank=True) city = models.CharField(max_length=100, blank=True) region = models.CharField(max_length=100, blank=True) zipcode = models.CharField(max_length=100, blank=True) country = EnumField(CountryEnum)
from django.db import models from django_enumfield import EnumField from librement.utils.user_data import PerUserData from .enums import AccountEnum, CountryEnum class Profile(PerUserData('profile')): account_type = EnumField(AccountEnum, default=AccountEnum.INDIVIDUAL) organisation = models.CharField(max_length=100, blank=True) address_1 = models.CharField(max_length=150, blank=True) address_2 = models.CharField(max_length=150, blank=True) city = models.CharField(max_length=100, blank=True) region = models.CharField(max_length=100, blank=True) zipcode = models.CharField(max_length=100, blank=True) country = EnumField(CountryEnum)
Make this the default so we can create User objects.
Make this the default so we can create User objects. Signed-off-by: Chris Lamb <29e6d179a8d73471df7861382db6dd7e64138033@debian.org>
Python
agpl-3.0
rhertzog/librement,rhertzog/librement,rhertzog/librement
from django.db import models from django_enumfield import EnumField from librement.utils.user_data import PerUserData from .enums import AccountEnum, CountryEnum class Profile(PerUserData('profile')): account_type = EnumField(AccountEnum) organisation = models.CharField(max_length=100, blank=True) address_1 = models.CharField(max_length=150, blank=True) address_2 = models.CharField(max_length=150, blank=True) city = models.CharField(max_length=100, blank=True) region = models.CharField(max_length=100, blank=True) zipcode = models.CharField(max_length=100, blank=True) country = EnumField(CountryEnum) Make this the default so we can create User objects. Signed-off-by: Chris Lamb <29e6d179a8d73471df7861382db6dd7e64138033@debian.org>
from django.db import models from django_enumfield import EnumField from librement.utils.user_data import PerUserData from .enums import AccountEnum, CountryEnum class Profile(PerUserData('profile')): account_type = EnumField(AccountEnum, default=AccountEnum.INDIVIDUAL) organisation = models.CharField(max_length=100, blank=True) address_1 = models.CharField(max_length=150, blank=True) address_2 = models.CharField(max_length=150, blank=True) city = models.CharField(max_length=100, blank=True) region = models.CharField(max_length=100, blank=True) zipcode = models.CharField(max_length=100, blank=True) country = EnumField(CountryEnum)
<commit_before>from django.db import models from django_enumfield import EnumField from librement.utils.user_data import PerUserData from .enums import AccountEnum, CountryEnum class Profile(PerUserData('profile')): account_type = EnumField(AccountEnum) organisation = models.CharField(max_length=100, blank=True) address_1 = models.CharField(max_length=150, blank=True) address_2 = models.CharField(max_length=150, blank=True) city = models.CharField(max_length=100, blank=True) region = models.CharField(max_length=100, blank=True) zipcode = models.CharField(max_length=100, blank=True) country = EnumField(CountryEnum) <commit_msg>Make this the default so we can create User objects. Signed-off-by: Chris Lamb <29e6d179a8d73471df7861382db6dd7e64138033@debian.org><commit_after>
from django.db import models from django_enumfield import EnumField from librement.utils.user_data import PerUserData from .enums import AccountEnum, CountryEnum class Profile(PerUserData('profile')): account_type = EnumField(AccountEnum, default=AccountEnum.INDIVIDUAL) organisation = models.CharField(max_length=100, blank=True) address_1 = models.CharField(max_length=150, blank=True) address_2 = models.CharField(max_length=150, blank=True) city = models.CharField(max_length=100, blank=True) region = models.CharField(max_length=100, blank=True) zipcode = models.CharField(max_length=100, blank=True) country = EnumField(CountryEnum)
from django.db import models from django_enumfield import EnumField from librement.utils.user_data import PerUserData from .enums import AccountEnum, CountryEnum class Profile(PerUserData('profile')): account_type = EnumField(AccountEnum) organisation = models.CharField(max_length=100, blank=True) address_1 = models.CharField(max_length=150, blank=True) address_2 = models.CharField(max_length=150, blank=True) city = models.CharField(max_length=100, blank=True) region = models.CharField(max_length=100, blank=True) zipcode = models.CharField(max_length=100, blank=True) country = EnumField(CountryEnum) Make this the default so we can create User objects. Signed-off-by: Chris Lamb <29e6d179a8d73471df7861382db6dd7e64138033@debian.org>from django.db import models from django_enumfield import EnumField from librement.utils.user_data import PerUserData from .enums import AccountEnum, CountryEnum class Profile(PerUserData('profile')): account_type = EnumField(AccountEnum, default=AccountEnum.INDIVIDUAL) organisation = models.CharField(max_length=100, blank=True) address_1 = models.CharField(max_length=150, blank=True) address_2 = models.CharField(max_length=150, blank=True) city = models.CharField(max_length=100, blank=True) region = models.CharField(max_length=100, blank=True) zipcode = models.CharField(max_length=100, blank=True) country = EnumField(CountryEnum)
<commit_before>from django.db import models from django_enumfield import EnumField from librement.utils.user_data import PerUserData from .enums import AccountEnum, CountryEnum class Profile(PerUserData('profile')): account_type = EnumField(AccountEnum) organisation = models.CharField(max_length=100, blank=True) address_1 = models.CharField(max_length=150, blank=True) address_2 = models.CharField(max_length=150, blank=True) city = models.CharField(max_length=100, blank=True) region = models.CharField(max_length=100, blank=True) zipcode = models.CharField(max_length=100, blank=True) country = EnumField(CountryEnum) <commit_msg>Make this the default so we can create User objects. Signed-off-by: Chris Lamb <29e6d179a8d73471df7861382db6dd7e64138033@debian.org><commit_after>from django.db import models from django_enumfield import EnumField from librement.utils.user_data import PerUserData from .enums import AccountEnum, CountryEnum class Profile(PerUserData('profile')): account_type = EnumField(AccountEnum, default=AccountEnum.INDIVIDUAL) organisation = models.CharField(max_length=100, blank=True) address_1 = models.CharField(max_length=150, blank=True) address_2 = models.CharField(max_length=150, blank=True) city = models.CharField(max_length=100, blank=True) region = models.CharField(max_length=100, blank=True) zipcode = models.CharField(max_length=100, blank=True) country = EnumField(CountryEnum)
e43345616e5240274e852a722c0c72c07f988b2a
registration/__init__.py
registration/__init__.py
VERSION = (0, 9, 0, 'beta', 1) def get_version(): from django.utils.version import get_version as django_get_version return django_get_version(VERSION) # pragma: no cover
VERSION = (1, 0, 0, 'final', 0) def get_version(): "Returns a PEP 386-compliant version number from VERSION." assert len(VERSION) == 5 assert VERSION[3] in ('alpha', 'beta', 'rc', 'final') # Now build the two parts of the version number: # main = X.Y[.Z] # sub = .devN - for pre-alpha releases # | {a|b|c}N - for alpha, beta and rc releases parts = 2 if VERSION[2] == 0 else 3 main = '.'.join(str(x) for x in VERSION[:parts]) sub = '' if VERSION[3] != 'final': mapping = {'alpha': 'a', 'beta': 'b', 'rc': 'c'} sub = mapping[VERSION[3]] + str(VERSION[4]) return str(main + sub)
Fix version number reporting so we can be installed before Django.
Fix version number reporting so we can be installed before Django.
Python
bsd-3-clause
myimages/django-registration,Troyhy/django-registration,mypebble/djregs,akvo/django-registration,Troyhy/django-registration,hacklabr/django-registration,gone/django-registration,akvo/django-registration,tdruez/django-registration,dirtycoder/django-registration,sandipagr/django-registration,kennydude/djregs,danielsamuels/django-registration,awakeup/django-registration,ubernostrum/django-registration,sandipagr/django-registration,gone/django-registration,hacklabr/django-registration
VERSION = (0, 9, 0, 'beta', 1) def get_version(): from django.utils.version import get_version as django_get_version return django_get_version(VERSION) # pragma: no cover Fix version number reporting so we can be installed before Django.
VERSION = (1, 0, 0, 'final', 0) def get_version(): "Returns a PEP 386-compliant version number from VERSION." assert len(VERSION) == 5 assert VERSION[3] in ('alpha', 'beta', 'rc', 'final') # Now build the two parts of the version number: # main = X.Y[.Z] # sub = .devN - for pre-alpha releases # | {a|b|c}N - for alpha, beta and rc releases parts = 2 if VERSION[2] == 0 else 3 main = '.'.join(str(x) for x in VERSION[:parts]) sub = '' if VERSION[3] != 'final': mapping = {'alpha': 'a', 'beta': 'b', 'rc': 'c'} sub = mapping[VERSION[3]] + str(VERSION[4]) return str(main + sub)
<commit_before>VERSION = (0, 9, 0, 'beta', 1) def get_version(): from django.utils.version import get_version as django_get_version return django_get_version(VERSION) # pragma: no cover <commit_msg>Fix version number reporting so we can be installed before Django.<commit_after>
VERSION = (1, 0, 0, 'final', 0) def get_version(): "Returns a PEP 386-compliant version number from VERSION." assert len(VERSION) == 5 assert VERSION[3] in ('alpha', 'beta', 'rc', 'final') # Now build the two parts of the version number: # main = X.Y[.Z] # sub = .devN - for pre-alpha releases # | {a|b|c}N - for alpha, beta and rc releases parts = 2 if VERSION[2] == 0 else 3 main = '.'.join(str(x) for x in VERSION[:parts]) sub = '' if VERSION[3] != 'final': mapping = {'alpha': 'a', 'beta': 'b', 'rc': 'c'} sub = mapping[VERSION[3]] + str(VERSION[4]) return str(main + sub)
VERSION = (0, 9, 0, 'beta', 1) def get_version(): from django.utils.version import get_version as django_get_version return django_get_version(VERSION) # pragma: no cover Fix version number reporting so we can be installed before Django.VERSION = (1, 0, 0, 'final', 0) def get_version(): "Returns a PEP 386-compliant version number from VERSION." assert len(VERSION) == 5 assert VERSION[3] in ('alpha', 'beta', 'rc', 'final') # Now build the two parts of the version number: # main = X.Y[.Z] # sub = .devN - for pre-alpha releases # | {a|b|c}N - for alpha, beta and rc releases parts = 2 if VERSION[2] == 0 else 3 main = '.'.join(str(x) for x in VERSION[:parts]) sub = '' if VERSION[3] != 'final': mapping = {'alpha': 'a', 'beta': 'b', 'rc': 'c'} sub = mapping[VERSION[3]] + str(VERSION[4]) return str(main + sub)
<commit_before>VERSION = (0, 9, 0, 'beta', 1) def get_version(): from django.utils.version import get_version as django_get_version return django_get_version(VERSION) # pragma: no cover <commit_msg>Fix version number reporting so we can be installed before Django.<commit_after>VERSION = (1, 0, 0, 'final', 0) def get_version(): "Returns a PEP 386-compliant version number from VERSION." assert len(VERSION) == 5 assert VERSION[3] in ('alpha', 'beta', 'rc', 'final') # Now build the two parts of the version number: # main = X.Y[.Z] # sub = .devN - for pre-alpha releases # | {a|b|c}N - for alpha, beta and rc releases parts = 2 if VERSION[2] == 0 else 3 main = '.'.join(str(x) for x in VERSION[:parts]) sub = '' if VERSION[3] != 'final': mapping = {'alpha': 'a', 'beta': 'b', 'rc': 'c'} sub = mapping[VERSION[3]] + str(VERSION[4]) return str(main + sub)
dc45f973faa5655e821364cfbdb96a3e17ff9893
app/__init__.py
app/__init__.py
import os from flask import Flask, Blueprint, request, jsonify app = Flask(__name__) # Read configuration to apply from environment config_name = os.environ.get('FLASK_CONFIG', 'development') # apply configuration cfg = os.path.join(os.getcwd(), 'config', config_name + '.py') app.config.from_pyfile(cfg) # Create a blueprint api = Blueprint('api', __name__) # Import the endpoints belonging to this blueprint from . import endpoints from . import errors @api.before_request def before_request(): """All routes in this blueprint require authentication.""" if app.config['AUTH_REQUIRED']: if request.args.get('secret_token'): token = request.headers.get(app.config['TOKEN_HEADER']) if token == app.config['SECRET_TOKEN']: app.logger.info('Validated request token') app.logger.warn('Unauthorized: Invalid request token') app.logger.warn('Unauthorized: No request token included') return jsonify({'error': 'unauthorized'}), 401 app.logger.info('No authentication required') # register blueprints app.register_blueprint(api, url_prefix=app.config['URL_PREFIX'])
import os from flask import Flask, Blueprint, request, jsonify app = Flask(__name__) # Read configuration to apply from environment config_name = os.environ.get('FLASK_CONFIG', 'development') # apply configuration cfg = os.path.join(os.getcwd(), 'config', config_name + '.py') app.config.from_pyfile(cfg) # Create a blueprint api = Blueprint('api', __name__) # Import the endpoints belonging to this blueprint from . import endpoints from . import errors @api.before_request def before_request(): """All routes in this blueprint require authentication.""" if app.config['AUTH_REQUIRED']: token_header = app.config['TOKEN_HEADER'] if request.headers.get(token_header): token = request.headers[token_header] if token == app.config['SECRET_TOKEN']: app.logger.info('Validated request token') app.logger.warn('Unauthorized: Invalid request token') app.logger.warn('Unauthorized: No request token included') return jsonify({'error': 'unauthorized'}), 401 app.logger.info('No authentication required') # register blueprints app.register_blueprint(api, url_prefix=app.config['URL_PREFIX'])
Fix bug in token authentication
Fix bug in token authentication
Python
apache-2.0
javicacheiro/salt-git-synchronizer-proxy
import os from flask import Flask, Blueprint, request, jsonify app = Flask(__name__) # Read configuration to apply from environment config_name = os.environ.get('FLASK_CONFIG', 'development') # apply configuration cfg = os.path.join(os.getcwd(), 'config', config_name + '.py') app.config.from_pyfile(cfg) # Create a blueprint api = Blueprint('api', __name__) # Import the endpoints belonging to this blueprint from . import endpoints from . import errors @api.before_request def before_request(): """All routes in this blueprint require authentication.""" if app.config['AUTH_REQUIRED']: if request.args.get('secret_token'): token = request.headers.get(app.config['TOKEN_HEADER']) if token == app.config['SECRET_TOKEN']: app.logger.info('Validated request token') app.logger.warn('Unauthorized: Invalid request token') app.logger.warn('Unauthorized: No request token included') return jsonify({'error': 'unauthorized'}), 401 app.logger.info('No authentication required') # register blueprints app.register_blueprint(api, url_prefix=app.config['URL_PREFIX']) Fix bug in token authentication
import os from flask import Flask, Blueprint, request, jsonify app = Flask(__name__) # Read configuration to apply from environment config_name = os.environ.get('FLASK_CONFIG', 'development') # apply configuration cfg = os.path.join(os.getcwd(), 'config', config_name + '.py') app.config.from_pyfile(cfg) # Create a blueprint api = Blueprint('api', __name__) # Import the endpoints belonging to this blueprint from . import endpoints from . import errors @api.before_request def before_request(): """All routes in this blueprint require authentication.""" if app.config['AUTH_REQUIRED']: token_header = app.config['TOKEN_HEADER'] if request.headers.get(token_header): token = request.headers[token_header] if token == app.config['SECRET_TOKEN']: app.logger.info('Validated request token') app.logger.warn('Unauthorized: Invalid request token') app.logger.warn('Unauthorized: No request token included') return jsonify({'error': 'unauthorized'}), 401 app.logger.info('No authentication required') # register blueprints app.register_blueprint(api, url_prefix=app.config['URL_PREFIX'])
<commit_before>import os from flask import Flask, Blueprint, request, jsonify app = Flask(__name__) # Read configuration to apply from environment config_name = os.environ.get('FLASK_CONFIG', 'development') # apply configuration cfg = os.path.join(os.getcwd(), 'config', config_name + '.py') app.config.from_pyfile(cfg) # Create a blueprint api = Blueprint('api', __name__) # Import the endpoints belonging to this blueprint from . import endpoints from . import errors @api.before_request def before_request(): """All routes in this blueprint require authentication.""" if app.config['AUTH_REQUIRED']: if request.args.get('secret_token'): token = request.headers.get(app.config['TOKEN_HEADER']) if token == app.config['SECRET_TOKEN']: app.logger.info('Validated request token') app.logger.warn('Unauthorized: Invalid request token') app.logger.warn('Unauthorized: No request token included') return jsonify({'error': 'unauthorized'}), 401 app.logger.info('No authentication required') # register blueprints app.register_blueprint(api, url_prefix=app.config['URL_PREFIX']) <commit_msg>Fix bug in token authentication<commit_after>
import os from flask import Flask, Blueprint, request, jsonify app = Flask(__name__) # Read configuration to apply from environment config_name = os.environ.get('FLASK_CONFIG', 'development') # apply configuration cfg = os.path.join(os.getcwd(), 'config', config_name + '.py') app.config.from_pyfile(cfg) # Create a blueprint api = Blueprint('api', __name__) # Import the endpoints belonging to this blueprint from . import endpoints from . import errors @api.before_request def before_request(): """All routes in this blueprint require authentication.""" if app.config['AUTH_REQUIRED']: token_header = app.config['TOKEN_HEADER'] if request.headers.get(token_header): token = request.headers[token_header] if token == app.config['SECRET_TOKEN']: app.logger.info('Validated request token') app.logger.warn('Unauthorized: Invalid request token') app.logger.warn('Unauthorized: No request token included') return jsonify({'error': 'unauthorized'}), 401 app.logger.info('No authentication required') # register blueprints app.register_blueprint(api, url_prefix=app.config['URL_PREFIX'])
import os from flask import Flask, Blueprint, request, jsonify app = Flask(__name__) # Read configuration to apply from environment config_name = os.environ.get('FLASK_CONFIG', 'development') # apply configuration cfg = os.path.join(os.getcwd(), 'config', config_name + '.py') app.config.from_pyfile(cfg) # Create a blueprint api = Blueprint('api', __name__) # Import the endpoints belonging to this blueprint from . import endpoints from . import errors @api.before_request def before_request(): """All routes in this blueprint require authentication.""" if app.config['AUTH_REQUIRED']: if request.args.get('secret_token'): token = request.headers.get(app.config['TOKEN_HEADER']) if token == app.config['SECRET_TOKEN']: app.logger.info('Validated request token') app.logger.warn('Unauthorized: Invalid request token') app.logger.warn('Unauthorized: No request token included') return jsonify({'error': 'unauthorized'}), 401 app.logger.info('No authentication required') # register blueprints app.register_blueprint(api, url_prefix=app.config['URL_PREFIX']) Fix bug in token authenticationimport os from flask import Flask, Blueprint, request, jsonify app = Flask(__name__) # Read configuration to apply from environment config_name = os.environ.get('FLASK_CONFIG', 'development') # apply configuration cfg = os.path.join(os.getcwd(), 'config', config_name + '.py') app.config.from_pyfile(cfg) # Create a blueprint api = Blueprint('api', __name__) # Import the endpoints belonging to this blueprint from . import endpoints from . import errors @api.before_request def before_request(): """All routes in this blueprint require authentication.""" if app.config['AUTH_REQUIRED']: token_header = app.config['TOKEN_HEADER'] if request.headers.get(token_header): token = request.headers[token_header] if token == app.config['SECRET_TOKEN']: app.logger.info('Validated request token') app.logger.warn('Unauthorized: Invalid request token') app.logger.warn('Unauthorized: No request token included') return jsonify({'error': 'unauthorized'}), 401 app.logger.info('No authentication required') # register blueprints app.register_blueprint(api, url_prefix=app.config['URL_PREFIX'])
<commit_before>import os from flask import Flask, Blueprint, request, jsonify app = Flask(__name__) # Read configuration to apply from environment config_name = os.environ.get('FLASK_CONFIG', 'development') # apply configuration cfg = os.path.join(os.getcwd(), 'config', config_name + '.py') app.config.from_pyfile(cfg) # Create a blueprint api = Blueprint('api', __name__) # Import the endpoints belonging to this blueprint from . import endpoints from . import errors @api.before_request def before_request(): """All routes in this blueprint require authentication.""" if app.config['AUTH_REQUIRED']: if request.args.get('secret_token'): token = request.headers.get(app.config['TOKEN_HEADER']) if token == app.config['SECRET_TOKEN']: app.logger.info('Validated request token') app.logger.warn('Unauthorized: Invalid request token') app.logger.warn('Unauthorized: No request token included') return jsonify({'error': 'unauthorized'}), 401 app.logger.info('No authentication required') # register blueprints app.register_blueprint(api, url_prefix=app.config['URL_PREFIX']) <commit_msg>Fix bug in token authentication<commit_after>import os from flask import Flask, Blueprint, request, jsonify app = Flask(__name__) # Read configuration to apply from environment config_name = os.environ.get('FLASK_CONFIG', 'development') # apply configuration cfg = os.path.join(os.getcwd(), 'config', config_name + '.py') app.config.from_pyfile(cfg) # Create a blueprint api = Blueprint('api', __name__) # Import the endpoints belonging to this blueprint from . import endpoints from . import errors @api.before_request def before_request(): """All routes in this blueprint require authentication.""" if app.config['AUTH_REQUIRED']: token_header = app.config['TOKEN_HEADER'] if request.headers.get(token_header): token = request.headers[token_header] if token == app.config['SECRET_TOKEN']: app.logger.info('Validated request token') app.logger.warn('Unauthorized: Invalid request token') app.logger.warn('Unauthorized: No request token included') return jsonify({'error': 'unauthorized'}), 401 app.logger.info('No authentication required') # register blueprints app.register_blueprint(api, url_prefix=app.config['URL_PREFIX'])
12df471eff4d5ece3da100c3691e9a3d577fa114
EC2/create_instance.py
EC2/create_instance.py
import boto3 import botocore import time ec2 = boto3.resource('ec2', region_name='us-east-1') client = boto3.client('ec2') # Create a security group try: sg = ec2.create_security_group(GroupName='jupyter', Description='EC2 for Jupyter Notebook') response = client.authorize_security_group_ingress(GroupName='jupyter', IpPermissions=[{'PrefixListIds': [], 'UserIdGroupPairs': [], 'IpRanges': [{'CidrIp': '0.0.0.0/0'}], 'IpProtocol': 'tcp', 'Ipv6Ranges': [{'CidrIpv6': '::/0'}], 'ToPort': 8888, 'FromPort': 8888}]) print("create a security group") except botocore.exceptions.ClientError as e: sg = client.describe_security_groups(GroupNames=['jupyter']) print("the security group exist") o = ec2.create_instances(ImageId='ami-e36637f5', MinCount=1, MaxCount=1, InstanceType='i3.xlarge', SecurityGroups=['jupyter']) print_res = False while (not print_res): time.sleep(1) for i in ec2.instances.filter(InstanceIds=[o[0].id]): if i.public_ip_address is not None: print("The public IP address: " + str(i.public_ip_address)) print_res = True
import boto3 import botocore import time ec2 = boto3.resource('ec2', region_name='us-east-1') client = boto3.client('ec2') # Create a security group try: sg = ec2.create_security_group(GroupName='jupyter', Description='EC2 for Jupyter Notebook') response = client.authorize_security_group_ingress(GroupName='jupyter', IpPermissions=[{'PrefixListIds': [], 'UserIdGroupPairs': [], 'IpRanges': [{'CidrIp': '0.0.0.0/0'}], 'IpProtocol': 'tcp', 'Ipv6Ranges': [{'CidrIpv6': '::/0'}], 'ToPort': 8888, 'FromPort': 8888}]) print("create a security group") except botocore.exceptions.ClientError as e: sg = client.describe_security_groups(GroupNames=['jupyter']) print("the security group exist") o = ec2.create_instances(ImageId='ami-622a0119', MinCount=1, MaxCount=1, InstanceType='i3.8xlarge', SecurityGroups=['jupyter']) print_res = False while (not print_res): time.sleep(1) for i in ec2.instances.filter(InstanceIds=[o[0].id]): if i.public_ip_address is not None: print("The public IP address: " + str(i.public_ip_address)) print_res = True
Update the script to create EC2 instance.
Update the script to create EC2 instance. This creates an EC2 i3.8xlarge.
Python
apache-2.0
icoming/FlashX,flashxio/FlashX,icoming/FlashX,icoming/FlashX,flashxio/FlashX,icoming/FlashX,flashxio/FlashX,icoming/FlashX,flashxio/FlashX,flashxio/FlashX,flashxio/FlashX
import boto3 import botocore import time ec2 = boto3.resource('ec2', region_name='us-east-1') client = boto3.client('ec2') # Create a security group try: sg = ec2.create_security_group(GroupName='jupyter', Description='EC2 for Jupyter Notebook') response = client.authorize_security_group_ingress(GroupName='jupyter', IpPermissions=[{'PrefixListIds': [], 'UserIdGroupPairs': [], 'IpRanges': [{'CidrIp': '0.0.0.0/0'}], 'IpProtocol': 'tcp', 'Ipv6Ranges': [{'CidrIpv6': '::/0'}], 'ToPort': 8888, 'FromPort': 8888}]) print("create a security group") except botocore.exceptions.ClientError as e: sg = client.describe_security_groups(GroupNames=['jupyter']) print("the security group exist") o = ec2.create_instances(ImageId='ami-e36637f5', MinCount=1, MaxCount=1, InstanceType='i3.xlarge', SecurityGroups=['jupyter']) print_res = False while (not print_res): time.sleep(1) for i in ec2.instances.filter(InstanceIds=[o[0].id]): if i.public_ip_address is not None: print("The public IP address: " + str(i.public_ip_address)) print_res = True Update the script to create EC2 instance. This creates an EC2 i3.8xlarge.
import boto3 import botocore import time ec2 = boto3.resource('ec2', region_name='us-east-1') client = boto3.client('ec2') # Create a security group try: sg = ec2.create_security_group(GroupName='jupyter', Description='EC2 for Jupyter Notebook') response = client.authorize_security_group_ingress(GroupName='jupyter', IpPermissions=[{'PrefixListIds': [], 'UserIdGroupPairs': [], 'IpRanges': [{'CidrIp': '0.0.0.0/0'}], 'IpProtocol': 'tcp', 'Ipv6Ranges': [{'CidrIpv6': '::/0'}], 'ToPort': 8888, 'FromPort': 8888}]) print("create a security group") except botocore.exceptions.ClientError as e: sg = client.describe_security_groups(GroupNames=['jupyter']) print("the security group exist") o = ec2.create_instances(ImageId='ami-622a0119', MinCount=1, MaxCount=1, InstanceType='i3.8xlarge', SecurityGroups=['jupyter']) print_res = False while (not print_res): time.sleep(1) for i in ec2.instances.filter(InstanceIds=[o[0].id]): if i.public_ip_address is not None: print("The public IP address: " + str(i.public_ip_address)) print_res = True
<commit_before>import boto3 import botocore import time ec2 = boto3.resource('ec2', region_name='us-east-1') client = boto3.client('ec2') # Create a security group try: sg = ec2.create_security_group(GroupName='jupyter', Description='EC2 for Jupyter Notebook') response = client.authorize_security_group_ingress(GroupName='jupyter', IpPermissions=[{'PrefixListIds': [], 'UserIdGroupPairs': [], 'IpRanges': [{'CidrIp': '0.0.0.0/0'}], 'IpProtocol': 'tcp', 'Ipv6Ranges': [{'CidrIpv6': '::/0'}], 'ToPort': 8888, 'FromPort': 8888}]) print("create a security group") except botocore.exceptions.ClientError as e: sg = client.describe_security_groups(GroupNames=['jupyter']) print("the security group exist") o = ec2.create_instances(ImageId='ami-e36637f5', MinCount=1, MaxCount=1, InstanceType='i3.xlarge', SecurityGroups=['jupyter']) print_res = False while (not print_res): time.sleep(1) for i in ec2.instances.filter(InstanceIds=[o[0].id]): if i.public_ip_address is not None: print("The public IP address: " + str(i.public_ip_address)) print_res = True <commit_msg>Update the script to create EC2 instance. This creates an EC2 i3.8xlarge.<commit_after>
import boto3 import botocore import time ec2 = boto3.resource('ec2', region_name='us-east-1') client = boto3.client('ec2') # Create a security group try: sg = ec2.create_security_group(GroupName='jupyter', Description='EC2 for Jupyter Notebook') response = client.authorize_security_group_ingress(GroupName='jupyter', IpPermissions=[{'PrefixListIds': [], 'UserIdGroupPairs': [], 'IpRanges': [{'CidrIp': '0.0.0.0/0'}], 'IpProtocol': 'tcp', 'Ipv6Ranges': [{'CidrIpv6': '::/0'}], 'ToPort': 8888, 'FromPort': 8888}]) print("create a security group") except botocore.exceptions.ClientError as e: sg = client.describe_security_groups(GroupNames=['jupyter']) print("the security group exist") o = ec2.create_instances(ImageId='ami-622a0119', MinCount=1, MaxCount=1, InstanceType='i3.8xlarge', SecurityGroups=['jupyter']) print_res = False while (not print_res): time.sleep(1) for i in ec2.instances.filter(InstanceIds=[o[0].id]): if i.public_ip_address is not None: print("The public IP address: " + str(i.public_ip_address)) print_res = True
import boto3 import botocore import time ec2 = boto3.resource('ec2', region_name='us-east-1') client = boto3.client('ec2') # Create a security group try: sg = ec2.create_security_group(GroupName='jupyter', Description='EC2 for Jupyter Notebook') response = client.authorize_security_group_ingress(GroupName='jupyter', IpPermissions=[{'PrefixListIds': [], 'UserIdGroupPairs': [], 'IpRanges': [{'CidrIp': '0.0.0.0/0'}], 'IpProtocol': 'tcp', 'Ipv6Ranges': [{'CidrIpv6': '::/0'}], 'ToPort': 8888, 'FromPort': 8888}]) print("create a security group") except botocore.exceptions.ClientError as e: sg = client.describe_security_groups(GroupNames=['jupyter']) print("the security group exist") o = ec2.create_instances(ImageId='ami-e36637f5', MinCount=1, MaxCount=1, InstanceType='i3.xlarge', SecurityGroups=['jupyter']) print_res = False while (not print_res): time.sleep(1) for i in ec2.instances.filter(InstanceIds=[o[0].id]): if i.public_ip_address is not None: print("The public IP address: " + str(i.public_ip_address)) print_res = True Update the script to create EC2 instance. This creates an EC2 i3.8xlarge.import boto3 import botocore import time ec2 = boto3.resource('ec2', region_name='us-east-1') client = boto3.client('ec2') # Create a security group try: sg = ec2.create_security_group(GroupName='jupyter', Description='EC2 for Jupyter Notebook') response = client.authorize_security_group_ingress(GroupName='jupyter', IpPermissions=[{'PrefixListIds': [], 'UserIdGroupPairs': [], 'IpRanges': [{'CidrIp': '0.0.0.0/0'}], 'IpProtocol': 'tcp', 'Ipv6Ranges': [{'CidrIpv6': '::/0'}], 'ToPort': 8888, 'FromPort': 8888}]) print("create a security group") except botocore.exceptions.ClientError as e: sg = client.describe_security_groups(GroupNames=['jupyter']) print("the security group exist") o = ec2.create_instances(ImageId='ami-622a0119', MinCount=1, MaxCount=1, InstanceType='i3.8xlarge', SecurityGroups=['jupyter']) print_res = False while (not print_res): time.sleep(1) for i in ec2.instances.filter(InstanceIds=[o[0].id]): if i.public_ip_address is not None: print("The public IP address: " + str(i.public_ip_address)) print_res = True
<commit_before>import boto3 import botocore import time ec2 = boto3.resource('ec2', region_name='us-east-1') client = boto3.client('ec2') # Create a security group try: sg = ec2.create_security_group(GroupName='jupyter', Description='EC2 for Jupyter Notebook') response = client.authorize_security_group_ingress(GroupName='jupyter', IpPermissions=[{'PrefixListIds': [], 'UserIdGroupPairs': [], 'IpRanges': [{'CidrIp': '0.0.0.0/0'}], 'IpProtocol': 'tcp', 'Ipv6Ranges': [{'CidrIpv6': '::/0'}], 'ToPort': 8888, 'FromPort': 8888}]) print("create a security group") except botocore.exceptions.ClientError as e: sg = client.describe_security_groups(GroupNames=['jupyter']) print("the security group exist") o = ec2.create_instances(ImageId='ami-e36637f5', MinCount=1, MaxCount=1, InstanceType='i3.xlarge', SecurityGroups=['jupyter']) print_res = False while (not print_res): time.sleep(1) for i in ec2.instances.filter(InstanceIds=[o[0].id]): if i.public_ip_address is not None: print("The public IP address: " + str(i.public_ip_address)) print_res = True <commit_msg>Update the script to create EC2 instance. This creates an EC2 i3.8xlarge.<commit_after>import boto3 import botocore import time ec2 = boto3.resource('ec2', region_name='us-east-1') client = boto3.client('ec2') # Create a security group try: sg = ec2.create_security_group(GroupName='jupyter', Description='EC2 for Jupyter Notebook') response = client.authorize_security_group_ingress(GroupName='jupyter', IpPermissions=[{'PrefixListIds': [], 'UserIdGroupPairs': [], 'IpRanges': [{'CidrIp': '0.0.0.0/0'}], 'IpProtocol': 'tcp', 'Ipv6Ranges': [{'CidrIpv6': '::/0'}], 'ToPort': 8888, 'FromPort': 8888}]) print("create a security group") except botocore.exceptions.ClientError as e: sg = client.describe_security_groups(GroupNames=['jupyter']) print("the security group exist") o = ec2.create_instances(ImageId='ami-622a0119', MinCount=1, MaxCount=1, InstanceType='i3.8xlarge', SecurityGroups=['jupyter']) print_res = False while (not print_res): time.sleep(1) for i in ec2.instances.filter(InstanceIds=[o[0].id]): if i.public_ip_address is not None: print("The public IP address: " + str(i.public_ip_address)) print_res = True
015ecbbe112edaa3ada4cb1af70f62f03654dfe4
py/app.py
py/app.py
import json import functools from flask import Flask, Response from foxgami.red import Story app = Flask(__name__) def return_as_json(inner_f): @functools.wraps(inner_f) def new_f(*args, **kwargs): result = inner_f(*args, **kwargs) return Response(json.dumps( result, indent=4, separators=(', ', ': ') ), mimetype='application/json') return new_f @app.route('/api/stories') @return_as_json def hardcoded_aww(): return Story.find() @app.route('/api/stories/<string:story_id>') def get_story(story_id): return Story.get(story_id) if __name__ == '__main__': app.run(debug=True)
import json import functools from flask import Flask, Response from foxgami.red import Story app = Flask(__name__) @app.after_response def add_content_headers(response): response.headers['Access-Control-Allow-Origin'] = '*' return response def return_as_json(inner_f): @functools.wraps(inner_f) def new_f(*args, **kwargs): result = inner_f(*args, **kwargs) return Response(json.dumps( result, indent=4, separators=(', ', ': ') ), mimetype='application/json') return new_f @app.route('/api/stories') @return_as_json def hardcoded_aww(): return Story.find() @app.route('/api/stories/<string:story_id>') def get_story(story_id): return Story.get(story_id) if __name__ == '__main__': app.run(debug=True)
Add Access-Control headers to python
Add Access-Control headers to python
Python
mit
flubstep/foxgami.com,flubstep/foxgami.com
import json import functools from flask import Flask, Response from foxgami.red import Story app = Flask(__name__) def return_as_json(inner_f): @functools.wraps(inner_f) def new_f(*args, **kwargs): result = inner_f(*args, **kwargs) return Response(json.dumps( result, indent=4, separators=(', ', ': ') ), mimetype='application/json') return new_f @app.route('/api/stories') @return_as_json def hardcoded_aww(): return Story.find() @app.route('/api/stories/<string:story_id>') def get_story(story_id): return Story.get(story_id) if __name__ == '__main__': app.run(debug=True)Add Access-Control headers to python
import json import functools from flask import Flask, Response from foxgami.red import Story app = Flask(__name__) @app.after_response def add_content_headers(response): response.headers['Access-Control-Allow-Origin'] = '*' return response def return_as_json(inner_f): @functools.wraps(inner_f) def new_f(*args, **kwargs): result = inner_f(*args, **kwargs) return Response(json.dumps( result, indent=4, separators=(', ', ': ') ), mimetype='application/json') return new_f @app.route('/api/stories') @return_as_json def hardcoded_aww(): return Story.find() @app.route('/api/stories/<string:story_id>') def get_story(story_id): return Story.get(story_id) if __name__ == '__main__': app.run(debug=True)
<commit_before>import json import functools from flask import Flask, Response from foxgami.red import Story app = Flask(__name__) def return_as_json(inner_f): @functools.wraps(inner_f) def new_f(*args, **kwargs): result = inner_f(*args, **kwargs) return Response(json.dumps( result, indent=4, separators=(', ', ': ') ), mimetype='application/json') return new_f @app.route('/api/stories') @return_as_json def hardcoded_aww(): return Story.find() @app.route('/api/stories/<string:story_id>') def get_story(story_id): return Story.get(story_id) if __name__ == '__main__': app.run(debug=True)<commit_msg>Add Access-Control headers to python<commit_after>
import json import functools from flask import Flask, Response from foxgami.red import Story app = Flask(__name__) @app.after_response def add_content_headers(response): response.headers['Access-Control-Allow-Origin'] = '*' return response def return_as_json(inner_f): @functools.wraps(inner_f) def new_f(*args, **kwargs): result = inner_f(*args, **kwargs) return Response(json.dumps( result, indent=4, separators=(', ', ': ') ), mimetype='application/json') return new_f @app.route('/api/stories') @return_as_json def hardcoded_aww(): return Story.find() @app.route('/api/stories/<string:story_id>') def get_story(story_id): return Story.get(story_id) if __name__ == '__main__': app.run(debug=True)
import json import functools from flask import Flask, Response from foxgami.red import Story app = Flask(__name__) def return_as_json(inner_f): @functools.wraps(inner_f) def new_f(*args, **kwargs): result = inner_f(*args, **kwargs) return Response(json.dumps( result, indent=4, separators=(', ', ': ') ), mimetype='application/json') return new_f @app.route('/api/stories') @return_as_json def hardcoded_aww(): return Story.find() @app.route('/api/stories/<string:story_id>') def get_story(story_id): return Story.get(story_id) if __name__ == '__main__': app.run(debug=True)Add Access-Control headers to pythonimport json import functools from flask import Flask, Response from foxgami.red import Story app = Flask(__name__) @app.after_response def add_content_headers(response): response.headers['Access-Control-Allow-Origin'] = '*' return response def return_as_json(inner_f): @functools.wraps(inner_f) def new_f(*args, **kwargs): result = inner_f(*args, **kwargs) return Response(json.dumps( result, indent=4, separators=(', ', ': ') ), mimetype='application/json') return new_f @app.route('/api/stories') @return_as_json def hardcoded_aww(): return Story.find() @app.route('/api/stories/<string:story_id>') def get_story(story_id): return Story.get(story_id) if __name__ == '__main__': app.run(debug=True)
<commit_before>import json import functools from flask import Flask, Response from foxgami.red import Story app = Flask(__name__) def return_as_json(inner_f): @functools.wraps(inner_f) def new_f(*args, **kwargs): result = inner_f(*args, **kwargs) return Response(json.dumps( result, indent=4, separators=(', ', ': ') ), mimetype='application/json') return new_f @app.route('/api/stories') @return_as_json def hardcoded_aww(): return Story.find() @app.route('/api/stories/<string:story_id>') def get_story(story_id): return Story.get(story_id) if __name__ == '__main__': app.run(debug=True)<commit_msg>Add Access-Control headers to python<commit_after>import json import functools from flask import Flask, Response from foxgami.red import Story app = Flask(__name__) @app.after_response def add_content_headers(response): response.headers['Access-Control-Allow-Origin'] = '*' return response def return_as_json(inner_f): @functools.wraps(inner_f) def new_f(*args, **kwargs): result = inner_f(*args, **kwargs) return Response(json.dumps( result, indent=4, separators=(', ', ': ') ), mimetype='application/json') return new_f @app.route('/api/stories') @return_as_json def hardcoded_aww(): return Story.find() @app.route('/api/stories/<string:story_id>') def get_story(story_id): return Story.get(story_id) if __name__ == '__main__': app.run(debug=True)
f323676f1d3717ed2c84d06374cffbe2f1882cb4
blimp_boards/notifications/serializers.py
blimp_boards/notifications/serializers.py
from rest_framework import serializers from .models import Notification class NotificationSerializer(serializers.ModelSerializer): target = serializers.Field(source='data.target') action_object = serializers.Field(source='data.action_object') actor = serializers.Field(source='data.sender') timesince = serializers.Field(source='timesince') class Meta: model = Notification fields = ('target', 'action_object', 'actor', 'verb', 'timesince', 'date_created', 'date_modified')
from django.utils.six.moves.urllib import parse from rest_framework import serializers from ..files.utils import sign_s3_url from .models import Notification class NotificationSerializer(serializers.ModelSerializer): target = serializers.SerializerMethodField('get_target_data') action_object = serializers.SerializerMethodField('get_action_object_data') actor = serializers.Field(source='data.sender') timesince = serializers.Field(source='timesince') class Meta: model = Notification fields = ('target', 'action_object', 'actor', 'verb', 'timesince', 'date_created', 'date_modified') def get_action_object_data(self, obj): return self._data_with_signed_urls(obj.data['action_object']) def get_target_data(self, obj): return self._data_with_signed_urls(obj.data['target']) def _data_with_signed_urls(self, data): thumbnail_keys = [ 'thumbnail_sm_path', 'thumbnail_md_path', 'thumbnail_lg_path' ] for key, value in data.items(): if key in thumbnail_keys and value: split_results = list(tuple(parse.urlsplit(value))) split_results[-2] = '' cleaned_url = parse.unquote(parse.urlunsplit(split_results)) data[key] = sign_s3_url(cleaned_url) return data
Fix action object and target thumbnail urls
Fix action object and target thumbnail urls
Python
agpl-3.0
jessamynsmith/boards-backend,jessamynsmith/boards-backend,GetBlimp/boards-backend
from rest_framework import serializers from .models import Notification class NotificationSerializer(serializers.ModelSerializer): target = serializers.Field(source='data.target') action_object = serializers.Field(source='data.action_object') actor = serializers.Field(source='data.sender') timesince = serializers.Field(source='timesince') class Meta: model = Notification fields = ('target', 'action_object', 'actor', 'verb', 'timesince', 'date_created', 'date_modified') Fix action object and target thumbnail urls
from django.utils.six.moves.urllib import parse from rest_framework import serializers from ..files.utils import sign_s3_url from .models import Notification class NotificationSerializer(serializers.ModelSerializer): target = serializers.SerializerMethodField('get_target_data') action_object = serializers.SerializerMethodField('get_action_object_data') actor = serializers.Field(source='data.sender') timesince = serializers.Field(source='timesince') class Meta: model = Notification fields = ('target', 'action_object', 'actor', 'verb', 'timesince', 'date_created', 'date_modified') def get_action_object_data(self, obj): return self._data_with_signed_urls(obj.data['action_object']) def get_target_data(self, obj): return self._data_with_signed_urls(obj.data['target']) def _data_with_signed_urls(self, data): thumbnail_keys = [ 'thumbnail_sm_path', 'thumbnail_md_path', 'thumbnail_lg_path' ] for key, value in data.items(): if key in thumbnail_keys and value: split_results = list(tuple(parse.urlsplit(value))) split_results[-2] = '' cleaned_url = parse.unquote(parse.urlunsplit(split_results)) data[key] = sign_s3_url(cleaned_url) return data
<commit_before>from rest_framework import serializers from .models import Notification class NotificationSerializer(serializers.ModelSerializer): target = serializers.Field(source='data.target') action_object = serializers.Field(source='data.action_object') actor = serializers.Field(source='data.sender') timesince = serializers.Field(source='timesince') class Meta: model = Notification fields = ('target', 'action_object', 'actor', 'verb', 'timesince', 'date_created', 'date_modified') <commit_msg>Fix action object and target thumbnail urls<commit_after>
from django.utils.six.moves.urllib import parse from rest_framework import serializers from ..files.utils import sign_s3_url from .models import Notification class NotificationSerializer(serializers.ModelSerializer): target = serializers.SerializerMethodField('get_target_data') action_object = serializers.SerializerMethodField('get_action_object_data') actor = serializers.Field(source='data.sender') timesince = serializers.Field(source='timesince') class Meta: model = Notification fields = ('target', 'action_object', 'actor', 'verb', 'timesince', 'date_created', 'date_modified') def get_action_object_data(self, obj): return self._data_with_signed_urls(obj.data['action_object']) def get_target_data(self, obj): return self._data_with_signed_urls(obj.data['target']) def _data_with_signed_urls(self, data): thumbnail_keys = [ 'thumbnail_sm_path', 'thumbnail_md_path', 'thumbnail_lg_path' ] for key, value in data.items(): if key in thumbnail_keys and value: split_results = list(tuple(parse.urlsplit(value))) split_results[-2] = '' cleaned_url = parse.unquote(parse.urlunsplit(split_results)) data[key] = sign_s3_url(cleaned_url) return data
from rest_framework import serializers from .models import Notification class NotificationSerializer(serializers.ModelSerializer): target = serializers.Field(source='data.target') action_object = serializers.Field(source='data.action_object') actor = serializers.Field(source='data.sender') timesince = serializers.Field(source='timesince') class Meta: model = Notification fields = ('target', 'action_object', 'actor', 'verb', 'timesince', 'date_created', 'date_modified') Fix action object and target thumbnail urlsfrom django.utils.six.moves.urllib import parse from rest_framework import serializers from ..files.utils import sign_s3_url from .models import Notification class NotificationSerializer(serializers.ModelSerializer): target = serializers.SerializerMethodField('get_target_data') action_object = serializers.SerializerMethodField('get_action_object_data') actor = serializers.Field(source='data.sender') timesince = serializers.Field(source='timesince') class Meta: model = Notification fields = ('target', 'action_object', 'actor', 'verb', 'timesince', 'date_created', 'date_modified') def get_action_object_data(self, obj): return self._data_with_signed_urls(obj.data['action_object']) def get_target_data(self, obj): return self._data_with_signed_urls(obj.data['target']) def _data_with_signed_urls(self, data): thumbnail_keys = [ 'thumbnail_sm_path', 'thumbnail_md_path', 'thumbnail_lg_path' ] for key, value in data.items(): if key in thumbnail_keys and value: split_results = list(tuple(parse.urlsplit(value))) split_results[-2] = '' cleaned_url = parse.unquote(parse.urlunsplit(split_results)) data[key] = sign_s3_url(cleaned_url) return data
<commit_before>from rest_framework import serializers from .models import Notification class NotificationSerializer(serializers.ModelSerializer): target = serializers.Field(source='data.target') action_object = serializers.Field(source='data.action_object') actor = serializers.Field(source='data.sender') timesince = serializers.Field(source='timesince') class Meta: model = Notification fields = ('target', 'action_object', 'actor', 'verb', 'timesince', 'date_created', 'date_modified') <commit_msg>Fix action object and target thumbnail urls<commit_after>from django.utils.six.moves.urllib import parse from rest_framework import serializers from ..files.utils import sign_s3_url from .models import Notification class NotificationSerializer(serializers.ModelSerializer): target = serializers.SerializerMethodField('get_target_data') action_object = serializers.SerializerMethodField('get_action_object_data') actor = serializers.Field(source='data.sender') timesince = serializers.Field(source='timesince') class Meta: model = Notification fields = ('target', 'action_object', 'actor', 'verb', 'timesince', 'date_created', 'date_modified') def get_action_object_data(self, obj): return self._data_with_signed_urls(obj.data['action_object']) def get_target_data(self, obj): return self._data_with_signed_urls(obj.data['target']) def _data_with_signed_urls(self, data): thumbnail_keys = [ 'thumbnail_sm_path', 'thumbnail_md_path', 'thumbnail_lg_path' ] for key, value in data.items(): if key in thumbnail_keys and value: split_results = list(tuple(parse.urlsplit(value))) split_results[-2] = '' cleaned_url = parse.unquote(parse.urlunsplit(split_results)) data[key] = sign_s3_url(cleaned_url) return data
7bfefe50c00d86b55c0620207e9848c97aa28227
rml/units.py
rml/units.py
import numpy as np from scipy.interpolate import PchipInterpolator class UcPoly(): def __init__(self, coef): self.p = np.poly1d(coef) def machine_to_physics(self, machine_value): return self.p(machine_value) def physics_to_machine(self, physics_value): roots = (self.p - physics_value).roots positive_roots = [root for root in roots if root > 0] if len(positive_roots) > 0: return positive_roots[0] else: raise ValueError("No corresponding positive machine value:", roots) class UcPchip(): def __init__(self, x, y): self.x = x self.y = y self.pp = PchipInterpolator(x, y) def machine_to_physics(self, machine_value): return self.pp(machine_value) def physics_to_machine(self, physics_value): pass
import numpy as np from scipy.interpolate import PchipInterpolator class UcPoly(object): def __init__(self, coef): self.p = np.poly1d(coef) def machine_to_physics(self, machine_value): return self.p(machine_value) def physics_to_machine(self, physics_value): roots = (self.p - physics_value).roots positive_roots = [root for root in roots if root > 0] if len(positive_roots) > 0: return positive_roots[0] else: raise ValueError("No corresponding positive machine value:", roots) class UcPchip(object): def __init__(self, x, y): self.x = x self.y = y self.pp = PchipInterpolator(x, y) def machine_to_physics(self, machine_value): return self.pp(machine_value) def physics_to_machine(self, physics_value): pass
Correct the definitions of old-style classes
Correct the definitions of old-style classes
Python
apache-2.0
razvanvasile/RML,willrogers/pml,willrogers/pml
import numpy as np from scipy.interpolate import PchipInterpolator class UcPoly(): def __init__(self, coef): self.p = np.poly1d(coef) def machine_to_physics(self, machine_value): return self.p(machine_value) def physics_to_machine(self, physics_value): roots = (self.p - physics_value).roots positive_roots = [root for root in roots if root > 0] if len(positive_roots) > 0: return positive_roots[0] else: raise ValueError("No corresponding positive machine value:", roots) class UcPchip(): def __init__(self, x, y): self.x = x self.y = y self.pp = PchipInterpolator(x, y) def machine_to_physics(self, machine_value): return self.pp(machine_value) def physics_to_machine(self, physics_value): pass Correct the definitions of old-style classes
import numpy as np from scipy.interpolate import PchipInterpolator class UcPoly(object): def __init__(self, coef): self.p = np.poly1d(coef) def machine_to_physics(self, machine_value): return self.p(machine_value) def physics_to_machine(self, physics_value): roots = (self.p - physics_value).roots positive_roots = [root for root in roots if root > 0] if len(positive_roots) > 0: return positive_roots[0] else: raise ValueError("No corresponding positive machine value:", roots) class UcPchip(object): def __init__(self, x, y): self.x = x self.y = y self.pp = PchipInterpolator(x, y) def machine_to_physics(self, machine_value): return self.pp(machine_value) def physics_to_machine(self, physics_value): pass
<commit_before>import numpy as np from scipy.interpolate import PchipInterpolator class UcPoly(): def __init__(self, coef): self.p = np.poly1d(coef) def machine_to_physics(self, machine_value): return self.p(machine_value) def physics_to_machine(self, physics_value): roots = (self.p - physics_value).roots positive_roots = [root for root in roots if root > 0] if len(positive_roots) > 0: return positive_roots[0] else: raise ValueError("No corresponding positive machine value:", roots) class UcPchip(): def __init__(self, x, y): self.x = x self.y = y self.pp = PchipInterpolator(x, y) def machine_to_physics(self, machine_value): return self.pp(machine_value) def physics_to_machine(self, physics_value): pass <commit_msg>Correct the definitions of old-style classes<commit_after>
import numpy as np from scipy.interpolate import PchipInterpolator class UcPoly(object): def __init__(self, coef): self.p = np.poly1d(coef) def machine_to_physics(self, machine_value): return self.p(machine_value) def physics_to_machine(self, physics_value): roots = (self.p - physics_value).roots positive_roots = [root for root in roots if root > 0] if len(positive_roots) > 0: return positive_roots[0] else: raise ValueError("No corresponding positive machine value:", roots) class UcPchip(object): def __init__(self, x, y): self.x = x self.y = y self.pp = PchipInterpolator(x, y) def machine_to_physics(self, machine_value): return self.pp(machine_value) def physics_to_machine(self, physics_value): pass
import numpy as np from scipy.interpolate import PchipInterpolator class UcPoly(): def __init__(self, coef): self.p = np.poly1d(coef) def machine_to_physics(self, machine_value): return self.p(machine_value) def physics_to_machine(self, physics_value): roots = (self.p - physics_value).roots positive_roots = [root for root in roots if root > 0] if len(positive_roots) > 0: return positive_roots[0] else: raise ValueError("No corresponding positive machine value:", roots) class UcPchip(): def __init__(self, x, y): self.x = x self.y = y self.pp = PchipInterpolator(x, y) def machine_to_physics(self, machine_value): return self.pp(machine_value) def physics_to_machine(self, physics_value): pass Correct the definitions of old-style classesimport numpy as np from scipy.interpolate import PchipInterpolator class UcPoly(object): def __init__(self, coef): self.p = np.poly1d(coef) def machine_to_physics(self, machine_value): return self.p(machine_value) def physics_to_machine(self, physics_value): roots = (self.p - physics_value).roots positive_roots = [root for root in roots if root > 0] if len(positive_roots) > 0: return positive_roots[0] else: raise ValueError("No corresponding positive machine value:", roots) class UcPchip(object): def __init__(self, x, y): self.x = x self.y = y self.pp = PchipInterpolator(x, y) def machine_to_physics(self, machine_value): return self.pp(machine_value) def physics_to_machine(self, physics_value): pass
<commit_before>import numpy as np from scipy.interpolate import PchipInterpolator class UcPoly(): def __init__(self, coef): self.p = np.poly1d(coef) def machine_to_physics(self, machine_value): return self.p(machine_value) def physics_to_machine(self, physics_value): roots = (self.p - physics_value).roots positive_roots = [root for root in roots if root > 0] if len(positive_roots) > 0: return positive_roots[0] else: raise ValueError("No corresponding positive machine value:", roots) class UcPchip(): def __init__(self, x, y): self.x = x self.y = y self.pp = PchipInterpolator(x, y) def machine_to_physics(self, machine_value): return self.pp(machine_value) def physics_to_machine(self, physics_value): pass <commit_msg>Correct the definitions of old-style classes<commit_after>import numpy as np from scipy.interpolate import PchipInterpolator class UcPoly(object): def __init__(self, coef): self.p = np.poly1d(coef) def machine_to_physics(self, machine_value): return self.p(machine_value) def physics_to_machine(self, physics_value): roots = (self.p - physics_value).roots positive_roots = [root for root in roots if root > 0] if len(positive_roots) > 0: return positive_roots[0] else: raise ValueError("No corresponding positive machine value:", roots) class UcPchip(object): def __init__(self, x, y): self.x = x self.y = y self.pp = PchipInterpolator(x, y) def machine_to_physics(self, machine_value): return self.pp(machine_value) def physics_to_machine(self, physics_value): pass
d487e8d74d8bbbadf003cd128f80868cc5651d21
shenfun/optimization/__init__.py
shenfun/optimization/__init__.py
"""Module for optimized functions Some methods performed in Python may be slowing down solvers. In this optimization module we place optimized functions that are to be used instead of default Python methods. Some methods are implemented solely in Cython and only called from within the regular Python modules. """ import os import importlib from functools import wraps from . import cython try: from . import numba except ModuleNotFoundError: numba = None def optimizer(func): """Decorator used to wrap calls to optimized versions of functions.""" from shenfun.config import config mod = config['optimization']['mode'] verbose = config['optimization']['verbose'] if mod.lower() not in ('cython', 'numba'): # Use python function if verbose: print(func.__name__ + ' not optimized') return func mod = importlib.import_module('shenfun.optimization.'+mod.lower()) fun = getattr(mod, func.__name__, func) if verbose: if fun is func: print(fun.__name__ + ' not optimized') @wraps(func) def wrapped_function(*args, **kwargs): u0 = fun(*args, **kwargs) return u0 return wrapped_function
"""Module for optimized functions Some methods performed in Python may be slowing down solvers. In this optimization module we place optimized functions that are to be used instead of default Python methods. Some methods are implemented solely in Cython and only called from within the regular Python modules. """ import os import importlib from functools import wraps from . import cython try: from . import numba except ModuleNotFoundError: numba = None def optimizer(func): """Decorator used to wrap calls to optimized versions of functions.""" from shenfun.config import config mod = 'cython' verbose = False try: mod = config['optimization']['mode'] verbose = config['optimization']['verbose'] except KeyError: pass if mod.lower() not in ('cython', 'numba'): # Use python function if verbose: print(func.__name__ + ' not optimized') return func mod = importlib.import_module('shenfun.optimization.'+mod.lower()) fun = getattr(mod, func.__name__, func) if verbose: if fun is func: print(fun.__name__ + ' not optimized') @wraps(func) def wrapped_function(*args, **kwargs): u0 = fun(*args, **kwargs) return u0 return wrapped_function
Use try clause for config in optimizer
Use try clause for config in optimizer
Python
bsd-2-clause
spectralDNS/shenfun,spectralDNS/shenfun,spectralDNS/shenfun
"""Module for optimized functions Some methods performed in Python may be slowing down solvers. In this optimization module we place optimized functions that are to be used instead of default Python methods. Some methods are implemented solely in Cython and only called from within the regular Python modules. """ import os import importlib from functools import wraps from . import cython try: from . import numba except ModuleNotFoundError: numba = None def optimizer(func): """Decorator used to wrap calls to optimized versions of functions.""" from shenfun.config import config mod = config['optimization']['mode'] verbose = config['optimization']['verbose'] if mod.lower() not in ('cython', 'numba'): # Use python function if verbose: print(func.__name__ + ' not optimized') return func mod = importlib.import_module('shenfun.optimization.'+mod.lower()) fun = getattr(mod, func.__name__, func) if verbose: if fun is func: print(fun.__name__ + ' not optimized') @wraps(func) def wrapped_function(*args, **kwargs): u0 = fun(*args, **kwargs) return u0 return wrapped_function Use try clause for config in optimizer
"""Module for optimized functions Some methods performed in Python may be slowing down solvers. In this optimization module we place optimized functions that are to be used instead of default Python methods. Some methods are implemented solely in Cython and only called from within the regular Python modules. """ import os import importlib from functools import wraps from . import cython try: from . import numba except ModuleNotFoundError: numba = None def optimizer(func): """Decorator used to wrap calls to optimized versions of functions.""" from shenfun.config import config mod = 'cython' verbose = False try: mod = config['optimization']['mode'] verbose = config['optimization']['verbose'] except KeyError: pass if mod.lower() not in ('cython', 'numba'): # Use python function if verbose: print(func.__name__ + ' not optimized') return func mod = importlib.import_module('shenfun.optimization.'+mod.lower()) fun = getattr(mod, func.__name__, func) if verbose: if fun is func: print(fun.__name__ + ' not optimized') @wraps(func) def wrapped_function(*args, **kwargs): u0 = fun(*args, **kwargs) return u0 return wrapped_function
<commit_before>"""Module for optimized functions Some methods performed in Python may be slowing down solvers. In this optimization module we place optimized functions that are to be used instead of default Python methods. Some methods are implemented solely in Cython and only called from within the regular Python modules. """ import os import importlib from functools import wraps from . import cython try: from . import numba except ModuleNotFoundError: numba = None def optimizer(func): """Decorator used to wrap calls to optimized versions of functions.""" from shenfun.config import config mod = config['optimization']['mode'] verbose = config['optimization']['verbose'] if mod.lower() not in ('cython', 'numba'): # Use python function if verbose: print(func.__name__ + ' not optimized') return func mod = importlib.import_module('shenfun.optimization.'+mod.lower()) fun = getattr(mod, func.__name__, func) if verbose: if fun is func: print(fun.__name__ + ' not optimized') @wraps(func) def wrapped_function(*args, **kwargs): u0 = fun(*args, **kwargs) return u0 return wrapped_function <commit_msg>Use try clause for config in optimizer<commit_after>
"""Module for optimized functions Some methods performed in Python may be slowing down solvers. In this optimization module we place optimized functions that are to be used instead of default Python methods. Some methods are implemented solely in Cython and only called from within the regular Python modules. """ import os import importlib from functools import wraps from . import cython try: from . import numba except ModuleNotFoundError: numba = None def optimizer(func): """Decorator used to wrap calls to optimized versions of functions.""" from shenfun.config import config mod = 'cython' verbose = False try: mod = config['optimization']['mode'] verbose = config['optimization']['verbose'] except KeyError: pass if mod.lower() not in ('cython', 'numba'): # Use python function if verbose: print(func.__name__ + ' not optimized') return func mod = importlib.import_module('shenfun.optimization.'+mod.lower()) fun = getattr(mod, func.__name__, func) if verbose: if fun is func: print(fun.__name__ + ' not optimized') @wraps(func) def wrapped_function(*args, **kwargs): u0 = fun(*args, **kwargs) return u0 return wrapped_function
"""Module for optimized functions Some methods performed in Python may be slowing down solvers. In this optimization module we place optimized functions that are to be used instead of default Python methods. Some methods are implemented solely in Cython and only called from within the regular Python modules. """ import os import importlib from functools import wraps from . import cython try: from . import numba except ModuleNotFoundError: numba = None def optimizer(func): """Decorator used to wrap calls to optimized versions of functions.""" from shenfun.config import config mod = config['optimization']['mode'] verbose = config['optimization']['verbose'] if mod.lower() not in ('cython', 'numba'): # Use python function if verbose: print(func.__name__ + ' not optimized') return func mod = importlib.import_module('shenfun.optimization.'+mod.lower()) fun = getattr(mod, func.__name__, func) if verbose: if fun is func: print(fun.__name__ + ' not optimized') @wraps(func) def wrapped_function(*args, **kwargs): u0 = fun(*args, **kwargs) return u0 return wrapped_function Use try clause for config in optimizer"""Module for optimized functions Some methods performed in Python may be slowing down solvers. In this optimization module we place optimized functions that are to be used instead of default Python methods. Some methods are implemented solely in Cython and only called from within the regular Python modules. """ import os import importlib from functools import wraps from . import cython try: from . import numba except ModuleNotFoundError: numba = None def optimizer(func): """Decorator used to wrap calls to optimized versions of functions.""" from shenfun.config import config mod = 'cython' verbose = False try: mod = config['optimization']['mode'] verbose = config['optimization']['verbose'] except KeyError: pass if mod.lower() not in ('cython', 'numba'): # Use python function if verbose: print(func.__name__ + ' not optimized') return func mod = importlib.import_module('shenfun.optimization.'+mod.lower()) fun = getattr(mod, func.__name__, func) if verbose: if fun is func: print(fun.__name__ + ' not optimized') @wraps(func) def wrapped_function(*args, **kwargs): u0 = fun(*args, **kwargs) return u0 return wrapped_function
<commit_before>"""Module for optimized functions Some methods performed in Python may be slowing down solvers. In this optimization module we place optimized functions that are to be used instead of default Python methods. Some methods are implemented solely in Cython and only called from within the regular Python modules. """ import os import importlib from functools import wraps from . import cython try: from . import numba except ModuleNotFoundError: numba = None def optimizer(func): """Decorator used to wrap calls to optimized versions of functions.""" from shenfun.config import config mod = config['optimization']['mode'] verbose = config['optimization']['verbose'] if mod.lower() not in ('cython', 'numba'): # Use python function if verbose: print(func.__name__ + ' not optimized') return func mod = importlib.import_module('shenfun.optimization.'+mod.lower()) fun = getattr(mod, func.__name__, func) if verbose: if fun is func: print(fun.__name__ + ' not optimized') @wraps(func) def wrapped_function(*args, **kwargs): u0 = fun(*args, **kwargs) return u0 return wrapped_function <commit_msg>Use try clause for config in optimizer<commit_after>"""Module for optimized functions Some methods performed in Python may be slowing down solvers. In this optimization module we place optimized functions that are to be used instead of default Python methods. Some methods are implemented solely in Cython and only called from within the regular Python modules. """ import os import importlib from functools import wraps from . import cython try: from . import numba except ModuleNotFoundError: numba = None def optimizer(func): """Decorator used to wrap calls to optimized versions of functions.""" from shenfun.config import config mod = 'cython' verbose = False try: mod = config['optimization']['mode'] verbose = config['optimization']['verbose'] except KeyError: pass if mod.lower() not in ('cython', 'numba'): # Use python function if verbose: print(func.__name__ + ' not optimized') return func mod = importlib.import_module('shenfun.optimization.'+mod.lower()) fun = getattr(mod, func.__name__, func) if verbose: if fun is func: print(fun.__name__ + ' not optimized') @wraps(func) def wrapped_function(*args, **kwargs): u0 = fun(*args, **kwargs) return u0 return wrapped_function
7c3c0fac58822bcfa7bbd69de5ce46b36f84740c
regulations/generator/link_flattener.py
regulations/generator/link_flattener.py
import re # <a> followed by another <a> without any intervening </a>s link_inside_link_regex = re.compile( ur"(?P<outer_link><a ((?!</a>).)*)(<a ((?!</a>).)*>" ur"(?P<internal_content>((?!</a>).)*)</a>)", re.IGNORECASE | re.DOTALL) def flatten_links(text): """ Fix <a> elements that have embedded <a> elements by replacing the internal <a> element with its content. Assumes that the text does not span multiple lines and that the <a> tags are lowercase. """ while True: text, sub_count = link_inside_link_regex.subn( ur"\g<outer_link>\g<internal_content>", text) if sub_count == 0: return text # Return only when no more subs possible
import re # <a> followed by another <a> without any intervening </a>s # outer_link - partial outer element up to the inner link # inner_content - content of the inner_link link_inside_link_regex = re.compile( ur"(?P<outer_link><a ((?!</a>).)*)<a .*?>(?P<inner_content>.*?)</a>", re.IGNORECASE | re.DOTALL) def flatten_links(text): """ Fix <a> elements that have embedded <a> elements by replacing the internal <a> element with its content. """ while True: text, sub_count = link_inside_link_regex.subn( ur"\g<outer_link>\g<inner_content>", text) if sub_count == 0: return text
Simplify regex using non-greedy qualifier
Simplify regex using non-greedy qualifier
Python
cc0-1.0
18F/regulations-site,tadhg-ohiggins/regulations-site,eregs/regulations-site,18F/regulations-site,tadhg-ohiggins/regulations-site,tadhg-ohiggins/regulations-site,eregs/regulations-site,18F/regulations-site,18F/regulations-site,eregs/regulations-site,eregs/regulations-site,tadhg-ohiggins/regulations-site
import re # <a> followed by another <a> without any intervening </a>s link_inside_link_regex = re.compile( ur"(?P<outer_link><a ((?!</a>).)*)(<a ((?!</a>).)*>" ur"(?P<internal_content>((?!</a>).)*)</a>)", re.IGNORECASE | re.DOTALL) def flatten_links(text): """ Fix <a> elements that have embedded <a> elements by replacing the internal <a> element with its content. Assumes that the text does not span multiple lines and that the <a> tags are lowercase. """ while True: text, sub_count = link_inside_link_regex.subn( ur"\g<outer_link>\g<internal_content>", text) if sub_count == 0: return text # Return only when no more subs possible Simplify regex using non-greedy qualifier
import re # <a> followed by another <a> without any intervening </a>s # outer_link - partial outer element up to the inner link # inner_content - content of the inner_link link_inside_link_regex = re.compile( ur"(?P<outer_link><a ((?!</a>).)*)<a .*?>(?P<inner_content>.*?)</a>", re.IGNORECASE | re.DOTALL) def flatten_links(text): """ Fix <a> elements that have embedded <a> elements by replacing the internal <a> element with its content. """ while True: text, sub_count = link_inside_link_regex.subn( ur"\g<outer_link>\g<inner_content>", text) if sub_count == 0: return text
<commit_before>import re # <a> followed by another <a> without any intervening </a>s link_inside_link_regex = re.compile( ur"(?P<outer_link><a ((?!</a>).)*)(<a ((?!</a>).)*>" ur"(?P<internal_content>((?!</a>).)*)</a>)", re.IGNORECASE | re.DOTALL) def flatten_links(text): """ Fix <a> elements that have embedded <a> elements by replacing the internal <a> element with its content. Assumes that the text does not span multiple lines and that the <a> tags are lowercase. """ while True: text, sub_count = link_inside_link_regex.subn( ur"\g<outer_link>\g<internal_content>", text) if sub_count == 0: return text # Return only when no more subs possible <commit_msg>Simplify regex using non-greedy qualifier<commit_after>
import re # <a> followed by another <a> without any intervening </a>s # outer_link - partial outer element up to the inner link # inner_content - content of the inner_link link_inside_link_regex = re.compile( ur"(?P<outer_link><a ((?!</a>).)*)<a .*?>(?P<inner_content>.*?)</a>", re.IGNORECASE | re.DOTALL) def flatten_links(text): """ Fix <a> elements that have embedded <a> elements by replacing the internal <a> element with its content. """ while True: text, sub_count = link_inside_link_regex.subn( ur"\g<outer_link>\g<inner_content>", text) if sub_count == 0: return text
import re # <a> followed by another <a> without any intervening </a>s link_inside_link_regex = re.compile( ur"(?P<outer_link><a ((?!</a>).)*)(<a ((?!</a>).)*>" ur"(?P<internal_content>((?!</a>).)*)</a>)", re.IGNORECASE | re.DOTALL) def flatten_links(text): """ Fix <a> elements that have embedded <a> elements by replacing the internal <a> element with its content. Assumes that the text does not span multiple lines and that the <a> tags are lowercase. """ while True: text, sub_count = link_inside_link_regex.subn( ur"\g<outer_link>\g<internal_content>", text) if sub_count == 0: return text # Return only when no more subs possible Simplify regex using non-greedy qualifierimport re # <a> followed by another <a> without any intervening </a>s # outer_link - partial outer element up to the inner link # inner_content - content of the inner_link link_inside_link_regex = re.compile( ur"(?P<outer_link><a ((?!</a>).)*)<a .*?>(?P<inner_content>.*?)</a>", re.IGNORECASE | re.DOTALL) def flatten_links(text): """ Fix <a> elements that have embedded <a> elements by replacing the internal <a> element with its content. """ while True: text, sub_count = link_inside_link_regex.subn( ur"\g<outer_link>\g<inner_content>", text) if sub_count == 0: return text
<commit_before>import re # <a> followed by another <a> without any intervening </a>s link_inside_link_regex = re.compile( ur"(?P<outer_link><a ((?!</a>).)*)(<a ((?!</a>).)*>" ur"(?P<internal_content>((?!</a>).)*)</a>)", re.IGNORECASE | re.DOTALL) def flatten_links(text): """ Fix <a> elements that have embedded <a> elements by replacing the internal <a> element with its content. Assumes that the text does not span multiple lines and that the <a> tags are lowercase. """ while True: text, sub_count = link_inside_link_regex.subn( ur"\g<outer_link>\g<internal_content>", text) if sub_count == 0: return text # Return only when no more subs possible <commit_msg>Simplify regex using non-greedy qualifier<commit_after>import re # <a> followed by another <a> without any intervening </a>s # outer_link - partial outer element up to the inner link # inner_content - content of the inner_link link_inside_link_regex = re.compile( ur"(?P<outer_link><a ((?!</a>).)*)<a .*?>(?P<inner_content>.*?)</a>", re.IGNORECASE | re.DOTALL) def flatten_links(text): """ Fix <a> elements that have embedded <a> elements by replacing the internal <a> element with its content. """ while True: text, sub_count = link_inside_link_regex.subn( ur"\g<outer_link>\g<inner_content>", text) if sub_count == 0: return text
accaf0bc0ab81d12927b55fdfd24ad75907b772f
runserver.py
runserver.py
from anser import Anser server = Anser(__name__, debug=True) @server.action('default') def action_a(message, address): print message['type'] print message['body'] server.run()
from anser import Anser server = Anser(__name__, debug=True) @server.action('default') def action_a(message, address): print "{0} - {1}".format(address, message) server.run()
Clean the output of the demo server
Clean the output of the demo server
Python
mit
iconpin/anser
from anser import Anser server = Anser(__name__, debug=True) @server.action('default') def action_a(message, address): print message['type'] print message['body'] server.run()Clean the output of the demo server
from anser import Anser server = Anser(__name__, debug=True) @server.action('default') def action_a(message, address): print "{0} - {1}".format(address, message) server.run()
<commit_before>from anser import Anser server = Anser(__name__, debug=True) @server.action('default') def action_a(message, address): print message['type'] print message['body'] server.run()<commit_msg>Clean the output of the demo server<commit_after>
from anser import Anser server = Anser(__name__, debug=True) @server.action('default') def action_a(message, address): print "{0} - {1}".format(address, message) server.run()
from anser import Anser server = Anser(__name__, debug=True) @server.action('default') def action_a(message, address): print message['type'] print message['body'] server.run()Clean the output of the demo serverfrom anser import Anser server = Anser(__name__, debug=True) @server.action('default') def action_a(message, address): print "{0} - {1}".format(address, message) server.run()
<commit_before>from anser import Anser server = Anser(__name__, debug=True) @server.action('default') def action_a(message, address): print message['type'] print message['body'] server.run()<commit_msg>Clean the output of the demo server<commit_after>from anser import Anser server = Anser(__name__, debug=True) @server.action('default') def action_a(message, address): print "{0} - {1}".format(address, message) server.run()
37b8cf1af7818fe78b31ed25622f3f91805ade01
test_bert_trainer.py
test_bert_trainer.py
import unittest import time import shutil import pandas as pd from bert_trainer import BERTTrainer from utils import * class TestBERT(unittest.TestCase): def __init__(self, *args, **kwargs): super(TestBERT, self).__init__(*args, **kwargs) self.output_dir = 'test_{}'.format(str(int(time.time()))) self.trainer = BERTTrainer(output_dir=self.output_dir) self.data = pd.DataFrame({ 'abstract': ['test one', 'test two', 'test three'] * 5, 'section': ['U.S.', 'Arts', 'U.S.'] * 5, }) def train_model(self): self.trainer.train(self.data['abstract'], self.data['section']) def test_train(self): self.train_model() shutil.rmtree(self.output_dir) def test_train_and_test(self): self.train_model() results = self.trainer.evaluate(self.data['abstract'], self.data['section']) results2 = self.trainer.evaluate(self.data['abstract'], self.data['section']) eval_acc1, eval_acc2 = results['eval_accuracy'], results2['eval_accuracy'] self.assertEqual(eval_acc1, eval_acc2) loss1, loss2 = results['loss'], results2['loss'] self.assertEqual(eval_acc1, eval_acc2) shutil.rmtree(self.output_dir) def test_train_and_predict(self): self.train_model() input_sentences = [ "test four", "test one", ] * 5 preds = self.trainer.predict(input_sentences) shutil.rmtree(self.output_dir) if __name__ == '__main__': unittest.main()
import unittest import time import shutil import pandas as pd from bert_trainer import BERTTrainer from utils import * class TestBERT(unittest.TestCase): def __init__(self, *args, **kwargs): super(TestBERT, self).__init__(*args, **kwargs) self.output_dir = 'test_{}'.format(str(int(time.time()))) self.trainer = BERTTrainer(output_dir=self.output_dir) self.data = pd.DataFrame({ 'abstract': ['test one', 'test two', 'test three'] * 5, 'section': ['U.S.', 'Arts', 'U.S.'] * 5, }) def train_model(self): self.trainer.train(self.data['abstract'], self.data['section']) def test_train(self): self.train_model() shutil.rmtree(self.output_dir) def test_train_and_predict(self): self.train_model() input_sentences = [ "test four", "test one", ] * 5 preds = self.trainer.predict(input_sentences) shutil.rmtree(self.output_dir) if __name__ == '__main__': unittest.main()
Fix merge conflict in bert_trainer_example.py
Fix merge conflict in bert_trainer_example.py
Python
apache-2.0
googleinterns/smart-news-query-embeddings,googleinterns/smart-news-query-embeddings
import unittest import time import shutil import pandas as pd from bert_trainer import BERTTrainer from utils import * class TestBERT(unittest.TestCase): def __init__(self, *args, **kwargs): super(TestBERT, self).__init__(*args, **kwargs) self.output_dir = 'test_{}'.format(str(int(time.time()))) self.trainer = BERTTrainer(output_dir=self.output_dir) self.data = pd.DataFrame({ 'abstract': ['test one', 'test two', 'test three'] * 5, 'section': ['U.S.', 'Arts', 'U.S.'] * 5, }) def train_model(self): self.trainer.train(self.data['abstract'], self.data['section']) def test_train(self): self.train_model() shutil.rmtree(self.output_dir) def test_train_and_test(self): self.train_model() results = self.trainer.evaluate(self.data['abstract'], self.data['section']) results2 = self.trainer.evaluate(self.data['abstract'], self.data['section']) eval_acc1, eval_acc2 = results['eval_accuracy'], results2['eval_accuracy'] self.assertEqual(eval_acc1, eval_acc2) loss1, loss2 = results['loss'], results2['loss'] self.assertEqual(eval_acc1, eval_acc2) shutil.rmtree(self.output_dir) def test_train_and_predict(self): self.train_model() input_sentences = [ "test four", "test one", ] * 5 preds = self.trainer.predict(input_sentences) shutil.rmtree(self.output_dir) if __name__ == '__main__': unittest.main() Fix merge conflict in bert_trainer_example.py
import unittest import time import shutil import pandas as pd from bert_trainer import BERTTrainer from utils import * class TestBERT(unittest.TestCase): def __init__(self, *args, **kwargs): super(TestBERT, self).__init__(*args, **kwargs) self.output_dir = 'test_{}'.format(str(int(time.time()))) self.trainer = BERTTrainer(output_dir=self.output_dir) self.data = pd.DataFrame({ 'abstract': ['test one', 'test two', 'test three'] * 5, 'section': ['U.S.', 'Arts', 'U.S.'] * 5, }) def train_model(self): self.trainer.train(self.data['abstract'], self.data['section']) def test_train(self): self.train_model() shutil.rmtree(self.output_dir) def test_train_and_predict(self): self.train_model() input_sentences = [ "test four", "test one", ] * 5 preds = self.trainer.predict(input_sentences) shutil.rmtree(self.output_dir) if __name__ == '__main__': unittest.main()
<commit_before>import unittest import time import shutil import pandas as pd from bert_trainer import BERTTrainer from utils import * class TestBERT(unittest.TestCase): def __init__(self, *args, **kwargs): super(TestBERT, self).__init__(*args, **kwargs) self.output_dir = 'test_{}'.format(str(int(time.time()))) self.trainer = BERTTrainer(output_dir=self.output_dir) self.data = pd.DataFrame({ 'abstract': ['test one', 'test two', 'test three'] * 5, 'section': ['U.S.', 'Arts', 'U.S.'] * 5, }) def train_model(self): self.trainer.train(self.data['abstract'], self.data['section']) def test_train(self): self.train_model() shutil.rmtree(self.output_dir) def test_train_and_test(self): self.train_model() results = self.trainer.evaluate(self.data['abstract'], self.data['section']) results2 = self.trainer.evaluate(self.data['abstract'], self.data['section']) eval_acc1, eval_acc2 = results['eval_accuracy'], results2['eval_accuracy'] self.assertEqual(eval_acc1, eval_acc2) loss1, loss2 = results['loss'], results2['loss'] self.assertEqual(eval_acc1, eval_acc2) shutil.rmtree(self.output_dir) def test_train_and_predict(self): self.train_model() input_sentences = [ "test four", "test one", ] * 5 preds = self.trainer.predict(input_sentences) shutil.rmtree(self.output_dir) if __name__ == '__main__': unittest.main() <commit_msg>Fix merge conflict in bert_trainer_example.py<commit_after>
import unittest import time import shutil import pandas as pd from bert_trainer import BERTTrainer from utils import * class TestBERT(unittest.TestCase): def __init__(self, *args, **kwargs): super(TestBERT, self).__init__(*args, **kwargs) self.output_dir = 'test_{}'.format(str(int(time.time()))) self.trainer = BERTTrainer(output_dir=self.output_dir) self.data = pd.DataFrame({ 'abstract': ['test one', 'test two', 'test three'] * 5, 'section': ['U.S.', 'Arts', 'U.S.'] * 5, }) def train_model(self): self.trainer.train(self.data['abstract'], self.data['section']) def test_train(self): self.train_model() shutil.rmtree(self.output_dir) def test_train_and_predict(self): self.train_model() input_sentences = [ "test four", "test one", ] * 5 preds = self.trainer.predict(input_sentences) shutil.rmtree(self.output_dir) if __name__ == '__main__': unittest.main()
import unittest import time import shutil import pandas as pd from bert_trainer import BERTTrainer from utils import * class TestBERT(unittest.TestCase): def __init__(self, *args, **kwargs): super(TestBERT, self).__init__(*args, **kwargs) self.output_dir = 'test_{}'.format(str(int(time.time()))) self.trainer = BERTTrainer(output_dir=self.output_dir) self.data = pd.DataFrame({ 'abstract': ['test one', 'test two', 'test three'] * 5, 'section': ['U.S.', 'Arts', 'U.S.'] * 5, }) def train_model(self): self.trainer.train(self.data['abstract'], self.data['section']) def test_train(self): self.train_model() shutil.rmtree(self.output_dir) def test_train_and_test(self): self.train_model() results = self.trainer.evaluate(self.data['abstract'], self.data['section']) results2 = self.trainer.evaluate(self.data['abstract'], self.data['section']) eval_acc1, eval_acc2 = results['eval_accuracy'], results2['eval_accuracy'] self.assertEqual(eval_acc1, eval_acc2) loss1, loss2 = results['loss'], results2['loss'] self.assertEqual(eval_acc1, eval_acc2) shutil.rmtree(self.output_dir) def test_train_and_predict(self): self.train_model() input_sentences = [ "test four", "test one", ] * 5 preds = self.trainer.predict(input_sentences) shutil.rmtree(self.output_dir) if __name__ == '__main__': unittest.main() Fix merge conflict in bert_trainer_example.pyimport unittest import time import shutil import pandas as pd from bert_trainer import BERTTrainer from utils import * class TestBERT(unittest.TestCase): def __init__(self, *args, **kwargs): super(TestBERT, self).__init__(*args, **kwargs) self.output_dir = 'test_{}'.format(str(int(time.time()))) self.trainer = BERTTrainer(output_dir=self.output_dir) self.data = pd.DataFrame({ 'abstract': ['test one', 'test two', 'test three'] * 5, 'section': ['U.S.', 'Arts', 'U.S.'] * 5, }) def train_model(self): self.trainer.train(self.data['abstract'], self.data['section']) def test_train(self): self.train_model() shutil.rmtree(self.output_dir) def test_train_and_predict(self): self.train_model() input_sentences = [ "test four", "test one", ] * 5 preds = self.trainer.predict(input_sentences) shutil.rmtree(self.output_dir) if __name__ == '__main__': unittest.main()
<commit_before>import unittest import time import shutil import pandas as pd from bert_trainer import BERTTrainer from utils import * class TestBERT(unittest.TestCase): def __init__(self, *args, **kwargs): super(TestBERT, self).__init__(*args, **kwargs) self.output_dir = 'test_{}'.format(str(int(time.time()))) self.trainer = BERTTrainer(output_dir=self.output_dir) self.data = pd.DataFrame({ 'abstract': ['test one', 'test two', 'test three'] * 5, 'section': ['U.S.', 'Arts', 'U.S.'] * 5, }) def train_model(self): self.trainer.train(self.data['abstract'], self.data['section']) def test_train(self): self.train_model() shutil.rmtree(self.output_dir) def test_train_and_test(self): self.train_model() results = self.trainer.evaluate(self.data['abstract'], self.data['section']) results2 = self.trainer.evaluate(self.data['abstract'], self.data['section']) eval_acc1, eval_acc2 = results['eval_accuracy'], results2['eval_accuracy'] self.assertEqual(eval_acc1, eval_acc2) loss1, loss2 = results['loss'], results2['loss'] self.assertEqual(eval_acc1, eval_acc2) shutil.rmtree(self.output_dir) def test_train_and_predict(self): self.train_model() input_sentences = [ "test four", "test one", ] * 5 preds = self.trainer.predict(input_sentences) shutil.rmtree(self.output_dir) if __name__ == '__main__': unittest.main() <commit_msg>Fix merge conflict in bert_trainer_example.py<commit_after>import unittest import time import shutil import pandas as pd from bert_trainer import BERTTrainer from utils import * class TestBERT(unittest.TestCase): def __init__(self, *args, **kwargs): super(TestBERT, self).__init__(*args, **kwargs) self.output_dir = 'test_{}'.format(str(int(time.time()))) self.trainer = BERTTrainer(output_dir=self.output_dir) self.data = pd.DataFrame({ 'abstract': ['test one', 'test two', 'test three'] * 5, 'section': ['U.S.', 'Arts', 'U.S.'] * 5, }) def train_model(self): self.trainer.train(self.data['abstract'], self.data['section']) def test_train(self): self.train_model() shutil.rmtree(self.output_dir) def test_train_and_predict(self): self.train_model() input_sentences = [ "test four", "test one", ] * 5 preds = self.trainer.predict(input_sentences) shutil.rmtree(self.output_dir) if __name__ == '__main__': unittest.main()
ceb123e78b4d15c0cfe30198aa3fbbe71603472d
project/forms.py
project/forms.py
#! coding: utf-8 from django import forms from django.utils.translation import ugettext_lazy as _ from django.contrib.auth import authenticate class LoginForm(forms.Form): username = forms.CharField(label=_('Naudotojo vardas'), max_length=100, help_text=_('VU MIF uosis.mif.vu.lt serverio.')) password = forms.CharField(label=_(u'Slaptažodis'), max_length=128, widget=forms.PasswordInput(render_value=False)) def clean(self): cleaned_data = super(LoginForm, self).clean() if self.errors: return cleaned_data user = authenticate(**cleaned_data) if not user: raise forms.ValidationError(_(u'Naudotojo vardas arba slaptažodis ' 'yra neteisingi')) return {'user': user}
#! coding: utf-8 from django import forms from django.utils.translation import ugettext_lazy as _ from django.contrib.auth import authenticate class LoginForm(forms.Form): username = forms.CharField(label=_('Naudotojo vardas'), max_length=100, help_text=_('VU MIF uosis.mif.vu.lt serverio.')) password = forms.CharField(label=_(u'Slaptažodis'), max_length=128, widget=forms.PasswordInput(render_value=False)) def clean(self): cleaned_data = super(LoginForm, self).clean() if self.errors: return cleaned_data user = authenticate(**cleaned_data) if not user: raise forms.ValidationError(_(u'Naudotojo vardas arba slaptažodis ' 'yra neteisingi')) cleaned_data['user'] = user return cleaned_data
Update login form clean method to return full cleaned data.
Update login form clean method to return full cleaned data.
Python
agpl-3.0
InScience/DAMIS-old,InScience/DAMIS-old
#! coding: utf-8 from django import forms from django.utils.translation import ugettext_lazy as _ from django.contrib.auth import authenticate class LoginForm(forms.Form): username = forms.CharField(label=_('Naudotojo vardas'), max_length=100, help_text=_('VU MIF uosis.mif.vu.lt serverio.')) password = forms.CharField(label=_(u'Slaptažodis'), max_length=128, widget=forms.PasswordInput(render_value=False)) def clean(self): cleaned_data = super(LoginForm, self).clean() if self.errors: return cleaned_data user = authenticate(**cleaned_data) if not user: raise forms.ValidationError(_(u'Naudotojo vardas arba slaptažodis ' 'yra neteisingi')) return {'user': user} Update login form clean method to return full cleaned data.
#! coding: utf-8 from django import forms from django.utils.translation import ugettext_lazy as _ from django.contrib.auth import authenticate class LoginForm(forms.Form): username = forms.CharField(label=_('Naudotojo vardas'), max_length=100, help_text=_('VU MIF uosis.mif.vu.lt serverio.')) password = forms.CharField(label=_(u'Slaptažodis'), max_length=128, widget=forms.PasswordInput(render_value=False)) def clean(self): cleaned_data = super(LoginForm, self).clean() if self.errors: return cleaned_data user = authenticate(**cleaned_data) if not user: raise forms.ValidationError(_(u'Naudotojo vardas arba slaptažodis ' 'yra neteisingi')) cleaned_data['user'] = user return cleaned_data
<commit_before>#! coding: utf-8 from django import forms from django.utils.translation import ugettext_lazy as _ from django.contrib.auth import authenticate class LoginForm(forms.Form): username = forms.CharField(label=_('Naudotojo vardas'), max_length=100, help_text=_('VU MIF uosis.mif.vu.lt serverio.')) password = forms.CharField(label=_(u'Slaptažodis'), max_length=128, widget=forms.PasswordInput(render_value=False)) def clean(self): cleaned_data = super(LoginForm, self).clean() if self.errors: return cleaned_data user = authenticate(**cleaned_data) if not user: raise forms.ValidationError(_(u'Naudotojo vardas arba slaptažodis ' 'yra neteisingi')) return {'user': user} <commit_msg>Update login form clean method to return full cleaned data.<commit_after>
#! coding: utf-8 from django import forms from django.utils.translation import ugettext_lazy as _ from django.contrib.auth import authenticate class LoginForm(forms.Form): username = forms.CharField(label=_('Naudotojo vardas'), max_length=100, help_text=_('VU MIF uosis.mif.vu.lt serverio.')) password = forms.CharField(label=_(u'Slaptažodis'), max_length=128, widget=forms.PasswordInput(render_value=False)) def clean(self): cleaned_data = super(LoginForm, self).clean() if self.errors: return cleaned_data user = authenticate(**cleaned_data) if not user: raise forms.ValidationError(_(u'Naudotojo vardas arba slaptažodis ' 'yra neteisingi')) cleaned_data['user'] = user return cleaned_data
#! coding: utf-8 from django import forms from django.utils.translation import ugettext_lazy as _ from django.contrib.auth import authenticate class LoginForm(forms.Form): username = forms.CharField(label=_('Naudotojo vardas'), max_length=100, help_text=_('VU MIF uosis.mif.vu.lt serverio.')) password = forms.CharField(label=_(u'Slaptažodis'), max_length=128, widget=forms.PasswordInput(render_value=False)) def clean(self): cleaned_data = super(LoginForm, self).clean() if self.errors: return cleaned_data user = authenticate(**cleaned_data) if not user: raise forms.ValidationError(_(u'Naudotojo vardas arba slaptažodis ' 'yra neteisingi')) return {'user': user} Update login form clean method to return full cleaned data.#! coding: utf-8 from django import forms from django.utils.translation import ugettext_lazy as _ from django.contrib.auth import authenticate class LoginForm(forms.Form): username = forms.CharField(label=_('Naudotojo vardas'), max_length=100, help_text=_('VU MIF uosis.mif.vu.lt serverio.')) password = forms.CharField(label=_(u'Slaptažodis'), max_length=128, widget=forms.PasswordInput(render_value=False)) def clean(self): cleaned_data = super(LoginForm, self).clean() if self.errors: return cleaned_data user = authenticate(**cleaned_data) if not user: raise forms.ValidationError(_(u'Naudotojo vardas arba slaptažodis ' 'yra neteisingi')) cleaned_data['user'] = user return cleaned_data
<commit_before>#! coding: utf-8 from django import forms from django.utils.translation import ugettext_lazy as _ from django.contrib.auth import authenticate class LoginForm(forms.Form): username = forms.CharField(label=_('Naudotojo vardas'), max_length=100, help_text=_('VU MIF uosis.mif.vu.lt serverio.')) password = forms.CharField(label=_(u'Slaptažodis'), max_length=128, widget=forms.PasswordInput(render_value=False)) def clean(self): cleaned_data = super(LoginForm, self).clean() if self.errors: return cleaned_data user = authenticate(**cleaned_data) if not user: raise forms.ValidationError(_(u'Naudotojo vardas arba slaptažodis ' 'yra neteisingi')) return {'user': user} <commit_msg>Update login form clean method to return full cleaned data.<commit_after>#! coding: utf-8 from django import forms from django.utils.translation import ugettext_lazy as _ from django.contrib.auth import authenticate class LoginForm(forms.Form): username = forms.CharField(label=_('Naudotojo vardas'), max_length=100, help_text=_('VU MIF uosis.mif.vu.lt serverio.')) password = forms.CharField(label=_(u'Slaptažodis'), max_length=128, widget=forms.PasswordInput(render_value=False)) def clean(self): cleaned_data = super(LoginForm, self).clean() if self.errors: return cleaned_data user = authenticate(**cleaned_data) if not user: raise forms.ValidationError(_(u'Naudotojo vardas arba slaptažodis ' 'yra neteisingi')) cleaned_data['user'] = user return cleaned_data
d38180f627cf421268f64d94e0eec36a19573754
test_runner/utils.py
test_runner/utils.py
import logging import os import uuid from contextlib import contextmanager from subprocess import check_call, CalledProcessError LOG = logging.getLogger(__name__) def touch(directory, filename=None): file_path = os.path.join(directory, filename) if os.path.exists(file_path): os.utime(file_path, None) else: os.makedirs(directory) if filename is not None: open(file_path, 'a').close() return file_path def run_cmd(command, **kwargs): """ Runs a command and returns an array of its results :param command: String of a command to run within a shell :returns: Dictionary with keys relating to the execution's success """ if kwargs['cwd']: cwd = kwargs['cwd'] try: ret = check_call(command, shell=True, cwd=cwd) return {'success': True, 'return': ret, 'exception': None} except CalledProcessError as exc: return {'success': False, 'return': None, 'exception': exc, 'command': command} @contextmanager def cleanup(stage): try: yield except (Exception, KeyboardInterrupt): LOG.exception('Run failed') finally: stage.destroy() def rand_name(prefix): return '{}-{}'.format(prefix, str(uuid.uuid4()))
import logging import os import uuid from contextlib import contextmanager from subprocess import check_call, CalledProcessError LOG = logging.getLogger(__name__) def touch(directory, filename=None): file_path = os.path.join(directory, filename) if os.path.exists(file_path): os.utime(file_path, None) else: os.makedirs(directory) if filename is not None: open(file_path, 'a').close() return file_path def run_cmd(command, **kwargs): """ Runs a command and returns an array of its results :param command: String of a command to run within a shell :returns: Dictionary with keys relating to the execution's success """ cwd = kwargs.get('cwd', None) try: ret = check_call(command, shell=True, cwd=cwd) return {'success': True, 'return': ret, 'exception': None} except CalledProcessError as exc: return {'success': False, 'return': None, 'exception': exc, 'command': command} @contextmanager def cleanup(stage): try: yield except (Exception, KeyboardInterrupt): LOG.exception('Run failed') finally: stage.destroy() def rand_name(prefix): return '{}-{}'.format(prefix, str(uuid.uuid4()))
Add fallback to 'cwd' if not defined
Add fallback to 'cwd' if not defined
Python
mit
rcbops-qa/test_runner
import logging import os import uuid from contextlib import contextmanager from subprocess import check_call, CalledProcessError LOG = logging.getLogger(__name__) def touch(directory, filename=None): file_path = os.path.join(directory, filename) if os.path.exists(file_path): os.utime(file_path, None) else: os.makedirs(directory) if filename is not None: open(file_path, 'a').close() return file_path def run_cmd(command, **kwargs): """ Runs a command and returns an array of its results :param command: String of a command to run within a shell :returns: Dictionary with keys relating to the execution's success """ if kwargs['cwd']: cwd = kwargs['cwd'] try: ret = check_call(command, shell=True, cwd=cwd) return {'success': True, 'return': ret, 'exception': None} except CalledProcessError as exc: return {'success': False, 'return': None, 'exception': exc, 'command': command} @contextmanager def cleanup(stage): try: yield except (Exception, KeyboardInterrupt): LOG.exception('Run failed') finally: stage.destroy() def rand_name(prefix): return '{}-{}'.format(prefix, str(uuid.uuid4())) Add fallback to 'cwd' if not defined
import logging import os import uuid from contextlib import contextmanager from subprocess import check_call, CalledProcessError LOG = logging.getLogger(__name__) def touch(directory, filename=None): file_path = os.path.join(directory, filename) if os.path.exists(file_path): os.utime(file_path, None) else: os.makedirs(directory) if filename is not None: open(file_path, 'a').close() return file_path def run_cmd(command, **kwargs): """ Runs a command and returns an array of its results :param command: String of a command to run within a shell :returns: Dictionary with keys relating to the execution's success """ cwd = kwargs.get('cwd', None) try: ret = check_call(command, shell=True, cwd=cwd) return {'success': True, 'return': ret, 'exception': None} except CalledProcessError as exc: return {'success': False, 'return': None, 'exception': exc, 'command': command} @contextmanager def cleanup(stage): try: yield except (Exception, KeyboardInterrupt): LOG.exception('Run failed') finally: stage.destroy() def rand_name(prefix): return '{}-{}'.format(prefix, str(uuid.uuid4()))
<commit_before>import logging import os import uuid from contextlib import contextmanager from subprocess import check_call, CalledProcessError LOG = logging.getLogger(__name__) def touch(directory, filename=None): file_path = os.path.join(directory, filename) if os.path.exists(file_path): os.utime(file_path, None) else: os.makedirs(directory) if filename is not None: open(file_path, 'a').close() return file_path def run_cmd(command, **kwargs): """ Runs a command and returns an array of its results :param command: String of a command to run within a shell :returns: Dictionary with keys relating to the execution's success """ if kwargs['cwd']: cwd = kwargs['cwd'] try: ret = check_call(command, shell=True, cwd=cwd) return {'success': True, 'return': ret, 'exception': None} except CalledProcessError as exc: return {'success': False, 'return': None, 'exception': exc, 'command': command} @contextmanager def cleanup(stage): try: yield except (Exception, KeyboardInterrupt): LOG.exception('Run failed') finally: stage.destroy() def rand_name(prefix): return '{}-{}'.format(prefix, str(uuid.uuid4())) <commit_msg>Add fallback to 'cwd' if not defined<commit_after>
import logging import os import uuid from contextlib import contextmanager from subprocess import check_call, CalledProcessError LOG = logging.getLogger(__name__) def touch(directory, filename=None): file_path = os.path.join(directory, filename) if os.path.exists(file_path): os.utime(file_path, None) else: os.makedirs(directory) if filename is not None: open(file_path, 'a').close() return file_path def run_cmd(command, **kwargs): """ Runs a command and returns an array of its results :param command: String of a command to run within a shell :returns: Dictionary with keys relating to the execution's success """ cwd = kwargs.get('cwd', None) try: ret = check_call(command, shell=True, cwd=cwd) return {'success': True, 'return': ret, 'exception': None} except CalledProcessError as exc: return {'success': False, 'return': None, 'exception': exc, 'command': command} @contextmanager def cleanup(stage): try: yield except (Exception, KeyboardInterrupt): LOG.exception('Run failed') finally: stage.destroy() def rand_name(prefix): return '{}-{}'.format(prefix, str(uuid.uuid4()))
import logging import os import uuid from contextlib import contextmanager from subprocess import check_call, CalledProcessError LOG = logging.getLogger(__name__) def touch(directory, filename=None): file_path = os.path.join(directory, filename) if os.path.exists(file_path): os.utime(file_path, None) else: os.makedirs(directory) if filename is not None: open(file_path, 'a').close() return file_path def run_cmd(command, **kwargs): """ Runs a command and returns an array of its results :param command: String of a command to run within a shell :returns: Dictionary with keys relating to the execution's success """ if kwargs['cwd']: cwd = kwargs['cwd'] try: ret = check_call(command, shell=True, cwd=cwd) return {'success': True, 'return': ret, 'exception': None} except CalledProcessError as exc: return {'success': False, 'return': None, 'exception': exc, 'command': command} @contextmanager def cleanup(stage): try: yield except (Exception, KeyboardInterrupt): LOG.exception('Run failed') finally: stage.destroy() def rand_name(prefix): return '{}-{}'.format(prefix, str(uuid.uuid4())) Add fallback to 'cwd' if not definedimport logging import os import uuid from contextlib import contextmanager from subprocess import check_call, CalledProcessError LOG = logging.getLogger(__name__) def touch(directory, filename=None): file_path = os.path.join(directory, filename) if os.path.exists(file_path): os.utime(file_path, None) else: os.makedirs(directory) if filename is not None: open(file_path, 'a').close() return file_path def run_cmd(command, **kwargs): """ Runs a command and returns an array of its results :param command: String of a command to run within a shell :returns: Dictionary with keys relating to the execution's success """ cwd = kwargs.get('cwd', None) try: ret = check_call(command, shell=True, cwd=cwd) return {'success': True, 'return': ret, 'exception': None} except CalledProcessError as exc: return {'success': False, 'return': None, 'exception': exc, 'command': command} @contextmanager def cleanup(stage): try: yield except (Exception, KeyboardInterrupt): LOG.exception('Run failed') finally: stage.destroy() def rand_name(prefix): return '{}-{}'.format(prefix, str(uuid.uuid4()))
<commit_before>import logging import os import uuid from contextlib import contextmanager from subprocess import check_call, CalledProcessError LOG = logging.getLogger(__name__) def touch(directory, filename=None): file_path = os.path.join(directory, filename) if os.path.exists(file_path): os.utime(file_path, None) else: os.makedirs(directory) if filename is not None: open(file_path, 'a').close() return file_path def run_cmd(command, **kwargs): """ Runs a command and returns an array of its results :param command: String of a command to run within a shell :returns: Dictionary with keys relating to the execution's success """ if kwargs['cwd']: cwd = kwargs['cwd'] try: ret = check_call(command, shell=True, cwd=cwd) return {'success': True, 'return': ret, 'exception': None} except CalledProcessError as exc: return {'success': False, 'return': None, 'exception': exc, 'command': command} @contextmanager def cleanup(stage): try: yield except (Exception, KeyboardInterrupt): LOG.exception('Run failed') finally: stage.destroy() def rand_name(prefix): return '{}-{}'.format(prefix, str(uuid.uuid4())) <commit_msg>Add fallback to 'cwd' if not defined<commit_after>import logging import os import uuid from contextlib import contextmanager from subprocess import check_call, CalledProcessError LOG = logging.getLogger(__name__) def touch(directory, filename=None): file_path = os.path.join(directory, filename) if os.path.exists(file_path): os.utime(file_path, None) else: os.makedirs(directory) if filename is not None: open(file_path, 'a').close() return file_path def run_cmd(command, **kwargs): """ Runs a command and returns an array of its results :param command: String of a command to run within a shell :returns: Dictionary with keys relating to the execution's success """ cwd = kwargs.get('cwd', None) try: ret = check_call(command, shell=True, cwd=cwd) return {'success': True, 'return': ret, 'exception': None} except CalledProcessError as exc: return {'success': False, 'return': None, 'exception': exc, 'command': command} @contextmanager def cleanup(stage): try: yield except (Exception, KeyboardInterrupt): LOG.exception('Run failed') finally: stage.destroy() def rand_name(prefix): return '{}-{}'.format(prefix, str(uuid.uuid4()))
a4d5a74c675b0c1d90dd9f254367975af9ea2735
opps/core/models/__init__.py
opps/core/models/__init__.py
# -*- coding: utf-8 -*- from opps.core.models.channel import * from opps.core.models.profile import * from opps.core.models.source import * from opps.core.models.publisher import *
# -*- coding: utf-8 -*- from opps.core.models.channel import * from opps.core.models.profile import * from opps.core.models.source import * from opps.core.models.published import *
Change package core mdoels init
Change package core mdoels init
Python
mit
jeanmask/opps,williamroot/opps,williamroot/opps,jeanmask/opps,opps/opps,YACOWS/opps,YACOWS/opps,opps/opps,williamroot/opps,jeanmask/opps,YACOWS/opps,opps/opps,opps/opps,jeanmask/opps,williamroot/opps,YACOWS/opps
# -*- coding: utf-8 -*- from opps.core.models.channel import * from opps.core.models.profile import * from opps.core.models.source import * from opps.core.models.publisher import * Change package core mdoels init
# -*- coding: utf-8 -*- from opps.core.models.channel import * from opps.core.models.profile import * from opps.core.models.source import * from opps.core.models.published import *
<commit_before># -*- coding: utf-8 -*- from opps.core.models.channel import * from opps.core.models.profile import * from opps.core.models.source import * from opps.core.models.publisher import * <commit_msg>Change package core mdoels init<commit_after>
# -*- coding: utf-8 -*- from opps.core.models.channel import * from opps.core.models.profile import * from opps.core.models.source import * from opps.core.models.published import *
# -*- coding: utf-8 -*- from opps.core.models.channel import * from opps.core.models.profile import * from opps.core.models.source import * from opps.core.models.publisher import * Change package core mdoels init# -*- coding: utf-8 -*- from opps.core.models.channel import * from opps.core.models.profile import * from opps.core.models.source import * from opps.core.models.published import *
<commit_before># -*- coding: utf-8 -*- from opps.core.models.channel import * from opps.core.models.profile import * from opps.core.models.source import * from opps.core.models.publisher import * <commit_msg>Change package core mdoels init<commit_after># -*- coding: utf-8 -*- from opps.core.models.channel import * from opps.core.models.profile import * from opps.core.models.source import * from opps.core.models.published import *
096bd7e3357e85c57ec56695bfc16f0b4eab9c4d
pystil.py
pystil.py
#!/usr/bin/env python # -*- coding: utf-8 -*- # Copyright (C) 2011 by Florian Mounier, Kozea # This file is part of pystil, licensed under a 3-clause BSD license. """ pystil - An elegant site web traffic analyzer """ from pystil import app, config import werkzeug.contrib import sys config.freeze() if 'soup' in sys.argv: from os import getenv from sqlalchemy import create_engine from sqlalchemy.ext.sqlsoup import SqlSoup import code engine = create_engine(config.CONFIG["DB_URL"], echo=True) db = SqlSoup(engine) Visit = db.visit Keys = db.keys class FunkyConsole(code.InteractiveConsole): def showtraceback(self): import pdb import sys code.InteractiveConsole.showtraceback(self) pdb.post_mortem(sys.exc_info()[2]) console = FunkyConsole(locals=globals()) if getenv("PYTHONSTARTUP"): execfile(getenv("PYTHONSTARTUP")) console.interact() else: from pystil.service.http import Application from gevent import monkey monkey.patch_all() import gevent.wsgi application = Application(werkzeug.contrib.fixers.ProxyFix(app())) ws = gevent.wsgi.WSGIServer(('', 1789), application) ws.serve_forever()
#!/usr/bin/env python # -*- coding: utf-8 -*- # Copyright (C) 2011 by Florian Mounier, Kozea # This file is part of pystil, licensed under a 3-clause BSD license. """ pystil - An elegant site web traffic analyzer """ from pystil import app, config import werkzeug.contrib.fixers import sys config.freeze() if 'soup' in sys.argv: from os import getenv from sqlalchemy import create_engine from sqlalchemy.ext.sqlsoup import SqlSoup import code engine = create_engine(config.CONFIG["DB_URL"], echo=True) db = SqlSoup(engine) Visit = db.visit Keys = db.keys class FunkyConsole(code.InteractiveConsole): def showtraceback(self): import pdb import sys code.InteractiveConsole.showtraceback(self) pdb.post_mortem(sys.exc_info()[2]) console = FunkyConsole(locals=globals()) if getenv("PYTHONSTARTUP"): execfile(getenv("PYTHONSTARTUP")) console.interact() else: from pystil.service.http import Application from gevent import monkey monkey.patch_all() import gevent.wsgi application = werkzeug.contrib.fixers.ProxyFix(Application(app())) ws = gevent.wsgi.WSGIServer(('', 1789), application) ws.serve_forever()
Add the good old middleware
Add the good old middleware
Python
bsd-3-clause
Kozea/pystil,Kozea/pystil,Kozea/pystil,Kozea/pystil,Kozea/pystil
#!/usr/bin/env python # -*- coding: utf-8 -*- # Copyright (C) 2011 by Florian Mounier, Kozea # This file is part of pystil, licensed under a 3-clause BSD license. """ pystil - An elegant site web traffic analyzer """ from pystil import app, config import werkzeug.contrib import sys config.freeze() if 'soup' in sys.argv: from os import getenv from sqlalchemy import create_engine from sqlalchemy.ext.sqlsoup import SqlSoup import code engine = create_engine(config.CONFIG["DB_URL"], echo=True) db = SqlSoup(engine) Visit = db.visit Keys = db.keys class FunkyConsole(code.InteractiveConsole): def showtraceback(self): import pdb import sys code.InteractiveConsole.showtraceback(self) pdb.post_mortem(sys.exc_info()[2]) console = FunkyConsole(locals=globals()) if getenv("PYTHONSTARTUP"): execfile(getenv("PYTHONSTARTUP")) console.interact() else: from pystil.service.http import Application from gevent import monkey monkey.patch_all() import gevent.wsgi application = Application(werkzeug.contrib.fixers.ProxyFix(app())) ws = gevent.wsgi.WSGIServer(('', 1789), application) ws.serve_forever() Add the good old middleware
#!/usr/bin/env python # -*- coding: utf-8 -*- # Copyright (C) 2011 by Florian Mounier, Kozea # This file is part of pystil, licensed under a 3-clause BSD license. """ pystil - An elegant site web traffic analyzer """ from pystil import app, config import werkzeug.contrib.fixers import sys config.freeze() if 'soup' in sys.argv: from os import getenv from sqlalchemy import create_engine from sqlalchemy.ext.sqlsoup import SqlSoup import code engine = create_engine(config.CONFIG["DB_URL"], echo=True) db = SqlSoup(engine) Visit = db.visit Keys = db.keys class FunkyConsole(code.InteractiveConsole): def showtraceback(self): import pdb import sys code.InteractiveConsole.showtraceback(self) pdb.post_mortem(sys.exc_info()[2]) console = FunkyConsole(locals=globals()) if getenv("PYTHONSTARTUP"): execfile(getenv("PYTHONSTARTUP")) console.interact() else: from pystil.service.http import Application from gevent import monkey monkey.patch_all() import gevent.wsgi application = werkzeug.contrib.fixers.ProxyFix(Application(app())) ws = gevent.wsgi.WSGIServer(('', 1789), application) ws.serve_forever()
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- # Copyright (C) 2011 by Florian Mounier, Kozea # This file is part of pystil, licensed under a 3-clause BSD license. """ pystil - An elegant site web traffic analyzer """ from pystil import app, config import werkzeug.contrib import sys config.freeze() if 'soup' in sys.argv: from os import getenv from sqlalchemy import create_engine from sqlalchemy.ext.sqlsoup import SqlSoup import code engine = create_engine(config.CONFIG["DB_URL"], echo=True) db = SqlSoup(engine) Visit = db.visit Keys = db.keys class FunkyConsole(code.InteractiveConsole): def showtraceback(self): import pdb import sys code.InteractiveConsole.showtraceback(self) pdb.post_mortem(sys.exc_info()[2]) console = FunkyConsole(locals=globals()) if getenv("PYTHONSTARTUP"): execfile(getenv("PYTHONSTARTUP")) console.interact() else: from pystil.service.http import Application from gevent import monkey monkey.patch_all() import gevent.wsgi application = Application(werkzeug.contrib.fixers.ProxyFix(app())) ws = gevent.wsgi.WSGIServer(('', 1789), application) ws.serve_forever() <commit_msg>Add the good old middleware<commit_after>
#!/usr/bin/env python # -*- coding: utf-8 -*- # Copyright (C) 2011 by Florian Mounier, Kozea # This file is part of pystil, licensed under a 3-clause BSD license. """ pystil - An elegant site web traffic analyzer """ from pystil import app, config import werkzeug.contrib.fixers import sys config.freeze() if 'soup' in sys.argv: from os import getenv from sqlalchemy import create_engine from sqlalchemy.ext.sqlsoup import SqlSoup import code engine = create_engine(config.CONFIG["DB_URL"], echo=True) db = SqlSoup(engine) Visit = db.visit Keys = db.keys class FunkyConsole(code.InteractiveConsole): def showtraceback(self): import pdb import sys code.InteractiveConsole.showtraceback(self) pdb.post_mortem(sys.exc_info()[2]) console = FunkyConsole(locals=globals()) if getenv("PYTHONSTARTUP"): execfile(getenv("PYTHONSTARTUP")) console.interact() else: from pystil.service.http import Application from gevent import monkey monkey.patch_all() import gevent.wsgi application = werkzeug.contrib.fixers.ProxyFix(Application(app())) ws = gevent.wsgi.WSGIServer(('', 1789), application) ws.serve_forever()
#!/usr/bin/env python # -*- coding: utf-8 -*- # Copyright (C) 2011 by Florian Mounier, Kozea # This file is part of pystil, licensed under a 3-clause BSD license. """ pystil - An elegant site web traffic analyzer """ from pystil import app, config import werkzeug.contrib import sys config.freeze() if 'soup' in sys.argv: from os import getenv from sqlalchemy import create_engine from sqlalchemy.ext.sqlsoup import SqlSoup import code engine = create_engine(config.CONFIG["DB_URL"], echo=True) db = SqlSoup(engine) Visit = db.visit Keys = db.keys class FunkyConsole(code.InteractiveConsole): def showtraceback(self): import pdb import sys code.InteractiveConsole.showtraceback(self) pdb.post_mortem(sys.exc_info()[2]) console = FunkyConsole(locals=globals()) if getenv("PYTHONSTARTUP"): execfile(getenv("PYTHONSTARTUP")) console.interact() else: from pystil.service.http import Application from gevent import monkey monkey.patch_all() import gevent.wsgi application = Application(werkzeug.contrib.fixers.ProxyFix(app())) ws = gevent.wsgi.WSGIServer(('', 1789), application) ws.serve_forever() Add the good old middleware#!/usr/bin/env python # -*- coding: utf-8 -*- # Copyright (C) 2011 by Florian Mounier, Kozea # This file is part of pystil, licensed under a 3-clause BSD license. """ pystil - An elegant site web traffic analyzer """ from pystil import app, config import werkzeug.contrib.fixers import sys config.freeze() if 'soup' in sys.argv: from os import getenv from sqlalchemy import create_engine from sqlalchemy.ext.sqlsoup import SqlSoup import code engine = create_engine(config.CONFIG["DB_URL"], echo=True) db = SqlSoup(engine) Visit = db.visit Keys = db.keys class FunkyConsole(code.InteractiveConsole): def showtraceback(self): import pdb import sys code.InteractiveConsole.showtraceback(self) pdb.post_mortem(sys.exc_info()[2]) console = FunkyConsole(locals=globals()) if getenv("PYTHONSTARTUP"): execfile(getenv("PYTHONSTARTUP")) console.interact() else: from pystil.service.http import Application from gevent import monkey monkey.patch_all() import gevent.wsgi application = werkzeug.contrib.fixers.ProxyFix(Application(app())) ws = gevent.wsgi.WSGIServer(('', 1789), application) ws.serve_forever()
<commit_before>#!/usr/bin/env python # -*- coding: utf-8 -*- # Copyright (C) 2011 by Florian Mounier, Kozea # This file is part of pystil, licensed under a 3-clause BSD license. """ pystil - An elegant site web traffic analyzer """ from pystil import app, config import werkzeug.contrib import sys config.freeze() if 'soup' in sys.argv: from os import getenv from sqlalchemy import create_engine from sqlalchemy.ext.sqlsoup import SqlSoup import code engine = create_engine(config.CONFIG["DB_URL"], echo=True) db = SqlSoup(engine) Visit = db.visit Keys = db.keys class FunkyConsole(code.InteractiveConsole): def showtraceback(self): import pdb import sys code.InteractiveConsole.showtraceback(self) pdb.post_mortem(sys.exc_info()[2]) console = FunkyConsole(locals=globals()) if getenv("PYTHONSTARTUP"): execfile(getenv("PYTHONSTARTUP")) console.interact() else: from pystil.service.http import Application from gevent import monkey monkey.patch_all() import gevent.wsgi application = Application(werkzeug.contrib.fixers.ProxyFix(app())) ws = gevent.wsgi.WSGIServer(('', 1789), application) ws.serve_forever() <commit_msg>Add the good old middleware<commit_after>#!/usr/bin/env python # -*- coding: utf-8 -*- # Copyright (C) 2011 by Florian Mounier, Kozea # This file is part of pystil, licensed under a 3-clause BSD license. """ pystil - An elegant site web traffic analyzer """ from pystil import app, config import werkzeug.contrib.fixers import sys config.freeze() if 'soup' in sys.argv: from os import getenv from sqlalchemy import create_engine from sqlalchemy.ext.sqlsoup import SqlSoup import code engine = create_engine(config.CONFIG["DB_URL"], echo=True) db = SqlSoup(engine) Visit = db.visit Keys = db.keys class FunkyConsole(code.InteractiveConsole): def showtraceback(self): import pdb import sys code.InteractiveConsole.showtraceback(self) pdb.post_mortem(sys.exc_info()[2]) console = FunkyConsole(locals=globals()) if getenv("PYTHONSTARTUP"): execfile(getenv("PYTHONSTARTUP")) console.interact() else: from pystil.service.http import Application from gevent import monkey monkey.patch_all() import gevent.wsgi application = werkzeug.contrib.fixers.ProxyFix(Application(app())) ws = gevent.wsgi.WSGIServer(('', 1789), application) ws.serve_forever()
e7c5e62da700f51e69662689758ffebf70fa1494
cms/djangoapps/contentstore/context_processors.py
cms/djangoapps/contentstore/context_processors.py
import ConfigParser from django.conf import settings def doc_url(request): config_file = open(settings.REPO_ROOT / "docs" / "config.ini") config = ConfigParser.ConfigParser() config.readfp(config_file) # in the future, we will detect the locale; for now, we will # hardcode en_us, since we only have English documentation locale = "en_us" def get_doc_url(token): try: return config.get(locale, token) except ConfigParser.NoOptionError: return config.get(locale, "default") return {"doc_url": get_doc_url}
import ConfigParser from django.conf import settings config_file = open(settings.REPO_ROOT / "docs" / "config.ini") config = ConfigParser.ConfigParser() config.readfp(config_file) def doc_url(request): # in the future, we will detect the locale; for now, we will # hardcode en_us, since we only have English documentation locale = "en_us" def get_doc_url(token): try: return config.get(locale, token) except ConfigParser.NoOptionError: return config.get(locale, "default") return {"doc_url": get_doc_url}
Read from doc url mapping file at load time, rather than once per request
Read from doc url mapping file at load time, rather than once per request
Python
agpl-3.0
ampax/edx-platform,Softmotions/edx-platform,dcosentino/edx-platform,chudaol/edx-platform,prarthitm/edxplatform,cyanna/edx-platform,jazztpt/edx-platform,motion2015/a3,nikolas/edx-platform,msegado/edx-platform,hkawasaki/kawasaki-aio8-0,LearnEra/LearnEraPlaftform,pabloborrego93/edx-platform,wwj718/ANALYSE,procangroup/edx-platform,ahmadiga/min_edx,wwj718/edx-platform,edx/edx-platform,morenopc/edx-platform,leansoft/edx-platform,cyanna/edx-platform,shubhdev/edxOnBaadal,leansoft/edx-platform,SivilTaram/edx-platform,shabab12/edx-platform,zadgroup/edx-platform,dsajkl/123,pku9104038/edx-platform,antoviaque/edx-platform,ovnicraft/edx-platform,mahendra-r/edx-platform,zhenzhai/edx-platform,eestay/edx-platform,abdoosh00/edraak,jjmiranda/edx-platform,leansoft/edx-platform,nikolas/edx-platform,longmen21/edx-platform,jruiperezv/ANALYSE,kamalx/edx-platform,Kalyzee/edx-platform,andyzsf/edx,CourseTalk/edx-platform,beacloudgenius/edx-platform,chauhanhardik/populo_2,fly19890211/edx-platform,rismalrv/edx-platform,hkawasaki/kawasaki-aio8-0,teltek/edx-platform,openfun/edx-platform,devs1991/test_edx_docmode,shubhdev/openedx,zadgroup/edx-platform,waheedahmed/edx-platform,Kalyzee/edx-platform,vasyarv/edx-platform,shubhdev/openedx,louyihua/edx-platform,Shrhawk/edx-platform,don-github/edx-platform,alu042/edx-platform,zhenzhai/edx-platform,mcgachey/edx-platform,mitocw/edx-platform,xuxiao19910803/edx-platform,LICEF/edx-platform,jswope00/griffinx,y12uc231/edx-platform,nanolearning/edx-platform,inares/edx-platform,benpatterson/edx-platform,ahmedaljazzar/edx-platform,mjirayu/sit_academy,gsehub/edx-platform,deepsrijit1105/edx-platform,Stanford-Online/edx-platform,zofuthan/edx-platform,playm2mboy/edx-platform,a-parhom/edx-platform,miptliot/edx-platform,playm2mboy/edx-platform,beni55/edx-platform,alu042/edx-platform,LICEF/edx-platform,naresh21/synergetics-edx-platform,bigdatauniversity/edx-platform,AkA84/edx-platform,caesar2164/edx-platform,WatanabeYasumasa/edx-platform,4eek/edx-platform,AkA84/edx-platform,UOMx/edx-platform,openfun/edx-platform,pku9104038/edx-platform,eestay/edx-platform,ampax/edx-platform,doganov/edx-platform,synergeticsedx/deployment-wipro,hkawasaki/kawasaki-aio8-0,jzoldak/edx-platform,franosincic/edx-platform,jswope00/griffinx,hkawasaki/kawasaki-aio8-2,proversity-org/edx-platform,dcosentino/edx-platform,dkarakats/edx-platform,tanmaykm/edx-platform,appliedx/edx-platform,nagyistoce/edx-platform,knehez/edx-platform,openfun/edx-platform,torchingloom/edx-platform,alu042/edx-platform,iivic/BoiseStateX,a-parhom/edx-platform,xuxiao19910803/edx-platform,etzhou/edx-platform,SravanthiSinha/edx-platform,pomegranited/edx-platform,EDUlib/edx-platform,cognitiveclass/edx-platform,arifsetiawan/edx-platform,cselis86/edx-platform,Ayub-Khan/edx-platform,vismartltd/edx-platform,unicri/edx-platform,kursitet/edx-platform,adoosii/edx-platform,fintech-circle/edx-platform,bitifirefly/edx-platform,mitocw/edx-platform,zerobatu/edx-platform,solashirai/edx-platform,doismellburning/edx-platform,proversity-org/edx-platform,xuxiao19910803/edx,ampax/edx-platform-backup,jazztpt/edx-platform,ak2703/edx-platform,cecep-edu/edx-platform,jelugbo/tundex,RPI-OPENEDX/edx-platform,chauhanhardik/populo_2,BehavioralInsightsTeam/edx-platform,rhndg/openedx,martynovp/edx-platform,tiagochiavericosta/edx-platform,jamiefolsom/edx-platform,nanolearning/edx-platform,BehavioralInsightsTeam/edx-platform,doganov/edx-platform,vasyarv/edx-platform,edry/edx-platform,tiagochiavericosta/edx-platform,fintech-circle/edx-platform,DefyVentures/edx-platform,OmarIthawi/edx-platform,jamesblunt/edx-platform,ZLLab-Mooc/edx-platform,alexthered/kienhoc-platform,devs1991/test_edx_docmode,cecep-edu/edx-platform,tiagochiavericosta/edx-platform,Edraak/edx-platform,jelugbo/tundex,motion2015/edx-platform,nanolearningllc/edx-platform-cypress-2,rismalrv/edx-platform,halvertoluke/edx-platform,mbareta/edx-platform-ft,marcore/edx-platform,miptliot/edx-platform,vikas1885/test1,B-MOOC/edx-platform,Kalyzee/edx-platform,dsajkl/reqiop,Livit/Livit.Learn.EdX,appsembler/edx-platform,amir-qayyum-khan/edx-platform,playm2mboy/edx-platform,nanolearningllc/edx-platform-cypress,procangroup/edx-platform,J861449197/edx-platform,nttks/jenkins-test,jbassen/edx-platform,ovnicraft/edx-platform,appliedx/edx-platform,gymnasium/edx-platform,mushtaqak/edx-platform,marcore/edx-platform,hkawasaki/kawasaki-aio8-2,ferabra/edx-platform,mbareta/edx-platform-ft,kursitet/edx-platform,nagyistoce/edx-platform,hkawasaki/kawasaki-aio8-1,ovnicraft/edx-platform,inares/edx-platform,jonathan-beard/edx-platform,IndonesiaX/edx-platform,UOMx/edx-platform,marcore/edx-platform,peterm-itr/edx-platform,shashank971/edx-platform,devs1991/test_edx_docmode,Shrhawk/edx-platform,naresh21/synergetics-edx-platform,morenopc/edx-platform,nanolearningllc/edx-platform-cypress-2,zofuthan/edx-platform,zhenzhai/edx-platform,EDUlib/edx-platform,polimediaupv/edx-platform,procangroup/edx-platform,jazztpt/edx-platform,atsolakid/edx-platform,stvstnfrd/edx-platform,doismellburning/edx-platform,xingyepei/edx-platform,jswope00/griffinx,hamzehd/edx-platform,motion2015/edx-platform,Edraak/circleci-edx-platform,sudheerchintala/LearnEraPlatForm,xinjiguaike/edx-platform,nttks/jenkins-test,mushtaqak/edx-platform,alu042/edx-platform,analyseuc3m/ANALYSE-v1,appliedx/edx-platform,chudaol/edx-platform,ahmadiga/min_edx,sameetb-cuelogic/edx-platform-test,romain-li/edx-platform,ESOedX/edx-platform,simbs/edx-platform,abdoosh00/edraak,chauhanhardik/populo,TeachAtTUM/edx-platform,shubhdev/edx-platform,gsehub/edx-platform,martynovp/edx-platform,ampax/edx-platform,appliedx/edx-platform,vikas1885/test1,IONISx/edx-platform,TeachAtTUM/edx-platform,shabab12/edx-platform,dsajkl/123,shubhdev/edxOnBaadal,hastexo/edx-platform,hkawasaki/kawasaki-aio8-2,prarthitm/edxplatform,cecep-edu/edx-platform,ubc/edx-platform,shurihell/testasia,kamalx/edx-platform,valtech-mooc/edx-platform,polimediaupv/edx-platform,motion2015/a3,antoviaque/edx-platform,JioEducation/edx-platform,prarthitm/edxplatform,benpatterson/edx-platform,defance/edx-platform,cognitiveclass/edx-platform,romain-li/edx-platform,hkawasaki/kawasaki-aio8-2,nanolearningllc/edx-platform-cypress,UXE/local-edx,zerobatu/edx-platform,xuxiao19910803/edx,peterm-itr/edx-platform,synergeticsedx/deployment-wipro,Stanford-Online/edx-platform,Semi-global/edx-platform,caesar2164/edx-platform,atsolakid/edx-platform,DNFcode/edx-platform,ahmadio/edx-platform,ahmadio/edx-platform,mahendra-r/edx-platform,polimediaupv/edx-platform,10clouds/edx-platform,rhndg/openedx,xuxiao19910803/edx-platform,kmoocdev2/edx-platform,shubhdev/edxOnBaadal,wwj718/edx-platform,vikas1885/test1,ahmedaljazzar/edx-platform,J861449197/edx-platform,inares/edx-platform,atsolakid/edx-platform,abdoosh00/edraak,UOMx/edx-platform,jazkarta/edx-platform-for-isc,simbs/edx-platform,nanolearningllc/edx-platform-cypress-2,OmarIthawi/edx-platform,shubhdev/openedx,ampax/edx-platform-backup,stvstnfrd/edx-platform,synergeticsedx/deployment-wipro,itsjeyd/edx-platform,romain-li/edx-platform,analyseuc3m/ANALYSE-v1,chand3040/cloud_that,SivilTaram/edx-platform,mushtaqak/edx-platform,msegado/edx-platform,beni55/edx-platform,hamzehd/edx-platform,Ayub-Khan/edx-platform,vikas1885/test1,kamalx/edx-platform,eduNEXT/edunext-platform,jazkarta/edx-platform,mahendra-r/edx-platform,OmarIthawi/edx-platform,defance/edx-platform,andyzsf/edx,CourseTalk/edx-platform,arifsetiawan/edx-platform,JioEducation/edx-platform,chrisndodge/edx-platform,openfun/edx-platform,kmoocdev2/edx-platform,IONISx/edx-platform,yokose-ks/edx-platform,shurihell/testasia,wwj718/ANALYSE,alexthered/kienhoc-platform,solashirai/edx-platform,tanmaykm/edx-platform,mjirayu/sit_academy,chudaol/edx-platform,cognitiveclass/edx-platform,kxliugang/edx-platform,ahmadiga/min_edx,Edraak/edraak-platform,Edraak/circleci-edx-platform,Edraak/circleci-edx-platform,knehez/edx-platform,fly19890211/edx-platform,jonathan-beard/edx-platform,sameetb-cuelogic/edx-platform-test,AkA84/edx-platform,cyanna/edx-platform,zhenzhai/edx-platform,kmoocdev/edx-platform,etzhou/edx-platform,jelugbo/tundex,iivic/BoiseStateX,msegado/edx-platform,angelapper/edx-platform,jolyonb/edx-platform,zofuthan/edx-platform,itsjeyd/edx-platform,rhndg/openedx,chudaol/edx-platform,mtlchun/edx,ESOedX/edx-platform,xuxiao19910803/edx-platform,jamiefolsom/edx-platform,CourseTalk/edx-platform,rue89-tech/edx-platform,ZLLab-Mooc/edx-platform,jazztpt/edx-platform,UXE/local-edx,Semi-global/edx-platform,unicri/edx-platform,raccoongang/edx-platform,polimediaupv/edx-platform,MSOpenTech/edx-platform,Kalyzee/edx-platform,jbassen/edx-platform,pomegranited/edx-platform,arifsetiawan/edx-platform,Softmotions/edx-platform,IndonesiaX/edx-platform,ovnicraft/edx-platform,SivilTaram/edx-platform,longmen21/edx-platform,jamiefolsom/edx-platform,hmcmooc/muddx-platform,ubc/edx-platform,valtech-mooc/edx-platform,JCBarahona/edX,halvertoluke/edx-platform,jswope00/griffinx,DefyVentures/edx-platform,kmoocdev/edx-platform,zerobatu/edx-platform,Stanford-Online/edx-platform,cecep-edu/edx-platform,utecuy/edx-platform,Edraak/edraak-platform,nanolearningllc/edx-platform-cypress-2,defance/edx-platform,pku9104038/edx-platform,jswope00/GAI,stvstnfrd/edx-platform,chand3040/cloud_that,OmarIthawi/edx-platform,openfun/edx-platform,playm2mboy/edx-platform,adoosii/edx-platform,ampax/edx-platform-backup,olexiim/edx-platform,don-github/edx-platform,DefyVentures/edx-platform,appliedx/edx-platform,edry/edx-platform,dsajkl/reqiop,fintech-circle/edx-platform,martynovp/edx-platform,angelapper/edx-platform,SivilTaram/edx-platform,leansoft/edx-platform,edx/edx-platform,hkawasaki/kawasaki-aio8-1,doganov/edx-platform,ak2703/edx-platform,pomegranited/edx-platform,philanthropy-u/edx-platform,pomegranited/edx-platform,hamzehd/edx-platform,jbassen/edx-platform,dsajkl/123,franosincic/edx-platform,tanmaykm/edx-platform,hmcmooc/muddx-platform,mcgachey/edx-platform,SravanthiSinha/edx-platform,gsehub/edx-platform,louyihua/edx-platform,eemirtekin/edx-platform,sudheerchintala/LearnEraPlatForm,ampax/edx-platform-backup,bdero/edx-platform,xinjiguaike/edx-platform,jolyonb/edx-platform,nttks/jenkins-test,J861449197/edx-platform,jjmiranda/edx-platform,Shrhawk/edx-platform,jonathan-beard/edx-platform,antonve/s4-project-mooc,hamzehd/edx-platform,antoviaque/edx-platform,RPI-OPENEDX/edx-platform,solashirai/edx-platform,chauhanhardik/populo_2,jbzdak/edx-platform,nttks/edx-platform,ESOedX/edx-platform,EDUlib/edx-platform,edx-solutions/edx-platform,jamesblunt/edx-platform,xingyepei/edx-platform,mcgachey/edx-platform,kmoocdev2/edx-platform,dcosentino/edx-platform,rue89-tech/edx-platform,xuxiao19910803/edx,ampax/edx-platform,4eek/edx-platform,ahmadiga/min_edx,deepsrijit1105/edx-platform,jolyonb/edx-platform,unicri/edx-platform,ubc/edx-platform,rismalrv/edx-platform,MakeHer/edx-platform,motion2015/edx-platform,CredoReference/edx-platform,solashirai/edx-platform,jbzdak/edx-platform,Livit/Livit.Learn.EdX,RPI-OPENEDX/edx-platform,eduNEXT/edx-platform,hkawasaki/kawasaki-aio8-1,kmoocdev/edx-platform,devs1991/test_edx_docmode,Softmotions/edx-platform,morenopc/edx-platform,beni55/edx-platform,J861449197/edx-platform,jolyonb/edx-platform,itsjeyd/edx-platform,morenopc/edx-platform,shubhdev/edxOnBaadal,jazkarta/edx-platform,dkarakats/edx-platform,simbs/edx-platform,Lektorium-LLC/edx-platform,zadgroup/edx-platform,TeachAtTUM/edx-platform,RPI-OPENEDX/edx-platform,sameetb-cuelogic/edx-platform-test,nagyistoce/edx-platform,kursitet/edx-platform,jswope00/GAI,unicri/edx-platform,nikolas/edx-platform,Softmotions/edx-platform,LICEF/edx-platform,cognitiveclass/edx-platform,atsolakid/edx-platform,MSOpenTech/edx-platform,ahmadio/edx-platform,bitifirefly/edx-platform,carsongee/edx-platform,beacloudgenius/edx-platform,ak2703/edx-platform,defance/edx-platform,pepeportela/edx-platform,amir-qayyum-khan/edx-platform,JCBarahona/edX,carsongee/edx-platform,rhndg/openedx,gymnasium/edx-platform,beacloudgenius/edx-platform,kxliugang/edx-platform,longmen21/edx-platform,4eek/edx-platform,bitifirefly/edx-platform,mbareta/edx-platform-ft,ferabra/edx-platform,doganov/edx-platform,jonathan-beard/edx-platform,abdoosh00/edraak,vikas1885/test1,cecep-edu/edx-platform,franosincic/edx-platform,beni55/edx-platform,cpennington/edx-platform,Semi-global/edx-platform,B-MOOC/edx-platform,angelapper/edx-platform,ak2703/edx-platform,MakeHer/edx-platform,utecuy/edx-platform,zadgroup/edx-platform,jbzdak/edx-platform,olexiim/edx-platform,adoosii/edx-platform,cselis86/edx-platform,deepsrijit1105/edx-platform,bdero/edx-platform,angelapper/edx-platform,knehez/edx-platform,xingyepei/edx-platform,mbareta/edx-platform-ft,shubhdev/edx-platform,fly19890211/edx-platform,shashank971/edx-platform,xuxiao19910803/edx,edx/edx-platform,caesar2164/edx-platform,simbs/edx-platform,mahendra-r/edx-platform,Edraak/circleci-edx-platform,lduarte1991/edx-platform,jswope00/GAI,B-MOOC/edx-platform,chand3040/cloud_that,valtech-mooc/edx-platform,amir-qayyum-khan/edx-platform,B-MOOC/edx-platform,martynovp/edx-platform,vasyarv/edx-platform,zerobatu/edx-platform,jzoldak/edx-platform,mitocw/edx-platform,alexthered/kienhoc-platform,lduarte1991/edx-platform,motion2015/a3,alexthered/kienhoc-platform,BehavioralInsightsTeam/edx-platform,edx-solutions/edx-platform,shurihell/testasia,wwj718/ANALYSE,mtlchun/edx,philanthropy-u/edx-platform,olexiim/edx-platform,tiagochiavericosta/edx-platform,jelugbo/tundex,IONISx/edx-platform,mtlchun/edx,doismellburning/edx-platform,xinjiguaike/edx-platform,eemirtekin/edx-platform,vasyarv/edx-platform,MSOpenTech/edx-platform,Livit/Livit.Learn.EdX,knehez/edx-platform,bdero/edx-platform,raccoongang/edx-platform,mcgachey/edx-platform,shashank971/edx-platform,jzoldak/edx-platform,EDUlib/edx-platform,dkarakats/edx-platform,ZLLab-Mooc/edx-platform,itsjeyd/edx-platform,shabab12/edx-platform,J861449197/edx-platform,antonve/s4-project-mooc,jamesblunt/edx-platform,hmcmooc/muddx-platform,gymnasium/edx-platform,zubair-arbi/edx-platform,doismellburning/edx-platform,xuxiao19910803/edx-platform,appsembler/edx-platform,kamalx/edx-platform,zubair-arbi/edx-platform,LearnEra/LearnEraPlaftform,mcgachey/edx-platform,edry/edx-platform,shubhdev/openedx,JioEducation/edx-platform,solashirai/edx-platform,nanolearningllc/edx-platform-cypress,jruiperezv/ANALYSE,halvertoluke/edx-platform,analyseuc3m/ANALYSE-v1,arifsetiawan/edx-platform,appsembler/edx-platform,lduarte1991/edx-platform,Stanford-Online/edx-platform,ovnicraft/edx-platform,vismartltd/edx-platform,nagyistoce/edx-platform,LICEF/edx-platform,Lektorium-LLC/edx-platform,chauhanhardik/populo,chand3040/cloud_that,chrisndodge/edx-platform,cyanna/edx-platform,benpatterson/edx-platform,franosincic/edx-platform,Ayub-Khan/edx-platform,cselis86/edx-platform,mtlchun/edx,Lektorium-LLC/edx-platform,CredoReference/edx-platform,dcosentino/edx-platform,cognitiveclass/edx-platform,bigdatauniversity/edx-platform,JCBarahona/edX,jazztpt/edx-platform,antonve/s4-project-mooc,Edraak/edx-platform,shashank971/edx-platform,nanolearning/edx-platform,nanolearning/edx-platform,rue89-tech/edx-platform,miptliot/edx-platform,jonathan-beard/edx-platform,beni55/edx-platform,edx/edx-platform,devs1991/test_edx_docmode,nttks/edx-platform,dsajkl/reqiop,proversity-org/edx-platform,devs1991/test_edx_docmode,yokose-ks/edx-platform,antonve/s4-project-mooc,arbrandes/edx-platform,mahendra-r/edx-platform,y12uc231/edx-platform,zofuthan/edx-platform,sudheerchintala/LearnEraPlatForm,Ayub-Khan/edx-platform,ferabra/edx-platform,halvertoluke/edx-platform,beacloudgenius/edx-platform,inares/edx-platform,olexiim/edx-platform,beacloudgenius/edx-platform,hkawasaki/kawasaki-aio8-1,B-MOOC/edx-platform,vasyarv/edx-platform,wwj718/ANALYSE,arbrandes/edx-platform,4eek/edx-platform,WatanabeYasumasa/edx-platform,waheedahmed/edx-platform,ahmadio/edx-platform,a-parhom/edx-platform,valtech-mooc/edx-platform,eemirtekin/edx-platform,jruiperezv/ANALYSE,eduNEXT/edunext-platform,eduNEXT/edx-platform,bitifirefly/edx-platform,don-github/edx-platform,nttks/edx-platform,a-parhom/edx-platform,mjirayu/sit_academy,rismalrv/edx-platform,etzhou/edx-platform,inares/edx-platform,teltek/edx-platform,chauhanhardik/populo,kxliugang/edx-platform,IndonesiaX/edx-platform,jruiperezv/ANALYSE,shubhdev/edx-platform,Lektorium-LLC/edx-platform,ubc/edx-platform,ZLLab-Mooc/edx-platform,tanmaykm/edx-platform,jruiperezv/ANALYSE,devs1991/test_edx_docmode,jbzdak/edx-platform,rue89-tech/edx-platform,edx-solutions/edx-platform,doganov/edx-platform,edry/edx-platform,jamiefolsom/edx-platform,pabloborrego93/edx-platform,chrisndodge/edx-platform,SravanthiSinha/edx-platform,antonve/s4-project-mooc,proversity-org/edx-platform,jelugbo/tundex,auferack08/edx-platform,synergeticsedx/deployment-wipro,arifsetiawan/edx-platform,cselis86/edx-platform,marcore/edx-platform,jamesblunt/edx-platform,SravanthiSinha/edx-platform,hkawasaki/kawasaki-aio8-0,shurihell/testasia,UXE/local-edx,LearnEra/LearnEraPlaftform,LearnEra/LearnEraPlaftform,naresh21/synergetics-edx-platform,xinjiguaike/edx-platform,jazkarta/edx-platform,vismartltd/edx-platform,msegado/edx-platform,nikolas/edx-platform,jazkarta/edx-platform-for-isc,torchingloom/edx-platform,olexiim/edx-platform,wwj718/edx-platform,jazkarta/edx-platform,peterm-itr/edx-platform,louyihua/edx-platform,xinjiguaike/edx-platform,fly19890211/edx-platform,pepeportela/edx-platform,MakeHer/edx-platform,Edraak/edx-platform,yokose-ks/edx-platform,etzhou/edx-platform,BehavioralInsightsTeam/edx-platform,Endika/edx-platform,utecuy/edx-platform,waheedahmed/edx-platform,don-github/edx-platform,auferack08/edx-platform,hamzehd/edx-platform,utecuy/edx-platform,jbassen/edx-platform,chauhanhardik/populo,kamalx/edx-platform,atsolakid/edx-platform,bitifirefly/edx-platform,kursitet/edx-platform,chrisndodge/edx-platform,Shrhawk/edx-platform,UXE/local-edx,Unow/edx-platform,Endika/edx-platform,valtech-mooc/edx-platform,bigdatauniversity/edx-platform,eemirtekin/edx-platform,jamiefolsom/edx-platform,MakeHer/edx-platform,Ayub-Khan/edx-platform,jjmiranda/edx-platform,kxliugang/edx-platform,SravanthiSinha/edx-platform,10clouds/edx-platform,y12uc231/edx-platform,ZLLab-Mooc/edx-platform,ahmadiga/min_edx,Kalyzee/edx-platform,sameetb-cuelogic/edx-platform-test,ampax/edx-platform-backup,miptliot/edx-platform,jzoldak/edx-platform,cpennington/edx-platform,hastexo/edx-platform,ahmedaljazzar/edx-platform,eduNEXT/edunext-platform,don-github/edx-platform,arbrandes/edx-platform,10clouds/edx-platform,DefyVentures/edx-platform,yokose-ks/edx-platform,unicri/edx-platform,chudaol/edx-platform,chand3040/cloud_that,10clouds/edx-platform,eestay/edx-platform,jswope00/GAI,etzhou/edx-platform,zubair-arbi/edx-platform,analyseuc3m/ANALYSE-v1,morenopc/edx-platform,JCBarahona/edX,carsongee/edx-platform,kmoocdev/edx-platform,simbs/edx-platform,utecuy/edx-platform,deepsrijit1105/edx-platform,shabab12/edx-platform,yokose-ks/edx-platform,jswope00/griffinx,lduarte1991/edx-platform,romain-li/edx-platform,wwj718/edx-platform,zadgroup/edx-platform,fly19890211/edx-platform,Edraak/edx-platform,Edraak/edx-platform,WatanabeYasumasa/edx-platform,sameetb-cuelogic/edx-platform-test,jazkarta/edx-platform-for-isc,arbrandes/edx-platform,prarthitm/edxplatform,mushtaqak/edx-platform,CourseTalk/edx-platform,bdero/edx-platform,kmoocdev2/edx-platform,IONISx/edx-platform,cpennington/edx-platform,cyanna/edx-platform,4eek/edx-platform,polimediaupv/edx-platform,adoosii/edx-platform,nanolearningllc/edx-platform-cypress,amir-qayyum-khan/edx-platform,devs1991/test_edx_docmode,UOMx/edx-platform,torchingloom/edx-platform,bigdatauniversity/edx-platform,andyzsf/edx,stvstnfrd/edx-platform,antoviaque/edx-platform,edx-solutions/edx-platform,andyzsf/edx,Shrhawk/edx-platform,dkarakats/edx-platform,ferabra/edx-platform,motion2015/a3,iivic/BoiseStateX,eemirtekin/edx-platform,teltek/edx-platform,WatanabeYasumasa/edx-platform,Endika/edx-platform,franosincic/edx-platform,romain-li/edx-platform,chauhanhardik/populo_2,mjirayu/sit_academy,ak2703/edx-platform,kmoocdev2/edx-platform,gsehub/edx-platform,halvertoluke/edx-platform,LICEF/edx-platform,fintech-circle/edx-platform,SivilTaram/edx-platform,Unow/edx-platform,waheedahmed/edx-platform,kxliugang/edx-platform,jjmiranda/edx-platform,benpatterson/edx-platform,Endika/edx-platform,nttks/edx-platform,teltek/edx-platform,xuxiao19910803/edx,peterm-itr/edx-platform,nttks/edx-platform,jazkarta/edx-platform,caesar2164/edx-platform,torchingloom/edx-platform,CredoReference/edx-platform,dsajkl/123,edry/edx-platform,Edraak/circleci-edx-platform,shubhdev/edxOnBaadal,iivic/BoiseStateX,chauhanhardik/populo_2,raccoongang/edx-platform,y12uc231/edx-platform,nttks/jenkins-test,pabloborrego93/edx-platform,playm2mboy/edx-platform,AkA84/edx-platform,nanolearning/edx-platform,xingyepei/edx-platform,martynovp/edx-platform,sudheerchintala/LearnEraPlatForm,mitocw/edx-platform,ferabra/edx-platform,Softmotions/edx-platform,zerobatu/edx-platform,DefyVentures/edx-platform,iivic/BoiseStateX,CredoReference/edx-platform,eduNEXT/edx-platform,doismellburning/edx-platform,shubhdev/openedx,DNFcode/edx-platform,waheedahmed/edx-platform,shurihell/testasia,y12uc231/edx-platform,dsajkl/123,gymnasium/edx-platform,eestay/edx-platform,dsajkl/reqiop,raccoongang/edx-platform,alexthered/kienhoc-platform,chauhanhardik/populo,Unow/edx-platform,kursitet/edx-platform,kmoocdev/edx-platform,ahmedaljazzar/edx-platform,philanthropy-u/edx-platform,hmcmooc/muddx-platform,dcosentino/edx-platform,auferack08/edx-platform,nikolas/edx-platform,adoosii/edx-platform,ubc/edx-platform,auferack08/edx-platform,MSOpenTech/edx-platform,mushtaqak/edx-platform,eduNEXT/edx-platform,MSOpenTech/edx-platform,Edraak/edraak-platform,zubair-arbi/edx-platform,motion2015/edx-platform,vismartltd/edx-platform,carsongee/edx-platform,mjirayu/sit_academy,TeachAtTUM/edx-platform,JCBarahona/edX,rue89-tech/edx-platform,ESOedX/edx-platform,shubhdev/edx-platform,Unow/edx-platform,nanolearningllc/edx-platform-cypress-2,cpennington/edx-platform,pku9104038/edx-platform,eduNEXT/edunext-platform,shubhdev/edx-platform,philanthropy-u/edx-platform,hastexo/edx-platform,motion2015/edx-platform,jbassen/edx-platform,longmen21/edx-platform,Edraak/edraak-platform,rismalrv/edx-platform,MakeHer/edx-platform,xingyepei/edx-platform,jamesblunt/edx-platform,RPI-OPENEDX/edx-platform,zhenzhai/edx-platform,zubair-arbi/edx-platform,pepeportela/edx-platform,nagyistoce/edx-platform,benpatterson/edx-platform,vismartltd/edx-platform,Semi-global/edx-platform,jazkarta/edx-platform-for-isc,DNFcode/edx-platform,procangroup/edx-platform,ahmadio/edx-platform,naresh21/synergetics-edx-platform,IONISx/edx-platform,torchingloom/edx-platform,leansoft/edx-platform,shashank971/edx-platform,msegado/edx-platform,longmen21/edx-platform,zofuthan/edx-platform,appsembler/edx-platform,AkA84/edx-platform,IndonesiaX/edx-platform,dkarakats/edx-platform,IndonesiaX/edx-platform,DNFcode/edx-platform,eestay/edx-platform,nanolearningllc/edx-platform-cypress,pabloborrego93/edx-platform,motion2015/a3,tiagochiavericosta/edx-platform,Semi-global/edx-platform,louyihua/edx-platform,knehez/edx-platform,wwj718/edx-platform,JioEducation/edx-platform,pomegranited/edx-platform,cselis86/edx-platform,nttks/jenkins-test,mtlchun/edx,jazkarta/edx-platform-for-isc,jbzdak/edx-platform,Livit/Livit.Learn.EdX,bigdatauniversity/edx-platform,pepeportela/edx-platform,wwj718/ANALYSE,DNFcode/edx-platform,rhndg/openedx,hastexo/edx-platform
import ConfigParser from django.conf import settings def doc_url(request): config_file = open(settings.REPO_ROOT / "docs" / "config.ini") config = ConfigParser.ConfigParser() config.readfp(config_file) # in the future, we will detect the locale; for now, we will # hardcode en_us, since we only have English documentation locale = "en_us" def get_doc_url(token): try: return config.get(locale, token) except ConfigParser.NoOptionError: return config.get(locale, "default") return {"doc_url": get_doc_url} Read from doc url mapping file at load time, rather than once per request
import ConfigParser from django.conf import settings config_file = open(settings.REPO_ROOT / "docs" / "config.ini") config = ConfigParser.ConfigParser() config.readfp(config_file) def doc_url(request): # in the future, we will detect the locale; for now, we will # hardcode en_us, since we only have English documentation locale = "en_us" def get_doc_url(token): try: return config.get(locale, token) except ConfigParser.NoOptionError: return config.get(locale, "default") return {"doc_url": get_doc_url}
<commit_before>import ConfigParser from django.conf import settings def doc_url(request): config_file = open(settings.REPO_ROOT / "docs" / "config.ini") config = ConfigParser.ConfigParser() config.readfp(config_file) # in the future, we will detect the locale; for now, we will # hardcode en_us, since we only have English documentation locale = "en_us" def get_doc_url(token): try: return config.get(locale, token) except ConfigParser.NoOptionError: return config.get(locale, "default") return {"doc_url": get_doc_url} <commit_msg>Read from doc url mapping file at load time, rather than once per request<commit_after>
import ConfigParser from django.conf import settings config_file = open(settings.REPO_ROOT / "docs" / "config.ini") config = ConfigParser.ConfigParser() config.readfp(config_file) def doc_url(request): # in the future, we will detect the locale; for now, we will # hardcode en_us, since we only have English documentation locale = "en_us" def get_doc_url(token): try: return config.get(locale, token) except ConfigParser.NoOptionError: return config.get(locale, "default") return {"doc_url": get_doc_url}
import ConfigParser from django.conf import settings def doc_url(request): config_file = open(settings.REPO_ROOT / "docs" / "config.ini") config = ConfigParser.ConfigParser() config.readfp(config_file) # in the future, we will detect the locale; for now, we will # hardcode en_us, since we only have English documentation locale = "en_us" def get_doc_url(token): try: return config.get(locale, token) except ConfigParser.NoOptionError: return config.get(locale, "default") return {"doc_url": get_doc_url} Read from doc url mapping file at load time, rather than once per requestimport ConfigParser from django.conf import settings config_file = open(settings.REPO_ROOT / "docs" / "config.ini") config = ConfigParser.ConfigParser() config.readfp(config_file) def doc_url(request): # in the future, we will detect the locale; for now, we will # hardcode en_us, since we only have English documentation locale = "en_us" def get_doc_url(token): try: return config.get(locale, token) except ConfigParser.NoOptionError: return config.get(locale, "default") return {"doc_url": get_doc_url}
<commit_before>import ConfigParser from django.conf import settings def doc_url(request): config_file = open(settings.REPO_ROOT / "docs" / "config.ini") config = ConfigParser.ConfigParser() config.readfp(config_file) # in the future, we will detect the locale; for now, we will # hardcode en_us, since we only have English documentation locale = "en_us" def get_doc_url(token): try: return config.get(locale, token) except ConfigParser.NoOptionError: return config.get(locale, "default") return {"doc_url": get_doc_url} <commit_msg>Read from doc url mapping file at load time, rather than once per request<commit_after>import ConfigParser from django.conf import settings config_file = open(settings.REPO_ROOT / "docs" / "config.ini") config = ConfigParser.ConfigParser() config.readfp(config_file) def doc_url(request): # in the future, we will detect the locale; for now, we will # hardcode en_us, since we only have English documentation locale = "en_us" def get_doc_url(token): try: return config.get(locale, token) except ConfigParser.NoOptionError: return config.get(locale, "default") return {"doc_url": get_doc_url}
09f649ac0b14269067c43df9f879d963ab99cdac
backend/breach/views.py
backend/breach/views.py
import json from django.http import Http404, JsonResponse from django.views.decorators.csrf import csrf_exempt from breach.strategy import Strategy from breach.models import Victim def get_work(request, victim_id=0): assert(victim_id) try: victim = Victim.objects.get(pk=victim_id) except: raise Http404('Victim not found') strategy = Strategy(victim) # Example work structure: # return {'url': 'https://www.dimkarakostas.com/?breach-test', # 'amount': 10, # 'timeout': 0} new_work = strategy.get_work() return HttpResponse(json.dumps(new_work), content_type='application/json') @csrf_exempt def work_completed(request, victim_id=0): assert(victim_id) try: victim = Victim.objects.get(pk=victim_id) except: raise Http404('Victim not found') strategy = Strategy(victim) victory = strategy.work_completed() return JsonResponse({ 'victory': victory })
import json from django.http import Http404, JsonResponse from django.views.decorators.csrf import csrf_exempt from breach.strategy import Strategy from breach.models import Victim def get_work(request, victim_id=0): assert(victim_id) try: victim = Victim.objects.get(pk=victim_id) except: raise Http404('Victim not found') strategy = Strategy(victim) # Example work structure: # return {'url': 'https://www.dimkarakostas.com/?breach-test', # 'amount': 10, # 'timeout': 0} new_work = strategy.get_work() return JsonResponse(new_work) @csrf_exempt def work_completed(request, victim_id=0): assert(victim_id) try: victim = Victim.objects.get(pk=victim_id) except: raise Http404('Victim not found') strategy = Strategy(victim) victory = strategy.work_completed() return JsonResponse({ 'victory': victory })
Fix response with json for get_work
Fix response with json for get_work
Python
mit
dionyziz/rupture,dimkarakostas/rupture,dionyziz/rupture,dimkarakostas/rupture,dimriou/rupture,esarafianou/rupture,dimriou/rupture,esarafianou/rupture,dionyziz/rupture,dimkarakostas/rupture,esarafianou/rupture,dionyziz/rupture,dimriou/rupture,dimkarakostas/rupture,dionyziz/rupture,dimriou/rupture,dimkarakostas/rupture,dimriou/rupture,esarafianou/rupture
import json from django.http import Http404, JsonResponse from django.views.decorators.csrf import csrf_exempt from breach.strategy import Strategy from breach.models import Victim def get_work(request, victim_id=0): assert(victim_id) try: victim = Victim.objects.get(pk=victim_id) except: raise Http404('Victim not found') strategy = Strategy(victim) # Example work structure: # return {'url': 'https://www.dimkarakostas.com/?breach-test', # 'amount': 10, # 'timeout': 0} new_work = strategy.get_work() return HttpResponse(json.dumps(new_work), content_type='application/json') @csrf_exempt def work_completed(request, victim_id=0): assert(victim_id) try: victim = Victim.objects.get(pk=victim_id) except: raise Http404('Victim not found') strategy = Strategy(victim) victory = strategy.work_completed() return JsonResponse({ 'victory': victory }) Fix response with json for get_work
import json from django.http import Http404, JsonResponse from django.views.decorators.csrf import csrf_exempt from breach.strategy import Strategy from breach.models import Victim def get_work(request, victim_id=0): assert(victim_id) try: victim = Victim.objects.get(pk=victim_id) except: raise Http404('Victim not found') strategy = Strategy(victim) # Example work structure: # return {'url': 'https://www.dimkarakostas.com/?breach-test', # 'amount': 10, # 'timeout': 0} new_work = strategy.get_work() return JsonResponse(new_work) @csrf_exempt def work_completed(request, victim_id=0): assert(victim_id) try: victim = Victim.objects.get(pk=victim_id) except: raise Http404('Victim not found') strategy = Strategy(victim) victory = strategy.work_completed() return JsonResponse({ 'victory': victory })
<commit_before>import json from django.http import Http404, JsonResponse from django.views.decorators.csrf import csrf_exempt from breach.strategy import Strategy from breach.models import Victim def get_work(request, victim_id=0): assert(victim_id) try: victim = Victim.objects.get(pk=victim_id) except: raise Http404('Victim not found') strategy = Strategy(victim) # Example work structure: # return {'url': 'https://www.dimkarakostas.com/?breach-test', # 'amount': 10, # 'timeout': 0} new_work = strategy.get_work() return HttpResponse(json.dumps(new_work), content_type='application/json') @csrf_exempt def work_completed(request, victim_id=0): assert(victim_id) try: victim = Victim.objects.get(pk=victim_id) except: raise Http404('Victim not found') strategy = Strategy(victim) victory = strategy.work_completed() return JsonResponse({ 'victory': victory }) <commit_msg>Fix response with json for get_work<commit_after>
import json from django.http import Http404, JsonResponse from django.views.decorators.csrf import csrf_exempt from breach.strategy import Strategy from breach.models import Victim def get_work(request, victim_id=0): assert(victim_id) try: victim = Victim.objects.get(pk=victim_id) except: raise Http404('Victim not found') strategy = Strategy(victim) # Example work structure: # return {'url': 'https://www.dimkarakostas.com/?breach-test', # 'amount': 10, # 'timeout': 0} new_work = strategy.get_work() return JsonResponse(new_work) @csrf_exempt def work_completed(request, victim_id=0): assert(victim_id) try: victim = Victim.objects.get(pk=victim_id) except: raise Http404('Victim not found') strategy = Strategy(victim) victory = strategy.work_completed() return JsonResponse({ 'victory': victory })
import json from django.http import Http404, JsonResponse from django.views.decorators.csrf import csrf_exempt from breach.strategy import Strategy from breach.models import Victim def get_work(request, victim_id=0): assert(victim_id) try: victim = Victim.objects.get(pk=victim_id) except: raise Http404('Victim not found') strategy = Strategy(victim) # Example work structure: # return {'url': 'https://www.dimkarakostas.com/?breach-test', # 'amount': 10, # 'timeout': 0} new_work = strategy.get_work() return HttpResponse(json.dumps(new_work), content_type='application/json') @csrf_exempt def work_completed(request, victim_id=0): assert(victim_id) try: victim = Victim.objects.get(pk=victim_id) except: raise Http404('Victim not found') strategy = Strategy(victim) victory = strategy.work_completed() return JsonResponse({ 'victory': victory }) Fix response with json for get_workimport json from django.http import Http404, JsonResponse from django.views.decorators.csrf import csrf_exempt from breach.strategy import Strategy from breach.models import Victim def get_work(request, victim_id=0): assert(victim_id) try: victim = Victim.objects.get(pk=victim_id) except: raise Http404('Victim not found') strategy = Strategy(victim) # Example work structure: # return {'url': 'https://www.dimkarakostas.com/?breach-test', # 'amount': 10, # 'timeout': 0} new_work = strategy.get_work() return JsonResponse(new_work) @csrf_exempt def work_completed(request, victim_id=0): assert(victim_id) try: victim = Victim.objects.get(pk=victim_id) except: raise Http404('Victim not found') strategy = Strategy(victim) victory = strategy.work_completed() return JsonResponse({ 'victory': victory })
<commit_before>import json from django.http import Http404, JsonResponse from django.views.decorators.csrf import csrf_exempt from breach.strategy import Strategy from breach.models import Victim def get_work(request, victim_id=0): assert(victim_id) try: victim = Victim.objects.get(pk=victim_id) except: raise Http404('Victim not found') strategy = Strategy(victim) # Example work structure: # return {'url': 'https://www.dimkarakostas.com/?breach-test', # 'amount': 10, # 'timeout': 0} new_work = strategy.get_work() return HttpResponse(json.dumps(new_work), content_type='application/json') @csrf_exempt def work_completed(request, victim_id=0): assert(victim_id) try: victim = Victim.objects.get(pk=victim_id) except: raise Http404('Victim not found') strategy = Strategy(victim) victory = strategy.work_completed() return JsonResponse({ 'victory': victory }) <commit_msg>Fix response with json for get_work<commit_after>import json from django.http import Http404, JsonResponse from django.views.decorators.csrf import csrf_exempt from breach.strategy import Strategy from breach.models import Victim def get_work(request, victim_id=0): assert(victim_id) try: victim = Victim.objects.get(pk=victim_id) except: raise Http404('Victim not found') strategy = Strategy(victim) # Example work structure: # return {'url': 'https://www.dimkarakostas.com/?breach-test', # 'amount': 10, # 'timeout': 0} new_work = strategy.get_work() return JsonResponse(new_work) @csrf_exempt def work_completed(request, victim_id=0): assert(victim_id) try: victim = Victim.objects.get(pk=victim_id) except: raise Http404('Victim not found') strategy = Strategy(victim) victory = strategy.work_completed() return JsonResponse({ 'victory': victory })
0d5c3b5f0c9278e834fc4df2a5d227972a1b513d
tests/unit/modules/file_test.py
tests/unit/modules/file_test.py
import tempfile from saltunittest import TestCase, TestLoader, TextTestRunner from salt import config as sconfig from salt.modules import file as filemod from salt.modules import cmdmod filemod.__salt__ = { 'cmd.run': cmdmod.run, } SED_CONTENT = """test some content /var/lib/foo/app/test here """ class FileModuleTestCase(TestCase): def test_sed_limit_escaped(self): with tempfile.NamedTemporaryFile() as tfile: tfile.write(SED_CONTENT) tfile.seek(0, 0) path = tfile.name before = '/var/lib/foo' after = '' limit = '^{0}'.format(before) filemod.sed(path, before, after, limit=limit) with open(path, 'rb') as newfile: self.assertEquals(SED_CONTENT.replace(before, ''), newfile.read()) if __name__ == "__main__": loader = TestLoader() tests = loader.loadTestsFromTestCase(FileModuleTestCase) TextTestRunner(verbosity=1).run(tests)
import tempfile from saltunittest import TestCase, TestLoader, TextTestRunner from salt import config as sconfig from salt.modules import file as filemod from salt.modules import cmdmod filemod.__salt__ = { 'cmd.run': cmdmod.run, 'cmd.run_all': cmdmod.run_all } SED_CONTENT = """test some content /var/lib/foo/app/test here """ class FileModuleTestCase(TestCase): def test_sed_limit_escaped(self): with tempfile.NamedTemporaryFile() as tfile: tfile.write(SED_CONTENT) tfile.seek(0, 0) path = tfile.name before = '/var/lib/foo' after = '' limit = '^{0}'.format(before) filemod.sed(path, before, after, limit=limit) with open(path, 'rb') as newfile: self.assertEquals( SED_CONTENT.replace(before, ''), newfile.read() ) if __name__ == "__main__": loader = TestLoader() tests = loader.loadTestsFromTestCase(FileModuleTestCase) TextTestRunner(verbosity=1).run(tests)
Add `cmd.run_all` to `__salt__`. Required for the unit test.
Add `cmd.run_all` to `__salt__`. Required for the unit test.
Python
apache-2.0
saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt
import tempfile from saltunittest import TestCase, TestLoader, TextTestRunner from salt import config as sconfig from salt.modules import file as filemod from salt.modules import cmdmod filemod.__salt__ = { 'cmd.run': cmdmod.run, } SED_CONTENT = """test some content /var/lib/foo/app/test here """ class FileModuleTestCase(TestCase): def test_sed_limit_escaped(self): with tempfile.NamedTemporaryFile() as tfile: tfile.write(SED_CONTENT) tfile.seek(0, 0) path = tfile.name before = '/var/lib/foo' after = '' limit = '^{0}'.format(before) filemod.sed(path, before, after, limit=limit) with open(path, 'rb') as newfile: self.assertEquals(SED_CONTENT.replace(before, ''), newfile.read()) if __name__ == "__main__": loader = TestLoader() tests = loader.loadTestsFromTestCase(FileModuleTestCase) TextTestRunner(verbosity=1).run(tests) Add `cmd.run_all` to `__salt__`. Required for the unit test.
import tempfile from saltunittest import TestCase, TestLoader, TextTestRunner from salt import config as sconfig from salt.modules import file as filemod from salt.modules import cmdmod filemod.__salt__ = { 'cmd.run': cmdmod.run, 'cmd.run_all': cmdmod.run_all } SED_CONTENT = """test some content /var/lib/foo/app/test here """ class FileModuleTestCase(TestCase): def test_sed_limit_escaped(self): with tempfile.NamedTemporaryFile() as tfile: tfile.write(SED_CONTENT) tfile.seek(0, 0) path = tfile.name before = '/var/lib/foo' after = '' limit = '^{0}'.format(before) filemod.sed(path, before, after, limit=limit) with open(path, 'rb') as newfile: self.assertEquals( SED_CONTENT.replace(before, ''), newfile.read() ) if __name__ == "__main__": loader = TestLoader() tests = loader.loadTestsFromTestCase(FileModuleTestCase) TextTestRunner(verbosity=1).run(tests)
<commit_before>import tempfile from saltunittest import TestCase, TestLoader, TextTestRunner from salt import config as sconfig from salt.modules import file as filemod from salt.modules import cmdmod filemod.__salt__ = { 'cmd.run': cmdmod.run, } SED_CONTENT = """test some content /var/lib/foo/app/test here """ class FileModuleTestCase(TestCase): def test_sed_limit_escaped(self): with tempfile.NamedTemporaryFile() as tfile: tfile.write(SED_CONTENT) tfile.seek(0, 0) path = tfile.name before = '/var/lib/foo' after = '' limit = '^{0}'.format(before) filemod.sed(path, before, after, limit=limit) with open(path, 'rb') as newfile: self.assertEquals(SED_CONTENT.replace(before, ''), newfile.read()) if __name__ == "__main__": loader = TestLoader() tests = loader.loadTestsFromTestCase(FileModuleTestCase) TextTestRunner(verbosity=1).run(tests) <commit_msg>Add `cmd.run_all` to `__salt__`. Required for the unit test.<commit_after>
import tempfile from saltunittest import TestCase, TestLoader, TextTestRunner from salt import config as sconfig from salt.modules import file as filemod from salt.modules import cmdmod filemod.__salt__ = { 'cmd.run': cmdmod.run, 'cmd.run_all': cmdmod.run_all } SED_CONTENT = """test some content /var/lib/foo/app/test here """ class FileModuleTestCase(TestCase): def test_sed_limit_escaped(self): with tempfile.NamedTemporaryFile() as tfile: tfile.write(SED_CONTENT) tfile.seek(0, 0) path = tfile.name before = '/var/lib/foo' after = '' limit = '^{0}'.format(before) filemod.sed(path, before, after, limit=limit) with open(path, 'rb') as newfile: self.assertEquals( SED_CONTENT.replace(before, ''), newfile.read() ) if __name__ == "__main__": loader = TestLoader() tests = loader.loadTestsFromTestCase(FileModuleTestCase) TextTestRunner(verbosity=1).run(tests)
import tempfile from saltunittest import TestCase, TestLoader, TextTestRunner from salt import config as sconfig from salt.modules import file as filemod from salt.modules import cmdmod filemod.__salt__ = { 'cmd.run': cmdmod.run, } SED_CONTENT = """test some content /var/lib/foo/app/test here """ class FileModuleTestCase(TestCase): def test_sed_limit_escaped(self): with tempfile.NamedTemporaryFile() as tfile: tfile.write(SED_CONTENT) tfile.seek(0, 0) path = tfile.name before = '/var/lib/foo' after = '' limit = '^{0}'.format(before) filemod.sed(path, before, after, limit=limit) with open(path, 'rb') as newfile: self.assertEquals(SED_CONTENT.replace(before, ''), newfile.read()) if __name__ == "__main__": loader = TestLoader() tests = loader.loadTestsFromTestCase(FileModuleTestCase) TextTestRunner(verbosity=1).run(tests) Add `cmd.run_all` to `__salt__`. Required for the unit test.import tempfile from saltunittest import TestCase, TestLoader, TextTestRunner from salt import config as sconfig from salt.modules import file as filemod from salt.modules import cmdmod filemod.__salt__ = { 'cmd.run': cmdmod.run, 'cmd.run_all': cmdmod.run_all } SED_CONTENT = """test some content /var/lib/foo/app/test here """ class FileModuleTestCase(TestCase): def test_sed_limit_escaped(self): with tempfile.NamedTemporaryFile() as tfile: tfile.write(SED_CONTENT) tfile.seek(0, 0) path = tfile.name before = '/var/lib/foo' after = '' limit = '^{0}'.format(before) filemod.sed(path, before, after, limit=limit) with open(path, 'rb') as newfile: self.assertEquals( SED_CONTENT.replace(before, ''), newfile.read() ) if __name__ == "__main__": loader = TestLoader() tests = loader.loadTestsFromTestCase(FileModuleTestCase) TextTestRunner(verbosity=1).run(tests)
<commit_before>import tempfile from saltunittest import TestCase, TestLoader, TextTestRunner from salt import config as sconfig from salt.modules import file as filemod from salt.modules import cmdmod filemod.__salt__ = { 'cmd.run': cmdmod.run, } SED_CONTENT = """test some content /var/lib/foo/app/test here """ class FileModuleTestCase(TestCase): def test_sed_limit_escaped(self): with tempfile.NamedTemporaryFile() as tfile: tfile.write(SED_CONTENT) tfile.seek(0, 0) path = tfile.name before = '/var/lib/foo' after = '' limit = '^{0}'.format(before) filemod.sed(path, before, after, limit=limit) with open(path, 'rb') as newfile: self.assertEquals(SED_CONTENT.replace(before, ''), newfile.read()) if __name__ == "__main__": loader = TestLoader() tests = loader.loadTestsFromTestCase(FileModuleTestCase) TextTestRunner(verbosity=1).run(tests) <commit_msg>Add `cmd.run_all` to `__salt__`. Required for the unit test.<commit_after>import tempfile from saltunittest import TestCase, TestLoader, TextTestRunner from salt import config as sconfig from salt.modules import file as filemod from salt.modules import cmdmod filemod.__salt__ = { 'cmd.run': cmdmod.run, 'cmd.run_all': cmdmod.run_all } SED_CONTENT = """test some content /var/lib/foo/app/test here """ class FileModuleTestCase(TestCase): def test_sed_limit_escaped(self): with tempfile.NamedTemporaryFile() as tfile: tfile.write(SED_CONTENT) tfile.seek(0, 0) path = tfile.name before = '/var/lib/foo' after = '' limit = '^{0}'.format(before) filemod.sed(path, before, after, limit=limit) with open(path, 'rb') as newfile: self.assertEquals( SED_CONTENT.replace(before, ''), newfile.read() ) if __name__ == "__main__": loader = TestLoader() tests = loader.loadTestsFromTestCase(FileModuleTestCase) TextTestRunner(verbosity=1).run(tests)
fc27b35dd1ac34b19bdc2d71dd71704bd9321b9d
src/test/ed/db/pythonnumbers.py
src/test/ed/db/pythonnumbers.py
import types db = connect( "test" ); t = db.pythonnumbers t.drop() thing = { "a" : 5 , "b" : 5.5 } assert( type( thing["a"] ) == types.IntType ); assert( type( thing["b"] ) == types.FloatType ); t.save( thing ) thing = t.findOne() assert( type( thing["b"] ) == types.FloatType ); assert( type( thing["a"] ) == types.IntType ); t.drop(); t.save( { "a" : 1 , "b" : 1.0 } ); assert( t.findOne() ); assert( t.findOne( { "a" : 1 } ) ); assert( t.findOne( { "b" : 1.0 } ) ); assert( not t.findOne( { "b" : 2.0 } ) ); assert( t.findOne( { "a" : 1.0 } ) ); assert( t.findOne( { "b" : 1 } ) );
import _10gen import types db = _10gen.connect( "test" ); t = db.pythonnumbers t.drop() thing = { "a" : 5 , "b" : 5.5 } assert( type( thing["a"] ) == types.IntType ); assert( type( thing["b"] ) == types.FloatType ); t.save( thing ) thing = t.findOne() assert( type( thing["b"] ) == types.FloatType ); assert( type( thing["a"] ) == types.IntType ); t.drop(); t.save( { "a" : 1 , "b" : 1.0 } ); assert( t.findOne() ); assert( t.findOne( { "a" : 1 } ) ); assert( t.findOne( { "b" : 1.0 } ) ); assert( not t.findOne( { "b" : 2.0 } ) ); assert( t.findOne( { "a" : 1.0 } ) ); assert( t.findOne( { "b" : 1 } ) );
Fix test for unwelded scopes.
Fix test for unwelded scopes.
Python
apache-2.0
babble/babble,babble/babble,babble/babble,babble/babble,babble/babble,babble/babble
import types db = connect( "test" ); t = db.pythonnumbers t.drop() thing = { "a" : 5 , "b" : 5.5 } assert( type( thing["a"] ) == types.IntType ); assert( type( thing["b"] ) == types.FloatType ); t.save( thing ) thing = t.findOne() assert( type( thing["b"] ) == types.FloatType ); assert( type( thing["a"] ) == types.IntType ); t.drop(); t.save( { "a" : 1 , "b" : 1.0 } ); assert( t.findOne() ); assert( t.findOne( { "a" : 1 } ) ); assert( t.findOne( { "b" : 1.0 } ) ); assert( not t.findOne( { "b" : 2.0 } ) ); assert( t.findOne( { "a" : 1.0 } ) ); assert( t.findOne( { "b" : 1 } ) ); Fix test for unwelded scopes.
import _10gen import types db = _10gen.connect( "test" ); t = db.pythonnumbers t.drop() thing = { "a" : 5 , "b" : 5.5 } assert( type( thing["a"] ) == types.IntType ); assert( type( thing["b"] ) == types.FloatType ); t.save( thing ) thing = t.findOne() assert( type( thing["b"] ) == types.FloatType ); assert( type( thing["a"] ) == types.IntType ); t.drop(); t.save( { "a" : 1 , "b" : 1.0 } ); assert( t.findOne() ); assert( t.findOne( { "a" : 1 } ) ); assert( t.findOne( { "b" : 1.0 } ) ); assert( not t.findOne( { "b" : 2.0 } ) ); assert( t.findOne( { "a" : 1.0 } ) ); assert( t.findOne( { "b" : 1 } ) );
<commit_before> import types db = connect( "test" ); t = db.pythonnumbers t.drop() thing = { "a" : 5 , "b" : 5.5 } assert( type( thing["a"] ) == types.IntType ); assert( type( thing["b"] ) == types.FloatType ); t.save( thing ) thing = t.findOne() assert( type( thing["b"] ) == types.FloatType ); assert( type( thing["a"] ) == types.IntType ); t.drop(); t.save( { "a" : 1 , "b" : 1.0 } ); assert( t.findOne() ); assert( t.findOne( { "a" : 1 } ) ); assert( t.findOne( { "b" : 1.0 } ) ); assert( not t.findOne( { "b" : 2.0 } ) ); assert( t.findOne( { "a" : 1.0 } ) ); assert( t.findOne( { "b" : 1 } ) ); <commit_msg>Fix test for unwelded scopes.<commit_after>
import _10gen import types db = _10gen.connect( "test" ); t = db.pythonnumbers t.drop() thing = { "a" : 5 , "b" : 5.5 } assert( type( thing["a"] ) == types.IntType ); assert( type( thing["b"] ) == types.FloatType ); t.save( thing ) thing = t.findOne() assert( type( thing["b"] ) == types.FloatType ); assert( type( thing["a"] ) == types.IntType ); t.drop(); t.save( { "a" : 1 , "b" : 1.0 } ); assert( t.findOne() ); assert( t.findOne( { "a" : 1 } ) ); assert( t.findOne( { "b" : 1.0 } ) ); assert( not t.findOne( { "b" : 2.0 } ) ); assert( t.findOne( { "a" : 1.0 } ) ); assert( t.findOne( { "b" : 1 } ) );
import types db = connect( "test" ); t = db.pythonnumbers t.drop() thing = { "a" : 5 , "b" : 5.5 } assert( type( thing["a"] ) == types.IntType ); assert( type( thing["b"] ) == types.FloatType ); t.save( thing ) thing = t.findOne() assert( type( thing["b"] ) == types.FloatType ); assert( type( thing["a"] ) == types.IntType ); t.drop(); t.save( { "a" : 1 , "b" : 1.0 } ); assert( t.findOne() ); assert( t.findOne( { "a" : 1 } ) ); assert( t.findOne( { "b" : 1.0 } ) ); assert( not t.findOne( { "b" : 2.0 } ) ); assert( t.findOne( { "a" : 1.0 } ) ); assert( t.findOne( { "b" : 1 } ) ); Fix test for unwelded scopes. import _10gen import types db = _10gen.connect( "test" ); t = db.pythonnumbers t.drop() thing = { "a" : 5 , "b" : 5.5 } assert( type( thing["a"] ) == types.IntType ); assert( type( thing["b"] ) == types.FloatType ); t.save( thing ) thing = t.findOne() assert( type( thing["b"] ) == types.FloatType ); assert( type( thing["a"] ) == types.IntType ); t.drop(); t.save( { "a" : 1 , "b" : 1.0 } ); assert( t.findOne() ); assert( t.findOne( { "a" : 1 } ) ); assert( t.findOne( { "b" : 1.0 } ) ); assert( not t.findOne( { "b" : 2.0 } ) ); assert( t.findOne( { "a" : 1.0 } ) ); assert( t.findOne( { "b" : 1 } ) );
<commit_before> import types db = connect( "test" ); t = db.pythonnumbers t.drop() thing = { "a" : 5 , "b" : 5.5 } assert( type( thing["a"] ) == types.IntType ); assert( type( thing["b"] ) == types.FloatType ); t.save( thing ) thing = t.findOne() assert( type( thing["b"] ) == types.FloatType ); assert( type( thing["a"] ) == types.IntType ); t.drop(); t.save( { "a" : 1 , "b" : 1.0 } ); assert( t.findOne() ); assert( t.findOne( { "a" : 1 } ) ); assert( t.findOne( { "b" : 1.0 } ) ); assert( not t.findOne( { "b" : 2.0 } ) ); assert( t.findOne( { "a" : 1.0 } ) ); assert( t.findOne( { "b" : 1 } ) ); <commit_msg>Fix test for unwelded scopes.<commit_after> import _10gen import types db = _10gen.connect( "test" ); t = db.pythonnumbers t.drop() thing = { "a" : 5 , "b" : 5.5 } assert( type( thing["a"] ) == types.IntType ); assert( type( thing["b"] ) == types.FloatType ); t.save( thing ) thing = t.findOne() assert( type( thing["b"] ) == types.FloatType ); assert( type( thing["a"] ) == types.IntType ); t.drop(); t.save( { "a" : 1 , "b" : 1.0 } ); assert( t.findOne() ); assert( t.findOne( { "a" : 1 } ) ); assert( t.findOne( { "b" : 1.0 } ) ); assert( not t.findOne( { "b" : 2.0 } ) ); assert( t.findOne( { "a" : 1.0 } ) ); assert( t.findOne( { "b" : 1 } ) );
94d99dea05bbd17220f3959a965cc6abd456bf12
demo/tests/conftest.py
demo/tests/conftest.py
"""Unit tests configuration file.""" import logging def pytest_configure(config): """Disable verbose output when running tests.""" logging.basicConfig(level=logging.DEBUG) terminal = config.pluginmanager.getplugin('terminal') base = terminal.TerminalReporter class QuietReporter(base): ""Reporter that only shows dots when running tests.""" def __init__(self, *args, **kwargs): super().__init__(*args, **kwargs) self.verbosity = 0 self.showlongtestinfo = False self.showfspath = False terminal.TerminalReporter = QuietReporter
"""Unit tests configuration file.""" import logging def pytest_configure(config): """Disable verbose output when running tests.""" logging.basicConfig(level=logging.DEBUG) terminal = config.pluginmanager.getplugin('terminal') base = terminal.TerminalReporter class QuietReporter(base): """Reporter that only shows dots when running tests.""" def __init__(self, *args, **kwargs): super().__init__(*args, **kwargs) self.verbosity = 0 self.showlongtestinfo = False self.showfspath = False terminal.TerminalReporter = QuietReporter
Deploy Travis CI build 976 to GitHub
Deploy Travis CI build 976 to GitHub
Python
mit
jacebrowning/template-python-demo
"""Unit tests configuration file.""" import logging def pytest_configure(config): """Disable verbose output when running tests.""" logging.basicConfig(level=logging.DEBUG) terminal = config.pluginmanager.getplugin('terminal') base = terminal.TerminalReporter class QuietReporter(base): ""Reporter that only shows dots when running tests.""" def __init__(self, *args, **kwargs): super().__init__(*args, **kwargs) self.verbosity = 0 self.showlongtestinfo = False self.showfspath = False terminal.TerminalReporter = QuietReporter Deploy Travis CI build 976 to GitHub
"""Unit tests configuration file.""" import logging def pytest_configure(config): """Disable verbose output when running tests.""" logging.basicConfig(level=logging.DEBUG) terminal = config.pluginmanager.getplugin('terminal') base = terminal.TerminalReporter class QuietReporter(base): """Reporter that only shows dots when running tests.""" def __init__(self, *args, **kwargs): super().__init__(*args, **kwargs) self.verbosity = 0 self.showlongtestinfo = False self.showfspath = False terminal.TerminalReporter = QuietReporter
<commit_before>"""Unit tests configuration file.""" import logging def pytest_configure(config): """Disable verbose output when running tests.""" logging.basicConfig(level=logging.DEBUG) terminal = config.pluginmanager.getplugin('terminal') base = terminal.TerminalReporter class QuietReporter(base): ""Reporter that only shows dots when running tests.""" def __init__(self, *args, **kwargs): super().__init__(*args, **kwargs) self.verbosity = 0 self.showlongtestinfo = False self.showfspath = False terminal.TerminalReporter = QuietReporter <commit_msg>Deploy Travis CI build 976 to GitHub<commit_after>
"""Unit tests configuration file.""" import logging def pytest_configure(config): """Disable verbose output when running tests.""" logging.basicConfig(level=logging.DEBUG) terminal = config.pluginmanager.getplugin('terminal') base = terminal.TerminalReporter class QuietReporter(base): """Reporter that only shows dots when running tests.""" def __init__(self, *args, **kwargs): super().__init__(*args, **kwargs) self.verbosity = 0 self.showlongtestinfo = False self.showfspath = False terminal.TerminalReporter = QuietReporter
"""Unit tests configuration file.""" import logging def pytest_configure(config): """Disable verbose output when running tests.""" logging.basicConfig(level=logging.DEBUG) terminal = config.pluginmanager.getplugin('terminal') base = terminal.TerminalReporter class QuietReporter(base): ""Reporter that only shows dots when running tests.""" def __init__(self, *args, **kwargs): super().__init__(*args, **kwargs) self.verbosity = 0 self.showlongtestinfo = False self.showfspath = False terminal.TerminalReporter = QuietReporter Deploy Travis CI build 976 to GitHub"""Unit tests configuration file.""" import logging def pytest_configure(config): """Disable verbose output when running tests.""" logging.basicConfig(level=logging.DEBUG) terminal = config.pluginmanager.getplugin('terminal') base = terminal.TerminalReporter class QuietReporter(base): """Reporter that only shows dots when running tests.""" def __init__(self, *args, **kwargs): super().__init__(*args, **kwargs) self.verbosity = 0 self.showlongtestinfo = False self.showfspath = False terminal.TerminalReporter = QuietReporter
<commit_before>"""Unit tests configuration file.""" import logging def pytest_configure(config): """Disable verbose output when running tests.""" logging.basicConfig(level=logging.DEBUG) terminal = config.pluginmanager.getplugin('terminal') base = terminal.TerminalReporter class QuietReporter(base): ""Reporter that only shows dots when running tests.""" def __init__(self, *args, **kwargs): super().__init__(*args, **kwargs) self.verbosity = 0 self.showlongtestinfo = False self.showfspath = False terminal.TerminalReporter = QuietReporter <commit_msg>Deploy Travis CI build 976 to GitHub<commit_after>"""Unit tests configuration file.""" import logging def pytest_configure(config): """Disable verbose output when running tests.""" logging.basicConfig(level=logging.DEBUG) terminal = config.pluginmanager.getplugin('terminal') base = terminal.TerminalReporter class QuietReporter(base): """Reporter that only shows dots when running tests.""" def __init__(self, *args, **kwargs): super().__init__(*args, **kwargs) self.verbosity = 0 self.showlongtestinfo = False self.showfspath = False terminal.TerminalReporter = QuietReporter
d87a44367c5542ab8052c212e6d51f1532086dd1
testsuite/python3.py
testsuite/python3.py
#!/usr/bin/env python3 from typing import ClassVar, List # Annotated function (Issue #29) def foo(x: int) -> int: return x + 1 # Annotated variables #575 CONST: int = 42 class Class: cls_var: ClassVar[str] for_var: ClassVar[str] while_var: ClassVar[str] def_var: ClassVar[str] if_var: ClassVar[str] elif_var: ClassVar[str] else_var: ClassVar[str] try_var: ClassVar[str] except_var: ClassVar[str] finally_var: ClassVar[str] with_var: ClassVar[str] For_var: ClassVar[str] While_var: ClassVar[str] Def_var: ClassVar[str] If_var: ClassVar[str] Elif_var: ClassVar[str] Else_var: ClassVar[str] Try_var: ClassVar[str] Except_var: ClassVar[str] Finally_var: ClassVar[str] With_var: ClassVar[str] def m(self): xs: List[int] = []
#!/usr/bin/env python3 from typing import ClassVar, List # Annotated function (Issue #29) def foo(x: int) -> int: return x + 1 # Annotated variables #575 CONST: int = 42 class Class: # Camel-caes cls_var: ClassVar[str] for_var: ClassVar[str] while_var: ClassVar[str] def_var: ClassVar[str] if_var: ClassVar[str] elif_var: ClassVar[str] else_var: ClassVar[str] try_var: ClassVar[str] except_var: ClassVar[str] finally_var: ClassVar[str] with_var: ClassVar[str] forVar: ClassVar[str] whileVar: ClassVar[str] defVar: ClassVar[str] ifVar: ClassVar[str] elifVar: ClassVar[str] elseVar: ClassVar[str] tryVar: ClassVar[str] exceptVar: ClassVar[str] finallyVar: ClassVar[str] withVar: ClassVar[str] def m(self): xs: List[int] = []
Make identifiers camel-case; remove redundant space.
Make identifiers camel-case; remove redundant space.
Python
mit
PyCQA/pep8
#!/usr/bin/env python3 from typing import ClassVar, List # Annotated function (Issue #29) def foo(x: int) -> int: return x + 1 # Annotated variables #575 CONST: int = 42 class Class: cls_var: ClassVar[str] for_var: ClassVar[str] while_var: ClassVar[str] def_var: ClassVar[str] if_var: ClassVar[str] elif_var: ClassVar[str] else_var: ClassVar[str] try_var: ClassVar[str] except_var: ClassVar[str] finally_var: ClassVar[str] with_var: ClassVar[str] For_var: ClassVar[str] While_var: ClassVar[str] Def_var: ClassVar[str] If_var: ClassVar[str] Elif_var: ClassVar[str] Else_var: ClassVar[str] Try_var: ClassVar[str] Except_var: ClassVar[str] Finally_var: ClassVar[str] With_var: ClassVar[str] def m(self): xs: List[int] = [] Make identifiers camel-case; remove redundant space.
#!/usr/bin/env python3 from typing import ClassVar, List # Annotated function (Issue #29) def foo(x: int) -> int: return x + 1 # Annotated variables #575 CONST: int = 42 class Class: # Camel-caes cls_var: ClassVar[str] for_var: ClassVar[str] while_var: ClassVar[str] def_var: ClassVar[str] if_var: ClassVar[str] elif_var: ClassVar[str] else_var: ClassVar[str] try_var: ClassVar[str] except_var: ClassVar[str] finally_var: ClassVar[str] with_var: ClassVar[str] forVar: ClassVar[str] whileVar: ClassVar[str] defVar: ClassVar[str] ifVar: ClassVar[str] elifVar: ClassVar[str] elseVar: ClassVar[str] tryVar: ClassVar[str] exceptVar: ClassVar[str] finallyVar: ClassVar[str] withVar: ClassVar[str] def m(self): xs: List[int] = []
<commit_before>#!/usr/bin/env python3 from typing import ClassVar, List # Annotated function (Issue #29) def foo(x: int) -> int: return x + 1 # Annotated variables #575 CONST: int = 42 class Class: cls_var: ClassVar[str] for_var: ClassVar[str] while_var: ClassVar[str] def_var: ClassVar[str] if_var: ClassVar[str] elif_var: ClassVar[str] else_var: ClassVar[str] try_var: ClassVar[str] except_var: ClassVar[str] finally_var: ClassVar[str] with_var: ClassVar[str] For_var: ClassVar[str] While_var: ClassVar[str] Def_var: ClassVar[str] If_var: ClassVar[str] Elif_var: ClassVar[str] Else_var: ClassVar[str] Try_var: ClassVar[str] Except_var: ClassVar[str] Finally_var: ClassVar[str] With_var: ClassVar[str] def m(self): xs: List[int] = [] <commit_msg>Make identifiers camel-case; remove redundant space.<commit_after>
#!/usr/bin/env python3 from typing import ClassVar, List # Annotated function (Issue #29) def foo(x: int) -> int: return x + 1 # Annotated variables #575 CONST: int = 42 class Class: # Camel-caes cls_var: ClassVar[str] for_var: ClassVar[str] while_var: ClassVar[str] def_var: ClassVar[str] if_var: ClassVar[str] elif_var: ClassVar[str] else_var: ClassVar[str] try_var: ClassVar[str] except_var: ClassVar[str] finally_var: ClassVar[str] with_var: ClassVar[str] forVar: ClassVar[str] whileVar: ClassVar[str] defVar: ClassVar[str] ifVar: ClassVar[str] elifVar: ClassVar[str] elseVar: ClassVar[str] tryVar: ClassVar[str] exceptVar: ClassVar[str] finallyVar: ClassVar[str] withVar: ClassVar[str] def m(self): xs: List[int] = []
#!/usr/bin/env python3 from typing import ClassVar, List # Annotated function (Issue #29) def foo(x: int) -> int: return x + 1 # Annotated variables #575 CONST: int = 42 class Class: cls_var: ClassVar[str] for_var: ClassVar[str] while_var: ClassVar[str] def_var: ClassVar[str] if_var: ClassVar[str] elif_var: ClassVar[str] else_var: ClassVar[str] try_var: ClassVar[str] except_var: ClassVar[str] finally_var: ClassVar[str] with_var: ClassVar[str] For_var: ClassVar[str] While_var: ClassVar[str] Def_var: ClassVar[str] If_var: ClassVar[str] Elif_var: ClassVar[str] Else_var: ClassVar[str] Try_var: ClassVar[str] Except_var: ClassVar[str] Finally_var: ClassVar[str] With_var: ClassVar[str] def m(self): xs: List[int] = [] Make identifiers camel-case; remove redundant space.#!/usr/bin/env python3 from typing import ClassVar, List # Annotated function (Issue #29) def foo(x: int) -> int: return x + 1 # Annotated variables #575 CONST: int = 42 class Class: # Camel-caes cls_var: ClassVar[str] for_var: ClassVar[str] while_var: ClassVar[str] def_var: ClassVar[str] if_var: ClassVar[str] elif_var: ClassVar[str] else_var: ClassVar[str] try_var: ClassVar[str] except_var: ClassVar[str] finally_var: ClassVar[str] with_var: ClassVar[str] forVar: ClassVar[str] whileVar: ClassVar[str] defVar: ClassVar[str] ifVar: ClassVar[str] elifVar: ClassVar[str] elseVar: ClassVar[str] tryVar: ClassVar[str] exceptVar: ClassVar[str] finallyVar: ClassVar[str] withVar: ClassVar[str] def m(self): xs: List[int] = []
<commit_before>#!/usr/bin/env python3 from typing import ClassVar, List # Annotated function (Issue #29) def foo(x: int) -> int: return x + 1 # Annotated variables #575 CONST: int = 42 class Class: cls_var: ClassVar[str] for_var: ClassVar[str] while_var: ClassVar[str] def_var: ClassVar[str] if_var: ClassVar[str] elif_var: ClassVar[str] else_var: ClassVar[str] try_var: ClassVar[str] except_var: ClassVar[str] finally_var: ClassVar[str] with_var: ClassVar[str] For_var: ClassVar[str] While_var: ClassVar[str] Def_var: ClassVar[str] If_var: ClassVar[str] Elif_var: ClassVar[str] Else_var: ClassVar[str] Try_var: ClassVar[str] Except_var: ClassVar[str] Finally_var: ClassVar[str] With_var: ClassVar[str] def m(self): xs: List[int] = [] <commit_msg>Make identifiers camel-case; remove redundant space.<commit_after>#!/usr/bin/env python3 from typing import ClassVar, List # Annotated function (Issue #29) def foo(x: int) -> int: return x + 1 # Annotated variables #575 CONST: int = 42 class Class: # Camel-caes cls_var: ClassVar[str] for_var: ClassVar[str] while_var: ClassVar[str] def_var: ClassVar[str] if_var: ClassVar[str] elif_var: ClassVar[str] else_var: ClassVar[str] try_var: ClassVar[str] except_var: ClassVar[str] finally_var: ClassVar[str] with_var: ClassVar[str] forVar: ClassVar[str] whileVar: ClassVar[str] defVar: ClassVar[str] ifVar: ClassVar[str] elifVar: ClassVar[str] elseVar: ClassVar[str] tryVar: ClassVar[str] exceptVar: ClassVar[str] finallyVar: ClassVar[str] withVar: ClassVar[str] def m(self): xs: List[int] = []
0d96ff52ca66de8afd95fb6dc342e8529764e89b
tflitehub/lit.cfg.py
tflitehub/lit.cfg.py
import os import sys import lit.formats import lit.util import lit.llvm # Configuration file for the 'lit' test runner. lit.llvm.initialize(lit_config, config) # name: The name of this test suite. config.name = 'TFLITEHUB' config.test_format = lit.formats.ShTest() # suffixes: A list of file extensions to treat as test files. config.suffixes = ['.py'] # test_source_root: The root path where tests are located. config.test_source_root = os.path.dirname(__file__) #config.use_default_substitutions() config.excludes = [ 'lit.cfg.py', 'lit.site.cfg.py', 'test_util.py', 'manual_test.py', ] config.substitutions.extend([ ('%PYTHON', sys.executable), ]) config.environment['PYTHONPATH'] = ":".join(sys.path) project_root = os.path.dirname(os.path.dirname(__file__)) # Enable features based on -D FEATURES=hugetest,vulkan # syntax. features_param = lit_config.params.get('FEATURES') if features_param: config.available_features.update(features_param.split(','))
import os import sys import lit.formats import lit.util import lit.llvm # Configuration file for the 'lit' test runner. lit.llvm.initialize(lit_config, config) # name: The name of this test suite. config.name = 'TFLITEHUB' config.test_format = lit.formats.ShTest() # suffixes: A list of file extensions to treat as test files. config.suffixes = ['.py'] # test_source_root: The root path where tests are located. config.test_source_root = os.path.dirname(__file__) #config.use_default_substitutions() config.excludes = [ 'imagenet_test_data.py', 'lit.cfg.py', 'lit.site.cfg.py', 'manual_test.py', 'squad_test_data.py', 'test_util.py', ] config.substitutions.extend([ ('%PYTHON', sys.executable), ]) config.environment['PYTHONPATH'] = ":".join(sys.path) project_root = os.path.dirname(os.path.dirname(__file__)) # Enable features based on -D FEATURES=hugetest,vulkan # syntax. features_param = lit_config.params.get('FEATURES') if features_param: config.available_features.update(features_param.split(','))
Exclude test data utilities from lit.
Exclude test data utilities from lit.
Python
apache-2.0
iree-org/iree-samples,iree-org/iree-samples,iree-org/iree-samples,iree-org/iree-samples
import os import sys import lit.formats import lit.util import lit.llvm # Configuration file for the 'lit' test runner. lit.llvm.initialize(lit_config, config) # name: The name of this test suite. config.name = 'TFLITEHUB' config.test_format = lit.formats.ShTest() # suffixes: A list of file extensions to treat as test files. config.suffixes = ['.py'] # test_source_root: The root path where tests are located. config.test_source_root = os.path.dirname(__file__) #config.use_default_substitutions() config.excludes = [ 'lit.cfg.py', 'lit.site.cfg.py', 'test_util.py', 'manual_test.py', ] config.substitutions.extend([ ('%PYTHON', sys.executable), ]) config.environment['PYTHONPATH'] = ":".join(sys.path) project_root = os.path.dirname(os.path.dirname(__file__)) # Enable features based on -D FEATURES=hugetest,vulkan # syntax. features_param = lit_config.params.get('FEATURES') if features_param: config.available_features.update(features_param.split(',')) Exclude test data utilities from lit.
import os import sys import lit.formats import lit.util import lit.llvm # Configuration file for the 'lit' test runner. lit.llvm.initialize(lit_config, config) # name: The name of this test suite. config.name = 'TFLITEHUB' config.test_format = lit.formats.ShTest() # suffixes: A list of file extensions to treat as test files. config.suffixes = ['.py'] # test_source_root: The root path where tests are located. config.test_source_root = os.path.dirname(__file__) #config.use_default_substitutions() config.excludes = [ 'imagenet_test_data.py', 'lit.cfg.py', 'lit.site.cfg.py', 'manual_test.py', 'squad_test_data.py', 'test_util.py', ] config.substitutions.extend([ ('%PYTHON', sys.executable), ]) config.environment['PYTHONPATH'] = ":".join(sys.path) project_root = os.path.dirname(os.path.dirname(__file__)) # Enable features based on -D FEATURES=hugetest,vulkan # syntax. features_param = lit_config.params.get('FEATURES') if features_param: config.available_features.update(features_param.split(','))
<commit_before>import os import sys import lit.formats import lit.util import lit.llvm # Configuration file for the 'lit' test runner. lit.llvm.initialize(lit_config, config) # name: The name of this test suite. config.name = 'TFLITEHUB' config.test_format = lit.formats.ShTest() # suffixes: A list of file extensions to treat as test files. config.suffixes = ['.py'] # test_source_root: The root path where tests are located. config.test_source_root = os.path.dirname(__file__) #config.use_default_substitutions() config.excludes = [ 'lit.cfg.py', 'lit.site.cfg.py', 'test_util.py', 'manual_test.py', ] config.substitutions.extend([ ('%PYTHON', sys.executable), ]) config.environment['PYTHONPATH'] = ":".join(sys.path) project_root = os.path.dirname(os.path.dirname(__file__)) # Enable features based on -D FEATURES=hugetest,vulkan # syntax. features_param = lit_config.params.get('FEATURES') if features_param: config.available_features.update(features_param.split(',')) <commit_msg>Exclude test data utilities from lit.<commit_after>
import os import sys import lit.formats import lit.util import lit.llvm # Configuration file for the 'lit' test runner. lit.llvm.initialize(lit_config, config) # name: The name of this test suite. config.name = 'TFLITEHUB' config.test_format = lit.formats.ShTest() # suffixes: A list of file extensions to treat as test files. config.suffixes = ['.py'] # test_source_root: The root path where tests are located. config.test_source_root = os.path.dirname(__file__) #config.use_default_substitutions() config.excludes = [ 'imagenet_test_data.py', 'lit.cfg.py', 'lit.site.cfg.py', 'manual_test.py', 'squad_test_data.py', 'test_util.py', ] config.substitutions.extend([ ('%PYTHON', sys.executable), ]) config.environment['PYTHONPATH'] = ":".join(sys.path) project_root = os.path.dirname(os.path.dirname(__file__)) # Enable features based on -D FEATURES=hugetest,vulkan # syntax. features_param = lit_config.params.get('FEATURES') if features_param: config.available_features.update(features_param.split(','))
import os import sys import lit.formats import lit.util import lit.llvm # Configuration file for the 'lit' test runner. lit.llvm.initialize(lit_config, config) # name: The name of this test suite. config.name = 'TFLITEHUB' config.test_format = lit.formats.ShTest() # suffixes: A list of file extensions to treat as test files. config.suffixes = ['.py'] # test_source_root: The root path where tests are located. config.test_source_root = os.path.dirname(__file__) #config.use_default_substitutions() config.excludes = [ 'lit.cfg.py', 'lit.site.cfg.py', 'test_util.py', 'manual_test.py', ] config.substitutions.extend([ ('%PYTHON', sys.executable), ]) config.environment['PYTHONPATH'] = ":".join(sys.path) project_root = os.path.dirname(os.path.dirname(__file__)) # Enable features based on -D FEATURES=hugetest,vulkan # syntax. features_param = lit_config.params.get('FEATURES') if features_param: config.available_features.update(features_param.split(',')) Exclude test data utilities from lit.import os import sys import lit.formats import lit.util import lit.llvm # Configuration file for the 'lit' test runner. lit.llvm.initialize(lit_config, config) # name: The name of this test suite. config.name = 'TFLITEHUB' config.test_format = lit.formats.ShTest() # suffixes: A list of file extensions to treat as test files. config.suffixes = ['.py'] # test_source_root: The root path where tests are located. config.test_source_root = os.path.dirname(__file__) #config.use_default_substitutions() config.excludes = [ 'imagenet_test_data.py', 'lit.cfg.py', 'lit.site.cfg.py', 'manual_test.py', 'squad_test_data.py', 'test_util.py', ] config.substitutions.extend([ ('%PYTHON', sys.executable), ]) config.environment['PYTHONPATH'] = ":".join(sys.path) project_root = os.path.dirname(os.path.dirname(__file__)) # Enable features based on -D FEATURES=hugetest,vulkan # syntax. features_param = lit_config.params.get('FEATURES') if features_param: config.available_features.update(features_param.split(','))
<commit_before>import os import sys import lit.formats import lit.util import lit.llvm # Configuration file for the 'lit' test runner. lit.llvm.initialize(lit_config, config) # name: The name of this test suite. config.name = 'TFLITEHUB' config.test_format = lit.formats.ShTest() # suffixes: A list of file extensions to treat as test files. config.suffixes = ['.py'] # test_source_root: The root path where tests are located. config.test_source_root = os.path.dirname(__file__) #config.use_default_substitutions() config.excludes = [ 'lit.cfg.py', 'lit.site.cfg.py', 'test_util.py', 'manual_test.py', ] config.substitutions.extend([ ('%PYTHON', sys.executable), ]) config.environment['PYTHONPATH'] = ":".join(sys.path) project_root = os.path.dirname(os.path.dirname(__file__)) # Enable features based on -D FEATURES=hugetest,vulkan # syntax. features_param = lit_config.params.get('FEATURES') if features_param: config.available_features.update(features_param.split(',')) <commit_msg>Exclude test data utilities from lit.<commit_after>import os import sys import lit.formats import lit.util import lit.llvm # Configuration file for the 'lit' test runner. lit.llvm.initialize(lit_config, config) # name: The name of this test suite. config.name = 'TFLITEHUB' config.test_format = lit.formats.ShTest() # suffixes: A list of file extensions to treat as test files. config.suffixes = ['.py'] # test_source_root: The root path where tests are located. config.test_source_root = os.path.dirname(__file__) #config.use_default_substitutions() config.excludes = [ 'imagenet_test_data.py', 'lit.cfg.py', 'lit.site.cfg.py', 'manual_test.py', 'squad_test_data.py', 'test_util.py', ] config.substitutions.extend([ ('%PYTHON', sys.executable), ]) config.environment['PYTHONPATH'] = ":".join(sys.path) project_root = os.path.dirname(os.path.dirname(__file__)) # Enable features based on -D FEATURES=hugetest,vulkan # syntax. features_param = lit_config.params.get('FEATURES') if features_param: config.available_features.update(features_param.split(','))
7a0b8550fa2f52519df81c7fa795d454e5e3b0bc
scripts/master/factory/dart/channels.py
scripts/master/factory/dart/channels.py
# Copyright 2013 The Chromium Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. class Channel(object): def __init__(self, name, branch, position, category_postfix, priority): self.branch = branch self.builder_postfix = '-' + name self.category_postfix = category_postfix self.name = name self.position = position self.priority = priority self.all_deps_path = '/' + branch + '/deps/all.deps' self.dartium_deps_path = '/' + branch + '/deps/dartium.deps' # The channel names are replicated in the slave.cfg files for all # dart waterfalls. If you change anything here please also change it there. CHANNELS = [ Channel('be', 'branches/bleeding_edge', 0, '', 3), Channel('dev', 'trunk', 1, '-dev', 2), Channel('stable', 'branches/0.6', 2, '-stable', 1), ] CHANNELS_BY_NAME = {} for c in CHANNELS: CHANNELS_BY_NAME[c.name] = c
# Copyright 2013 The Chromium Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. class Channel(object): def __init__(self, name, branch, position, category_postfix, priority): self.branch = branch self.builder_postfix = '-' + name self.category_postfix = category_postfix self.name = name self.position = position self.priority = priority self.all_deps_path = '/' + branch + '/deps/all.deps' self.dartium_deps_path = '/' + branch + '/deps/dartium.deps' # The channel names are replicated in the slave.cfg files for all # dart waterfalls. If you change anything here please also change it there. CHANNELS = [ Channel('be', 'branches/bleeding_edge', 0, '', 3), Channel('dev', 'trunk', 1, '-dev', 2), Channel('stable', 'branches/0.7', 2, '-stable', 1), ] CHANNELS_BY_NAME = {} for c in CHANNELS: CHANNELS_BY_NAME[c.name] = c
Update the build branch for stable to 0.7
Update the build branch for stable to 0.7 TBR=ricow Review URL: https://codereview.chromium.org/26993005 git-svn-id: 239fca9b83025a0b6f823aeeca02ba5be3d9fd76@228644 0039d316-1c4b-4281-b951-d872f2087c98
Python
bsd-3-clause
eunchong/build,eunchong/build,eunchong/build,eunchong/build
# Copyright 2013 The Chromium Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. class Channel(object): def __init__(self, name, branch, position, category_postfix, priority): self.branch = branch self.builder_postfix = '-' + name self.category_postfix = category_postfix self.name = name self.position = position self.priority = priority self.all_deps_path = '/' + branch + '/deps/all.deps' self.dartium_deps_path = '/' + branch + '/deps/dartium.deps' # The channel names are replicated in the slave.cfg files for all # dart waterfalls. If you change anything here please also change it there. CHANNELS = [ Channel('be', 'branches/bleeding_edge', 0, '', 3), Channel('dev', 'trunk', 1, '-dev', 2), Channel('stable', 'branches/0.6', 2, '-stable', 1), ] CHANNELS_BY_NAME = {} for c in CHANNELS: CHANNELS_BY_NAME[c.name] = c Update the build branch for stable to 0.7 TBR=ricow Review URL: https://codereview.chromium.org/26993005 git-svn-id: 239fca9b83025a0b6f823aeeca02ba5be3d9fd76@228644 0039d316-1c4b-4281-b951-d872f2087c98
# Copyright 2013 The Chromium Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. class Channel(object): def __init__(self, name, branch, position, category_postfix, priority): self.branch = branch self.builder_postfix = '-' + name self.category_postfix = category_postfix self.name = name self.position = position self.priority = priority self.all_deps_path = '/' + branch + '/deps/all.deps' self.dartium_deps_path = '/' + branch + '/deps/dartium.deps' # The channel names are replicated in the slave.cfg files for all # dart waterfalls. If you change anything here please also change it there. CHANNELS = [ Channel('be', 'branches/bleeding_edge', 0, '', 3), Channel('dev', 'trunk', 1, '-dev', 2), Channel('stable', 'branches/0.7', 2, '-stable', 1), ] CHANNELS_BY_NAME = {} for c in CHANNELS: CHANNELS_BY_NAME[c.name] = c
<commit_before># Copyright 2013 The Chromium Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. class Channel(object): def __init__(self, name, branch, position, category_postfix, priority): self.branch = branch self.builder_postfix = '-' + name self.category_postfix = category_postfix self.name = name self.position = position self.priority = priority self.all_deps_path = '/' + branch + '/deps/all.deps' self.dartium_deps_path = '/' + branch + '/deps/dartium.deps' # The channel names are replicated in the slave.cfg files for all # dart waterfalls. If you change anything here please also change it there. CHANNELS = [ Channel('be', 'branches/bleeding_edge', 0, '', 3), Channel('dev', 'trunk', 1, '-dev', 2), Channel('stable', 'branches/0.6', 2, '-stable', 1), ] CHANNELS_BY_NAME = {} for c in CHANNELS: CHANNELS_BY_NAME[c.name] = c <commit_msg>Update the build branch for stable to 0.7 TBR=ricow Review URL: https://codereview.chromium.org/26993005 git-svn-id: 239fca9b83025a0b6f823aeeca02ba5be3d9fd76@228644 0039d316-1c4b-4281-b951-d872f2087c98<commit_after>
# Copyright 2013 The Chromium Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. class Channel(object): def __init__(self, name, branch, position, category_postfix, priority): self.branch = branch self.builder_postfix = '-' + name self.category_postfix = category_postfix self.name = name self.position = position self.priority = priority self.all_deps_path = '/' + branch + '/deps/all.deps' self.dartium_deps_path = '/' + branch + '/deps/dartium.deps' # The channel names are replicated in the slave.cfg files for all # dart waterfalls. If you change anything here please also change it there. CHANNELS = [ Channel('be', 'branches/bleeding_edge', 0, '', 3), Channel('dev', 'trunk', 1, '-dev', 2), Channel('stable', 'branches/0.7', 2, '-stable', 1), ] CHANNELS_BY_NAME = {} for c in CHANNELS: CHANNELS_BY_NAME[c.name] = c
# Copyright 2013 The Chromium Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. class Channel(object): def __init__(self, name, branch, position, category_postfix, priority): self.branch = branch self.builder_postfix = '-' + name self.category_postfix = category_postfix self.name = name self.position = position self.priority = priority self.all_deps_path = '/' + branch + '/deps/all.deps' self.dartium_deps_path = '/' + branch + '/deps/dartium.deps' # The channel names are replicated in the slave.cfg files for all # dart waterfalls. If you change anything here please also change it there. CHANNELS = [ Channel('be', 'branches/bleeding_edge', 0, '', 3), Channel('dev', 'trunk', 1, '-dev', 2), Channel('stable', 'branches/0.6', 2, '-stable', 1), ] CHANNELS_BY_NAME = {} for c in CHANNELS: CHANNELS_BY_NAME[c.name] = c Update the build branch for stable to 0.7 TBR=ricow Review URL: https://codereview.chromium.org/26993005 git-svn-id: 239fca9b83025a0b6f823aeeca02ba5be3d9fd76@228644 0039d316-1c4b-4281-b951-d872f2087c98# Copyright 2013 The Chromium Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. class Channel(object): def __init__(self, name, branch, position, category_postfix, priority): self.branch = branch self.builder_postfix = '-' + name self.category_postfix = category_postfix self.name = name self.position = position self.priority = priority self.all_deps_path = '/' + branch + '/deps/all.deps' self.dartium_deps_path = '/' + branch + '/deps/dartium.deps' # The channel names are replicated in the slave.cfg files for all # dart waterfalls. If you change anything here please also change it there. CHANNELS = [ Channel('be', 'branches/bleeding_edge', 0, '', 3), Channel('dev', 'trunk', 1, '-dev', 2), Channel('stable', 'branches/0.7', 2, '-stable', 1), ] CHANNELS_BY_NAME = {} for c in CHANNELS: CHANNELS_BY_NAME[c.name] = c
<commit_before># Copyright 2013 The Chromium Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. class Channel(object): def __init__(self, name, branch, position, category_postfix, priority): self.branch = branch self.builder_postfix = '-' + name self.category_postfix = category_postfix self.name = name self.position = position self.priority = priority self.all_deps_path = '/' + branch + '/deps/all.deps' self.dartium_deps_path = '/' + branch + '/deps/dartium.deps' # The channel names are replicated in the slave.cfg files for all # dart waterfalls. If you change anything here please also change it there. CHANNELS = [ Channel('be', 'branches/bleeding_edge', 0, '', 3), Channel('dev', 'trunk', 1, '-dev', 2), Channel('stable', 'branches/0.6', 2, '-stable', 1), ] CHANNELS_BY_NAME = {} for c in CHANNELS: CHANNELS_BY_NAME[c.name] = c <commit_msg>Update the build branch for stable to 0.7 TBR=ricow Review URL: https://codereview.chromium.org/26993005 git-svn-id: 239fca9b83025a0b6f823aeeca02ba5be3d9fd76@228644 0039d316-1c4b-4281-b951-d872f2087c98<commit_after># Copyright 2013 The Chromium Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. class Channel(object): def __init__(self, name, branch, position, category_postfix, priority): self.branch = branch self.builder_postfix = '-' + name self.category_postfix = category_postfix self.name = name self.position = position self.priority = priority self.all_deps_path = '/' + branch + '/deps/all.deps' self.dartium_deps_path = '/' + branch + '/deps/dartium.deps' # The channel names are replicated in the slave.cfg files for all # dart waterfalls. If you change anything here please also change it there. CHANNELS = [ Channel('be', 'branches/bleeding_edge', 0, '', 3), Channel('dev', 'trunk', 1, '-dev', 2), Channel('stable', 'branches/0.7', 2, '-stable', 1), ] CHANNELS_BY_NAME = {} for c in CHANNELS: CHANNELS_BY_NAME[c.name] = c
4840763265675407ed9fc612dc8884d859bdc28e
recipe_scrapers/_abstract.py
recipe_scrapers/_abstract.py
from urllib import request from bs4 import BeautifulSoup HEADERS = { 'User-Agent': 'Mozilla/5.0 (Windows; U; Windows NT 5.1; en-US; rv:1.9.0.7) Gecko/2009021910 Firefox/3.0.7' } class AbstractScraper(): def __init__(self, url, test=False): if test: # when testing, we simply load a file self.soup = BeautifulSoup(url.read(), "html.parser") else: self.soup = BeautifulSoup(request.urlopen( request.Request(url, headers=HEADERS)).read(), "html.parser") def host(self): """ get the host of the url, so we can use the correct scraper (check __init__.py) """ raise NotImplementedError("This should be implemented.") def publisher_site(self): """ the original url of the publisher site """ raise NotImplementedError("This should be implemented.") def title(self): """ title of the recipe itself """ raise NotImplementedError("This should be implemented.") def total_time(self): """ total time it takes to preparate the recipe in minutes """ raise NotImplementedError("This should be implemented.") def ingredients(self): """ list of ingredients needed for the recipe """ raise NotImplementedError("This should be implemented.") def instructions(self): """ directions provided on the recipe link """ raise NotImplementedError("This should be implemented.") def social_rating(self): """ social rating of the recipe in 0 - 100 scale """ raise NotImplementedError("This should be implemented.")
from urllib import request from bs4 import BeautifulSoup HEADERS = { 'User-Agent': 'Mozilla/5.0 (Windows; U; Windows NT 5.1; en-US; rv:1.9.0.7) Gecko/2009021910 Firefox/3.0.7' } class AbstractScraper(): def __init__(self, url, test=False): if test: # when testing, we simply load a file with url: self.soup = BeautifulSoup(url.read(), "html.parser") else: self.soup = BeautifulSoup(request.urlopen( request.Request(url, headers=HEADERS)).read(), "html.parser") def host(self): """ get the host of the url, so we can use the correct scraper (check __init__.py) """ raise NotImplementedError("This should be implemented.") def publisher_site(self): """ the original url of the publisher site """ raise NotImplementedError("This should be implemented.") def title(self): """ title of the recipe itself """ raise NotImplementedError("This should be implemented.") def total_time(self): """ total time it takes to preparate the recipe in minutes """ raise NotImplementedError("This should be implemented.") def ingredients(self): """ list of ingredients needed for the recipe """ raise NotImplementedError("This should be implemented.") def instructions(self): """ directions provided on the recipe link """ raise NotImplementedError("This should be implemented.") def social_rating(self): """ social rating of the recipe in 0 - 100 scale """ raise NotImplementedError("This should be implemented.")
Use context so the test file to get closed after tests
Use context so the test file to get closed after tests
Python
mit
hhursev/recipe-scraper
from urllib import request from bs4 import BeautifulSoup HEADERS = { 'User-Agent': 'Mozilla/5.0 (Windows; U; Windows NT 5.1; en-US; rv:1.9.0.7) Gecko/2009021910 Firefox/3.0.7' } class AbstractScraper(): def __init__(self, url, test=False): if test: # when testing, we simply load a file self.soup = BeautifulSoup(url.read(), "html.parser") else: self.soup = BeautifulSoup(request.urlopen( request.Request(url, headers=HEADERS)).read(), "html.parser") def host(self): """ get the host of the url, so we can use the correct scraper (check __init__.py) """ raise NotImplementedError("This should be implemented.") def publisher_site(self): """ the original url of the publisher site """ raise NotImplementedError("This should be implemented.") def title(self): """ title of the recipe itself """ raise NotImplementedError("This should be implemented.") def total_time(self): """ total time it takes to preparate the recipe in minutes """ raise NotImplementedError("This should be implemented.") def ingredients(self): """ list of ingredients needed for the recipe """ raise NotImplementedError("This should be implemented.") def instructions(self): """ directions provided on the recipe link """ raise NotImplementedError("This should be implemented.") def social_rating(self): """ social rating of the recipe in 0 - 100 scale """ raise NotImplementedError("This should be implemented.") Use context so the test file to get closed after tests
from urllib import request from bs4 import BeautifulSoup HEADERS = { 'User-Agent': 'Mozilla/5.0 (Windows; U; Windows NT 5.1; en-US; rv:1.9.0.7) Gecko/2009021910 Firefox/3.0.7' } class AbstractScraper(): def __init__(self, url, test=False): if test: # when testing, we simply load a file with url: self.soup = BeautifulSoup(url.read(), "html.parser") else: self.soup = BeautifulSoup(request.urlopen( request.Request(url, headers=HEADERS)).read(), "html.parser") def host(self): """ get the host of the url, so we can use the correct scraper (check __init__.py) """ raise NotImplementedError("This should be implemented.") def publisher_site(self): """ the original url of the publisher site """ raise NotImplementedError("This should be implemented.") def title(self): """ title of the recipe itself """ raise NotImplementedError("This should be implemented.") def total_time(self): """ total time it takes to preparate the recipe in minutes """ raise NotImplementedError("This should be implemented.") def ingredients(self): """ list of ingredients needed for the recipe """ raise NotImplementedError("This should be implemented.") def instructions(self): """ directions provided on the recipe link """ raise NotImplementedError("This should be implemented.") def social_rating(self): """ social rating of the recipe in 0 - 100 scale """ raise NotImplementedError("This should be implemented.")
<commit_before>from urllib import request from bs4 import BeautifulSoup HEADERS = { 'User-Agent': 'Mozilla/5.0 (Windows; U; Windows NT 5.1; en-US; rv:1.9.0.7) Gecko/2009021910 Firefox/3.0.7' } class AbstractScraper(): def __init__(self, url, test=False): if test: # when testing, we simply load a file self.soup = BeautifulSoup(url.read(), "html.parser") else: self.soup = BeautifulSoup(request.urlopen( request.Request(url, headers=HEADERS)).read(), "html.parser") def host(self): """ get the host of the url, so we can use the correct scraper (check __init__.py) """ raise NotImplementedError("This should be implemented.") def publisher_site(self): """ the original url of the publisher site """ raise NotImplementedError("This should be implemented.") def title(self): """ title of the recipe itself """ raise NotImplementedError("This should be implemented.") def total_time(self): """ total time it takes to preparate the recipe in minutes """ raise NotImplementedError("This should be implemented.") def ingredients(self): """ list of ingredients needed for the recipe """ raise NotImplementedError("This should be implemented.") def instructions(self): """ directions provided on the recipe link """ raise NotImplementedError("This should be implemented.") def social_rating(self): """ social rating of the recipe in 0 - 100 scale """ raise NotImplementedError("This should be implemented.") <commit_msg>Use context so the test file to get closed after tests<commit_after>
from urllib import request from bs4 import BeautifulSoup HEADERS = { 'User-Agent': 'Mozilla/5.0 (Windows; U; Windows NT 5.1; en-US; rv:1.9.0.7) Gecko/2009021910 Firefox/3.0.7' } class AbstractScraper(): def __init__(self, url, test=False): if test: # when testing, we simply load a file with url: self.soup = BeautifulSoup(url.read(), "html.parser") else: self.soup = BeautifulSoup(request.urlopen( request.Request(url, headers=HEADERS)).read(), "html.parser") def host(self): """ get the host of the url, so we can use the correct scraper (check __init__.py) """ raise NotImplementedError("This should be implemented.") def publisher_site(self): """ the original url of the publisher site """ raise NotImplementedError("This should be implemented.") def title(self): """ title of the recipe itself """ raise NotImplementedError("This should be implemented.") def total_time(self): """ total time it takes to preparate the recipe in minutes """ raise NotImplementedError("This should be implemented.") def ingredients(self): """ list of ingredients needed for the recipe """ raise NotImplementedError("This should be implemented.") def instructions(self): """ directions provided on the recipe link """ raise NotImplementedError("This should be implemented.") def social_rating(self): """ social rating of the recipe in 0 - 100 scale """ raise NotImplementedError("This should be implemented.")
from urllib import request from bs4 import BeautifulSoup HEADERS = { 'User-Agent': 'Mozilla/5.0 (Windows; U; Windows NT 5.1; en-US; rv:1.9.0.7) Gecko/2009021910 Firefox/3.0.7' } class AbstractScraper(): def __init__(self, url, test=False): if test: # when testing, we simply load a file self.soup = BeautifulSoup(url.read(), "html.parser") else: self.soup = BeautifulSoup(request.urlopen( request.Request(url, headers=HEADERS)).read(), "html.parser") def host(self): """ get the host of the url, so we can use the correct scraper (check __init__.py) """ raise NotImplementedError("This should be implemented.") def publisher_site(self): """ the original url of the publisher site """ raise NotImplementedError("This should be implemented.") def title(self): """ title of the recipe itself """ raise NotImplementedError("This should be implemented.") def total_time(self): """ total time it takes to preparate the recipe in minutes """ raise NotImplementedError("This should be implemented.") def ingredients(self): """ list of ingredients needed for the recipe """ raise NotImplementedError("This should be implemented.") def instructions(self): """ directions provided on the recipe link """ raise NotImplementedError("This should be implemented.") def social_rating(self): """ social rating of the recipe in 0 - 100 scale """ raise NotImplementedError("This should be implemented.") Use context so the test file to get closed after testsfrom urllib import request from bs4 import BeautifulSoup HEADERS = { 'User-Agent': 'Mozilla/5.0 (Windows; U; Windows NT 5.1; en-US; rv:1.9.0.7) Gecko/2009021910 Firefox/3.0.7' } class AbstractScraper(): def __init__(self, url, test=False): if test: # when testing, we simply load a file with url: self.soup = BeautifulSoup(url.read(), "html.parser") else: self.soup = BeautifulSoup(request.urlopen( request.Request(url, headers=HEADERS)).read(), "html.parser") def host(self): """ get the host of the url, so we can use the correct scraper (check __init__.py) """ raise NotImplementedError("This should be implemented.") def publisher_site(self): """ the original url of the publisher site """ raise NotImplementedError("This should be implemented.") def title(self): """ title of the recipe itself """ raise NotImplementedError("This should be implemented.") def total_time(self): """ total time it takes to preparate the recipe in minutes """ raise NotImplementedError("This should be implemented.") def ingredients(self): """ list of ingredients needed for the recipe """ raise NotImplementedError("This should be implemented.") def instructions(self): """ directions provided on the recipe link """ raise NotImplementedError("This should be implemented.") def social_rating(self): """ social rating of the recipe in 0 - 100 scale """ raise NotImplementedError("This should be implemented.")
<commit_before>from urllib import request from bs4 import BeautifulSoup HEADERS = { 'User-Agent': 'Mozilla/5.0 (Windows; U; Windows NT 5.1; en-US; rv:1.9.0.7) Gecko/2009021910 Firefox/3.0.7' } class AbstractScraper(): def __init__(self, url, test=False): if test: # when testing, we simply load a file self.soup = BeautifulSoup(url.read(), "html.parser") else: self.soup = BeautifulSoup(request.urlopen( request.Request(url, headers=HEADERS)).read(), "html.parser") def host(self): """ get the host of the url, so we can use the correct scraper (check __init__.py) """ raise NotImplementedError("This should be implemented.") def publisher_site(self): """ the original url of the publisher site """ raise NotImplementedError("This should be implemented.") def title(self): """ title of the recipe itself """ raise NotImplementedError("This should be implemented.") def total_time(self): """ total time it takes to preparate the recipe in minutes """ raise NotImplementedError("This should be implemented.") def ingredients(self): """ list of ingredients needed for the recipe """ raise NotImplementedError("This should be implemented.") def instructions(self): """ directions provided on the recipe link """ raise NotImplementedError("This should be implemented.") def social_rating(self): """ social rating of the recipe in 0 - 100 scale """ raise NotImplementedError("This should be implemented.") <commit_msg>Use context so the test file to get closed after tests<commit_after>from urllib import request from bs4 import BeautifulSoup HEADERS = { 'User-Agent': 'Mozilla/5.0 (Windows; U; Windows NT 5.1; en-US; rv:1.9.0.7) Gecko/2009021910 Firefox/3.0.7' } class AbstractScraper(): def __init__(self, url, test=False): if test: # when testing, we simply load a file with url: self.soup = BeautifulSoup(url.read(), "html.parser") else: self.soup = BeautifulSoup(request.urlopen( request.Request(url, headers=HEADERS)).read(), "html.parser") def host(self): """ get the host of the url, so we can use the correct scraper (check __init__.py) """ raise NotImplementedError("This should be implemented.") def publisher_site(self): """ the original url of the publisher site """ raise NotImplementedError("This should be implemented.") def title(self): """ title of the recipe itself """ raise NotImplementedError("This should be implemented.") def total_time(self): """ total time it takes to preparate the recipe in minutes """ raise NotImplementedError("This should be implemented.") def ingredients(self): """ list of ingredients needed for the recipe """ raise NotImplementedError("This should be implemented.") def instructions(self): """ directions provided on the recipe link """ raise NotImplementedError("This should be implemented.") def social_rating(self): """ social rating of the recipe in 0 - 100 scale """ raise NotImplementedError("This should be implemented.")
e59055e29b5cc6a027d3a24803cc05fd709cca90
functest/opnfv_tests/features/odl_sfc.py
functest/opnfv_tests/features/odl_sfc.py
#!/usr/bin/python # # Copyright (c) 2016 All rights reserved # This program and the accompanying materials # are made available under the terms of the Apache License, Version 2.0 # which accompanies this distribution, and is available at # # http://www.apache.org/licenses/LICENSE-2.0 # import functest.core.feature_base as base from sfc.tests.functest import run_tests class OpenDaylightSFC(base.FeatureBase): def __init__(self): super(OpenDaylightSFC, self).__init__(project='sfc', case='functest-odl-sfc', repo='dir_repo_sfc') def execute(self): return run_tests.main()
#!/usr/bin/python # # Copyright (c) 2016 All rights reserved # This program and the accompanying materials # are made available under the terms of the Apache License, Version 2.0 # which accompanies this distribution, and is available at # # http://www.apache.org/licenses/LICENSE-2.0 # import functest.core.feature_base as base class OpenDaylightSFC(base.FeatureBase): def __init__(self): super(OpenDaylightSFC, self).__init__(project='sfc', case='functest-odl-sfc', repo='dir_repo_sfc') dir_sfc_functest = '{}/sfc/tests/functest'.format(self.repo) self.cmd = 'cd %s && python ./run_tests.py' % dir_sfc_functest
Revert "Make SFC test a python call to main()"
Revert "Make SFC test a python call to main()" This reverts commit d5820bef80ea4bdb871380dbfe41db12290fc5f8. Robot test runs before SFC test and it imports https://github.com/robotframework/SSHLibrary which does a monkey patching in the python runtime / paramiko. Untill now sfc run in a new python process (clean) because it run using the bash command. But when importing it as a module and call main() from python, it will run in the patched runtime and it will error out. https://hastebin.com/iyobuxutib.py Change-Id: I54237c32c957718b363d302efe84e01bc78e4f47 Signed-off-by: George Paraskevopoulos <97fa9f4b61869a8ee92863df90135c882f7b3aa0@intracom-telecom.com>
Python
apache-2.0
opnfv/functest,mywulin/functest,opnfv/functest,mywulin/functest
#!/usr/bin/python # # Copyright (c) 2016 All rights reserved # This program and the accompanying materials # are made available under the terms of the Apache License, Version 2.0 # which accompanies this distribution, and is available at # # http://www.apache.org/licenses/LICENSE-2.0 # import functest.core.feature_base as base from sfc.tests.functest import run_tests class OpenDaylightSFC(base.FeatureBase): def __init__(self): super(OpenDaylightSFC, self).__init__(project='sfc', case='functest-odl-sfc', repo='dir_repo_sfc') def execute(self): return run_tests.main() Revert "Make SFC test a python call to main()" This reverts commit d5820bef80ea4bdb871380dbfe41db12290fc5f8. Robot test runs before SFC test and it imports https://github.com/robotframework/SSHLibrary which does a monkey patching in the python runtime / paramiko. Untill now sfc run in a new python process (clean) because it run using the bash command. But when importing it as a module and call main() from python, it will run in the patched runtime and it will error out. https://hastebin.com/iyobuxutib.py Change-Id: I54237c32c957718b363d302efe84e01bc78e4f47 Signed-off-by: George Paraskevopoulos <97fa9f4b61869a8ee92863df90135c882f7b3aa0@intracom-telecom.com>
#!/usr/bin/python # # Copyright (c) 2016 All rights reserved # This program and the accompanying materials # are made available under the terms of the Apache License, Version 2.0 # which accompanies this distribution, and is available at # # http://www.apache.org/licenses/LICENSE-2.0 # import functest.core.feature_base as base class OpenDaylightSFC(base.FeatureBase): def __init__(self): super(OpenDaylightSFC, self).__init__(project='sfc', case='functest-odl-sfc', repo='dir_repo_sfc') dir_sfc_functest = '{}/sfc/tests/functest'.format(self.repo) self.cmd = 'cd %s && python ./run_tests.py' % dir_sfc_functest
<commit_before>#!/usr/bin/python # # Copyright (c) 2016 All rights reserved # This program and the accompanying materials # are made available under the terms of the Apache License, Version 2.0 # which accompanies this distribution, and is available at # # http://www.apache.org/licenses/LICENSE-2.0 # import functest.core.feature_base as base from sfc.tests.functest import run_tests class OpenDaylightSFC(base.FeatureBase): def __init__(self): super(OpenDaylightSFC, self).__init__(project='sfc', case='functest-odl-sfc', repo='dir_repo_sfc') def execute(self): return run_tests.main() <commit_msg>Revert "Make SFC test a python call to main()" This reverts commit d5820bef80ea4bdb871380dbfe41db12290fc5f8. Robot test runs before SFC test and it imports https://github.com/robotframework/SSHLibrary which does a monkey patching in the python runtime / paramiko. Untill now sfc run in a new python process (clean) because it run using the bash command. But when importing it as a module and call main() from python, it will run in the patched runtime and it will error out. https://hastebin.com/iyobuxutib.py Change-Id: I54237c32c957718b363d302efe84e01bc78e4f47 Signed-off-by: George Paraskevopoulos <97fa9f4b61869a8ee92863df90135c882f7b3aa0@intracom-telecom.com><commit_after>
#!/usr/bin/python # # Copyright (c) 2016 All rights reserved # This program and the accompanying materials # are made available under the terms of the Apache License, Version 2.0 # which accompanies this distribution, and is available at # # http://www.apache.org/licenses/LICENSE-2.0 # import functest.core.feature_base as base class OpenDaylightSFC(base.FeatureBase): def __init__(self): super(OpenDaylightSFC, self).__init__(project='sfc', case='functest-odl-sfc', repo='dir_repo_sfc') dir_sfc_functest = '{}/sfc/tests/functest'.format(self.repo) self.cmd = 'cd %s && python ./run_tests.py' % dir_sfc_functest
#!/usr/bin/python # # Copyright (c) 2016 All rights reserved # This program and the accompanying materials # are made available under the terms of the Apache License, Version 2.0 # which accompanies this distribution, and is available at # # http://www.apache.org/licenses/LICENSE-2.0 # import functest.core.feature_base as base from sfc.tests.functest import run_tests class OpenDaylightSFC(base.FeatureBase): def __init__(self): super(OpenDaylightSFC, self).__init__(project='sfc', case='functest-odl-sfc', repo='dir_repo_sfc') def execute(self): return run_tests.main() Revert "Make SFC test a python call to main()" This reverts commit d5820bef80ea4bdb871380dbfe41db12290fc5f8. Robot test runs before SFC test and it imports https://github.com/robotframework/SSHLibrary which does a monkey patching in the python runtime / paramiko. Untill now sfc run in a new python process (clean) because it run using the bash command. But when importing it as a module and call main() from python, it will run in the patched runtime and it will error out. https://hastebin.com/iyobuxutib.py Change-Id: I54237c32c957718b363d302efe84e01bc78e4f47 Signed-off-by: George Paraskevopoulos <97fa9f4b61869a8ee92863df90135c882f7b3aa0@intracom-telecom.com>#!/usr/bin/python # # Copyright (c) 2016 All rights reserved # This program and the accompanying materials # are made available under the terms of the Apache License, Version 2.0 # which accompanies this distribution, and is available at # # http://www.apache.org/licenses/LICENSE-2.0 # import functest.core.feature_base as base class OpenDaylightSFC(base.FeatureBase): def __init__(self): super(OpenDaylightSFC, self).__init__(project='sfc', case='functest-odl-sfc', repo='dir_repo_sfc') dir_sfc_functest = '{}/sfc/tests/functest'.format(self.repo) self.cmd = 'cd %s && python ./run_tests.py' % dir_sfc_functest
<commit_before>#!/usr/bin/python # # Copyright (c) 2016 All rights reserved # This program and the accompanying materials # are made available under the terms of the Apache License, Version 2.0 # which accompanies this distribution, and is available at # # http://www.apache.org/licenses/LICENSE-2.0 # import functest.core.feature_base as base from sfc.tests.functest import run_tests class OpenDaylightSFC(base.FeatureBase): def __init__(self): super(OpenDaylightSFC, self).__init__(project='sfc', case='functest-odl-sfc', repo='dir_repo_sfc') def execute(self): return run_tests.main() <commit_msg>Revert "Make SFC test a python call to main()" This reverts commit d5820bef80ea4bdb871380dbfe41db12290fc5f8. Robot test runs before SFC test and it imports https://github.com/robotframework/SSHLibrary which does a monkey patching in the python runtime / paramiko. Untill now sfc run in a new python process (clean) because it run using the bash command. But when importing it as a module and call main() from python, it will run in the patched runtime and it will error out. https://hastebin.com/iyobuxutib.py Change-Id: I54237c32c957718b363d302efe84e01bc78e4f47 Signed-off-by: George Paraskevopoulos <97fa9f4b61869a8ee92863df90135c882f7b3aa0@intracom-telecom.com><commit_after>#!/usr/bin/python # # Copyright (c) 2016 All rights reserved # This program and the accompanying materials # are made available under the terms of the Apache License, Version 2.0 # which accompanies this distribution, and is available at # # http://www.apache.org/licenses/LICENSE-2.0 # import functest.core.feature_base as base class OpenDaylightSFC(base.FeatureBase): def __init__(self): super(OpenDaylightSFC, self).__init__(project='sfc', case='functest-odl-sfc', repo='dir_repo_sfc') dir_sfc_functest = '{}/sfc/tests/functest'.format(self.repo) self.cmd = 'cd %s && python ./run_tests.py' % dir_sfc_functest
f3702870cf912322061678cf7c827959cde2ac02
south/introspection_plugins/__init__.py
south/introspection_plugins/__init__.py
# This module contains built-in introspector plugins for various common # Django apps. # These imports trigger the lower-down files import south.introspection_plugins.geodjango import south.introspection_plugins.django_tagging import south.introspection_plugins.django_taggit import south.introspection_plugins.django_objectpermissions
# This module contains built-in introspector plugins for various common # Django apps. # These imports trigger the lower-down files import south.introspection_plugins.geodjango import south.introspection_plugins.django_tagging import south.introspection_plugins.django_taggit import south.introspection_plugins.django_objectpermissions import south.introspection_plugins.annoying_autoonetoone
Add import of django-annoying patch
Add import of django-annoying patch
Python
apache-2.0
matthiask/south,matthiask/south
# This module contains built-in introspector plugins for various common # Django apps. # These imports trigger the lower-down files import south.introspection_plugins.geodjango import south.introspection_plugins.django_tagging import south.introspection_plugins.django_taggit import south.introspection_plugins.django_objectpermissions Add import of django-annoying patch
# This module contains built-in introspector plugins for various common # Django apps. # These imports trigger the lower-down files import south.introspection_plugins.geodjango import south.introspection_plugins.django_tagging import south.introspection_plugins.django_taggit import south.introspection_plugins.django_objectpermissions import south.introspection_plugins.annoying_autoonetoone
<commit_before># This module contains built-in introspector plugins for various common # Django apps. # These imports trigger the lower-down files import south.introspection_plugins.geodjango import south.introspection_plugins.django_tagging import south.introspection_plugins.django_taggit import south.introspection_plugins.django_objectpermissions <commit_msg>Add import of django-annoying patch<commit_after>
# This module contains built-in introspector plugins for various common # Django apps. # These imports trigger the lower-down files import south.introspection_plugins.geodjango import south.introspection_plugins.django_tagging import south.introspection_plugins.django_taggit import south.introspection_plugins.django_objectpermissions import south.introspection_plugins.annoying_autoonetoone
# This module contains built-in introspector plugins for various common # Django apps. # These imports trigger the lower-down files import south.introspection_plugins.geodjango import south.introspection_plugins.django_tagging import south.introspection_plugins.django_taggit import south.introspection_plugins.django_objectpermissions Add import of django-annoying patch# This module contains built-in introspector plugins for various common # Django apps. # These imports trigger the lower-down files import south.introspection_plugins.geodjango import south.introspection_plugins.django_tagging import south.introspection_plugins.django_taggit import south.introspection_plugins.django_objectpermissions import south.introspection_plugins.annoying_autoonetoone
<commit_before># This module contains built-in introspector plugins for various common # Django apps. # These imports trigger the lower-down files import south.introspection_plugins.geodjango import south.introspection_plugins.django_tagging import south.introspection_plugins.django_taggit import south.introspection_plugins.django_objectpermissions <commit_msg>Add import of django-annoying patch<commit_after># This module contains built-in introspector plugins for various common # Django apps. # These imports trigger the lower-down files import south.introspection_plugins.geodjango import south.introspection_plugins.django_tagging import south.introspection_plugins.django_taggit import south.introspection_plugins.django_objectpermissions import south.introspection_plugins.annoying_autoonetoone
47faf3ad38eff5de2edc529a7347c01ddeddf4c4
talks/users/models.py
talks/users/models.py
from django.db import models from django.contrib.auth.models import User from django.contrib.contenttypes.fields import GenericForeignKey from django.contrib.contenttypes.models import ContentType class TalksUser(models.Model): user = models.OneToOneField(User) class CollectionFollow(models.Model): """User following a Collection """ user = models.ForeignKey(TalksUser) collection = models.ForeignKey(Collection) is_owner = models.BooleanField() is_main = models.BooleanField() # main collection of the user class Meta: unique_together = ('user', 'collection') class Collection(models.Model): slug = models.SlugField() title = models.CharField(max_length=250) # TODO list private or public/shared? # TODO qualify list? (e.g. "Talks I want to attend"?) class CollectionItem(models.Model): collection = models.ForeignKey(Collection) content_type = models.ForeignKey(ContentType) object_id = models.PositiveIntegerField() item = GenericForeignKey('content_type', 'object_id') # atm: Talk, Event, Series class DepartmentFollow(models.Model): pass class LocationFollow(models.Model): pass
from django.db import models from django.dispatch import receiver from django.contrib.auth.models import User from django.contrib.contenttypes.fields import GenericForeignKey from django.contrib.contenttypes.models import ContentType class TalksUser(models.Model): user = models.OneToOneField(User) @receiver(models.signals.post_save, sender=User) def ensure_profile_exists(sender, instance, created, **kwargs): """If the User has just been created we use a signal to also create a TalksUser """ if created: TalksUser.objects.get_or_create(user=instance) class Collection(models.Model): slug = models.SlugField() title = models.CharField(max_length=250) # TODO list private or public/shared? # TODO qualify list? (e.g. "Talks I want to attend"?) class CollectionFollow(models.Model): """User following a Collection """ user = models.ForeignKey(TalksUser) collection = models.ForeignKey(Collection) is_owner = models.BooleanField(default=False) # Main/default collection of the user is_main = models.BooleanField(default=False) class Meta: unique_together = ('user', 'collection') class CollectionItem(models.Model): collection = models.ForeignKey(Collection) content_type = models.ForeignKey(ContentType) object_id = models.PositiveIntegerField() # Currently item can be an Event or EventGroup item = GenericForeignKey('content_type', 'object_id') class DepartmentFollow(models.Model): pass class LocationFollow(models.Model): pass
Create TalksUser model when a user is created
Create TalksUser model when a user is created
Python
apache-2.0
ox-it/talks.ox,ox-it/talks.ox,ox-it/talks.ox
from django.db import models from django.contrib.auth.models import User from django.contrib.contenttypes.fields import GenericForeignKey from django.contrib.contenttypes.models import ContentType class TalksUser(models.Model): user = models.OneToOneField(User) class CollectionFollow(models.Model): """User following a Collection """ user = models.ForeignKey(TalksUser) collection = models.ForeignKey(Collection) is_owner = models.BooleanField() is_main = models.BooleanField() # main collection of the user class Meta: unique_together = ('user', 'collection') class Collection(models.Model): slug = models.SlugField() title = models.CharField(max_length=250) # TODO list private or public/shared? # TODO qualify list? (e.g. "Talks I want to attend"?) class CollectionItem(models.Model): collection = models.ForeignKey(Collection) content_type = models.ForeignKey(ContentType) object_id = models.PositiveIntegerField() item = GenericForeignKey('content_type', 'object_id') # atm: Talk, Event, Series class DepartmentFollow(models.Model): pass class LocationFollow(models.Model): pass Create TalksUser model when a user is created
from django.db import models from django.dispatch import receiver from django.contrib.auth.models import User from django.contrib.contenttypes.fields import GenericForeignKey from django.contrib.contenttypes.models import ContentType class TalksUser(models.Model): user = models.OneToOneField(User) @receiver(models.signals.post_save, sender=User) def ensure_profile_exists(sender, instance, created, **kwargs): """If the User has just been created we use a signal to also create a TalksUser """ if created: TalksUser.objects.get_or_create(user=instance) class Collection(models.Model): slug = models.SlugField() title = models.CharField(max_length=250) # TODO list private or public/shared? # TODO qualify list? (e.g. "Talks I want to attend"?) class CollectionFollow(models.Model): """User following a Collection """ user = models.ForeignKey(TalksUser) collection = models.ForeignKey(Collection) is_owner = models.BooleanField(default=False) # Main/default collection of the user is_main = models.BooleanField(default=False) class Meta: unique_together = ('user', 'collection') class CollectionItem(models.Model): collection = models.ForeignKey(Collection) content_type = models.ForeignKey(ContentType) object_id = models.PositiveIntegerField() # Currently item can be an Event or EventGroup item = GenericForeignKey('content_type', 'object_id') class DepartmentFollow(models.Model): pass class LocationFollow(models.Model): pass
<commit_before>from django.db import models from django.contrib.auth.models import User from django.contrib.contenttypes.fields import GenericForeignKey from django.contrib.contenttypes.models import ContentType class TalksUser(models.Model): user = models.OneToOneField(User) class CollectionFollow(models.Model): """User following a Collection """ user = models.ForeignKey(TalksUser) collection = models.ForeignKey(Collection) is_owner = models.BooleanField() is_main = models.BooleanField() # main collection of the user class Meta: unique_together = ('user', 'collection') class Collection(models.Model): slug = models.SlugField() title = models.CharField(max_length=250) # TODO list private or public/shared? # TODO qualify list? (e.g. "Talks I want to attend"?) class CollectionItem(models.Model): collection = models.ForeignKey(Collection) content_type = models.ForeignKey(ContentType) object_id = models.PositiveIntegerField() item = GenericForeignKey('content_type', 'object_id') # atm: Talk, Event, Series class DepartmentFollow(models.Model): pass class LocationFollow(models.Model): pass <commit_msg>Create TalksUser model when a user is created<commit_after>
from django.db import models from django.dispatch import receiver from django.contrib.auth.models import User from django.contrib.contenttypes.fields import GenericForeignKey from django.contrib.contenttypes.models import ContentType class TalksUser(models.Model): user = models.OneToOneField(User) @receiver(models.signals.post_save, sender=User) def ensure_profile_exists(sender, instance, created, **kwargs): """If the User has just been created we use a signal to also create a TalksUser """ if created: TalksUser.objects.get_or_create(user=instance) class Collection(models.Model): slug = models.SlugField() title = models.CharField(max_length=250) # TODO list private or public/shared? # TODO qualify list? (e.g. "Talks I want to attend"?) class CollectionFollow(models.Model): """User following a Collection """ user = models.ForeignKey(TalksUser) collection = models.ForeignKey(Collection) is_owner = models.BooleanField(default=False) # Main/default collection of the user is_main = models.BooleanField(default=False) class Meta: unique_together = ('user', 'collection') class CollectionItem(models.Model): collection = models.ForeignKey(Collection) content_type = models.ForeignKey(ContentType) object_id = models.PositiveIntegerField() # Currently item can be an Event or EventGroup item = GenericForeignKey('content_type', 'object_id') class DepartmentFollow(models.Model): pass class LocationFollow(models.Model): pass
from django.db import models from django.contrib.auth.models import User from django.contrib.contenttypes.fields import GenericForeignKey from django.contrib.contenttypes.models import ContentType class TalksUser(models.Model): user = models.OneToOneField(User) class CollectionFollow(models.Model): """User following a Collection """ user = models.ForeignKey(TalksUser) collection = models.ForeignKey(Collection) is_owner = models.BooleanField() is_main = models.BooleanField() # main collection of the user class Meta: unique_together = ('user', 'collection') class Collection(models.Model): slug = models.SlugField() title = models.CharField(max_length=250) # TODO list private or public/shared? # TODO qualify list? (e.g. "Talks I want to attend"?) class CollectionItem(models.Model): collection = models.ForeignKey(Collection) content_type = models.ForeignKey(ContentType) object_id = models.PositiveIntegerField() item = GenericForeignKey('content_type', 'object_id') # atm: Talk, Event, Series class DepartmentFollow(models.Model): pass class LocationFollow(models.Model): pass Create TalksUser model when a user is createdfrom django.db import models from django.dispatch import receiver from django.contrib.auth.models import User from django.contrib.contenttypes.fields import GenericForeignKey from django.contrib.contenttypes.models import ContentType class TalksUser(models.Model): user = models.OneToOneField(User) @receiver(models.signals.post_save, sender=User) def ensure_profile_exists(sender, instance, created, **kwargs): """If the User has just been created we use a signal to also create a TalksUser """ if created: TalksUser.objects.get_or_create(user=instance) class Collection(models.Model): slug = models.SlugField() title = models.CharField(max_length=250) # TODO list private or public/shared? # TODO qualify list? (e.g. "Talks I want to attend"?) class CollectionFollow(models.Model): """User following a Collection """ user = models.ForeignKey(TalksUser) collection = models.ForeignKey(Collection) is_owner = models.BooleanField(default=False) # Main/default collection of the user is_main = models.BooleanField(default=False) class Meta: unique_together = ('user', 'collection') class CollectionItem(models.Model): collection = models.ForeignKey(Collection) content_type = models.ForeignKey(ContentType) object_id = models.PositiveIntegerField() # Currently item can be an Event or EventGroup item = GenericForeignKey('content_type', 'object_id') class DepartmentFollow(models.Model): pass class LocationFollow(models.Model): pass
<commit_before>from django.db import models from django.contrib.auth.models import User from django.contrib.contenttypes.fields import GenericForeignKey from django.contrib.contenttypes.models import ContentType class TalksUser(models.Model): user = models.OneToOneField(User) class CollectionFollow(models.Model): """User following a Collection """ user = models.ForeignKey(TalksUser) collection = models.ForeignKey(Collection) is_owner = models.BooleanField() is_main = models.BooleanField() # main collection of the user class Meta: unique_together = ('user', 'collection') class Collection(models.Model): slug = models.SlugField() title = models.CharField(max_length=250) # TODO list private or public/shared? # TODO qualify list? (e.g. "Talks I want to attend"?) class CollectionItem(models.Model): collection = models.ForeignKey(Collection) content_type = models.ForeignKey(ContentType) object_id = models.PositiveIntegerField() item = GenericForeignKey('content_type', 'object_id') # atm: Talk, Event, Series class DepartmentFollow(models.Model): pass class LocationFollow(models.Model): pass <commit_msg>Create TalksUser model when a user is created<commit_after>from django.db import models from django.dispatch import receiver from django.contrib.auth.models import User from django.contrib.contenttypes.fields import GenericForeignKey from django.contrib.contenttypes.models import ContentType class TalksUser(models.Model): user = models.OneToOneField(User) @receiver(models.signals.post_save, sender=User) def ensure_profile_exists(sender, instance, created, **kwargs): """If the User has just been created we use a signal to also create a TalksUser """ if created: TalksUser.objects.get_or_create(user=instance) class Collection(models.Model): slug = models.SlugField() title = models.CharField(max_length=250) # TODO list private or public/shared? # TODO qualify list? (e.g. "Talks I want to attend"?) class CollectionFollow(models.Model): """User following a Collection """ user = models.ForeignKey(TalksUser) collection = models.ForeignKey(Collection) is_owner = models.BooleanField(default=False) # Main/default collection of the user is_main = models.BooleanField(default=False) class Meta: unique_together = ('user', 'collection') class CollectionItem(models.Model): collection = models.ForeignKey(Collection) content_type = models.ForeignKey(ContentType) object_id = models.PositiveIntegerField() # Currently item can be an Event or EventGroup item = GenericForeignKey('content_type', 'object_id') class DepartmentFollow(models.Model): pass class LocationFollow(models.Model): pass
f39982f0cab45ba898bf8de7a4170e615cd9569c
aldryn_newsblog/managers.py
aldryn_newsblog/managers.py
from collections import Counter import datetime from parler.managers import TranslatableManager class RelatedManager(TranslatableManager): # TODO: uncomment this when we'll have image field # def get_query_set(self): # qs = super(RelatedManager, self).get_query_set() # return qs.select_related('image_fk') def get_months(self, namespace): """ Get months with posts count for given namespace string. This means how much posts there are in each month. Results are ordered by date. """ # done in a naive way as Django is having tough time while aggregating # on date fields entries = self.filter(namespace__namespace=namespace) dates = entries.values_list('publishing_date', flat=True) dates = [(x.year, x.month) for x in dates] date_counter = Counter(dates) dates = set(dates) dates = sorted(dates, reverse=True) months = [ # Use day=3 to make sure timezone won't affect this hacks' # month value. There are UTC+14 and UTC-12 timezones. {'date': datetime.date(year=year, month=month, day=3), 'count': date_counter[year, month]} for year, month in dates] return months
try: from collections import Counter except ImportError: from backport_collections import Counter import datetime from parler.managers import TranslatableManager class RelatedManager(TranslatableManager): # TODO: uncomment this when we'll have image field # def get_query_set(self): # qs = super(RelatedManager, self).get_query_set() # return qs.select_related('image_fk') def get_months(self, namespace): """ Get months with posts count for given namespace string. This means how much posts there are in each month. Results are ordered by date. """ # done in a naive way as Django is having tough time while aggregating # on date fields entries = self.filter(namespace__namespace=namespace) dates = entries.values_list('publishing_date', flat=True) dates = [(x.year, x.month) for x in dates] date_counter = Counter(dates) dates = set(dates) dates = sorted(dates, reverse=True) months = [ # Use day=3 to make sure timezone won't affect this hacks' # month value. There are UTC+14 and UTC-12 timezones. {'date': datetime.date(year=year, month=month, day=3), 'count': date_counter[year, month]} for year, month in dates] return months
Add fallback: from backport_collections import Counter
Add fallback: from backport_collections import Counter
Python
bsd-3-clause
mkoistinen/aldryn-newsblog,mkoistinen/aldryn-newsblog,mkoistinen/aldryn-newsblog,czpython/aldryn-newsblog,czpython/aldryn-newsblog,czpython/aldryn-newsblog,czpython/aldryn-newsblog
from collections import Counter import datetime from parler.managers import TranslatableManager class RelatedManager(TranslatableManager): # TODO: uncomment this when we'll have image field # def get_query_set(self): # qs = super(RelatedManager, self).get_query_set() # return qs.select_related('image_fk') def get_months(self, namespace): """ Get months with posts count for given namespace string. This means how much posts there are in each month. Results are ordered by date. """ # done in a naive way as Django is having tough time while aggregating # on date fields entries = self.filter(namespace__namespace=namespace) dates = entries.values_list('publishing_date', flat=True) dates = [(x.year, x.month) for x in dates] date_counter = Counter(dates) dates = set(dates) dates = sorted(dates, reverse=True) months = [ # Use day=3 to make sure timezone won't affect this hacks' # month value. There are UTC+14 and UTC-12 timezones. {'date': datetime.date(year=year, month=month, day=3), 'count': date_counter[year, month]} for year, month in dates] return months Add fallback: from backport_collections import Counter
try: from collections import Counter except ImportError: from backport_collections import Counter import datetime from parler.managers import TranslatableManager class RelatedManager(TranslatableManager): # TODO: uncomment this when we'll have image field # def get_query_set(self): # qs = super(RelatedManager, self).get_query_set() # return qs.select_related('image_fk') def get_months(self, namespace): """ Get months with posts count for given namespace string. This means how much posts there are in each month. Results are ordered by date. """ # done in a naive way as Django is having tough time while aggregating # on date fields entries = self.filter(namespace__namespace=namespace) dates = entries.values_list('publishing_date', flat=True) dates = [(x.year, x.month) for x in dates] date_counter = Counter(dates) dates = set(dates) dates = sorted(dates, reverse=True) months = [ # Use day=3 to make sure timezone won't affect this hacks' # month value. There are UTC+14 and UTC-12 timezones. {'date': datetime.date(year=year, month=month, day=3), 'count': date_counter[year, month]} for year, month in dates] return months
<commit_before>from collections import Counter import datetime from parler.managers import TranslatableManager class RelatedManager(TranslatableManager): # TODO: uncomment this when we'll have image field # def get_query_set(self): # qs = super(RelatedManager, self).get_query_set() # return qs.select_related('image_fk') def get_months(self, namespace): """ Get months with posts count for given namespace string. This means how much posts there are in each month. Results are ordered by date. """ # done in a naive way as Django is having tough time while aggregating # on date fields entries = self.filter(namespace__namespace=namespace) dates = entries.values_list('publishing_date', flat=True) dates = [(x.year, x.month) for x in dates] date_counter = Counter(dates) dates = set(dates) dates = sorted(dates, reverse=True) months = [ # Use day=3 to make sure timezone won't affect this hacks' # month value. There are UTC+14 and UTC-12 timezones. {'date': datetime.date(year=year, month=month, day=3), 'count': date_counter[year, month]} for year, month in dates] return months <commit_msg>Add fallback: from backport_collections import Counter<commit_after>
try: from collections import Counter except ImportError: from backport_collections import Counter import datetime from parler.managers import TranslatableManager class RelatedManager(TranslatableManager): # TODO: uncomment this when we'll have image field # def get_query_set(self): # qs = super(RelatedManager, self).get_query_set() # return qs.select_related('image_fk') def get_months(self, namespace): """ Get months with posts count for given namespace string. This means how much posts there are in each month. Results are ordered by date. """ # done in a naive way as Django is having tough time while aggregating # on date fields entries = self.filter(namespace__namespace=namespace) dates = entries.values_list('publishing_date', flat=True) dates = [(x.year, x.month) for x in dates] date_counter = Counter(dates) dates = set(dates) dates = sorted(dates, reverse=True) months = [ # Use day=3 to make sure timezone won't affect this hacks' # month value. There are UTC+14 and UTC-12 timezones. {'date': datetime.date(year=year, month=month, day=3), 'count': date_counter[year, month]} for year, month in dates] return months
from collections import Counter import datetime from parler.managers import TranslatableManager class RelatedManager(TranslatableManager): # TODO: uncomment this when we'll have image field # def get_query_set(self): # qs = super(RelatedManager, self).get_query_set() # return qs.select_related('image_fk') def get_months(self, namespace): """ Get months with posts count for given namespace string. This means how much posts there are in each month. Results are ordered by date. """ # done in a naive way as Django is having tough time while aggregating # on date fields entries = self.filter(namespace__namespace=namespace) dates = entries.values_list('publishing_date', flat=True) dates = [(x.year, x.month) for x in dates] date_counter = Counter(dates) dates = set(dates) dates = sorted(dates, reverse=True) months = [ # Use day=3 to make sure timezone won't affect this hacks' # month value. There are UTC+14 and UTC-12 timezones. {'date': datetime.date(year=year, month=month, day=3), 'count': date_counter[year, month]} for year, month in dates] return months Add fallback: from backport_collections import Countertry: from collections import Counter except ImportError: from backport_collections import Counter import datetime from parler.managers import TranslatableManager class RelatedManager(TranslatableManager): # TODO: uncomment this when we'll have image field # def get_query_set(self): # qs = super(RelatedManager, self).get_query_set() # return qs.select_related('image_fk') def get_months(self, namespace): """ Get months with posts count for given namespace string. This means how much posts there are in each month. Results are ordered by date. """ # done in a naive way as Django is having tough time while aggregating # on date fields entries = self.filter(namespace__namespace=namespace) dates = entries.values_list('publishing_date', flat=True) dates = [(x.year, x.month) for x in dates] date_counter = Counter(dates) dates = set(dates) dates = sorted(dates, reverse=True) months = [ # Use day=3 to make sure timezone won't affect this hacks' # month value. There are UTC+14 and UTC-12 timezones. {'date': datetime.date(year=year, month=month, day=3), 'count': date_counter[year, month]} for year, month in dates] return months
<commit_before>from collections import Counter import datetime from parler.managers import TranslatableManager class RelatedManager(TranslatableManager): # TODO: uncomment this when we'll have image field # def get_query_set(self): # qs = super(RelatedManager, self).get_query_set() # return qs.select_related('image_fk') def get_months(self, namespace): """ Get months with posts count for given namespace string. This means how much posts there are in each month. Results are ordered by date. """ # done in a naive way as Django is having tough time while aggregating # on date fields entries = self.filter(namespace__namespace=namespace) dates = entries.values_list('publishing_date', flat=True) dates = [(x.year, x.month) for x in dates] date_counter = Counter(dates) dates = set(dates) dates = sorted(dates, reverse=True) months = [ # Use day=3 to make sure timezone won't affect this hacks' # month value. There are UTC+14 and UTC-12 timezones. {'date': datetime.date(year=year, month=month, day=3), 'count': date_counter[year, month]} for year, month in dates] return months <commit_msg>Add fallback: from backport_collections import Counter<commit_after>try: from collections import Counter except ImportError: from backport_collections import Counter import datetime from parler.managers import TranslatableManager class RelatedManager(TranslatableManager): # TODO: uncomment this when we'll have image field # def get_query_set(self): # qs = super(RelatedManager, self).get_query_set() # return qs.select_related('image_fk') def get_months(self, namespace): """ Get months with posts count for given namespace string. This means how much posts there are in each month. Results are ordered by date. """ # done in a naive way as Django is having tough time while aggregating # on date fields entries = self.filter(namespace__namespace=namespace) dates = entries.values_list('publishing_date', flat=True) dates = [(x.year, x.month) for x in dates] date_counter = Counter(dates) dates = set(dates) dates = sorted(dates, reverse=True) months = [ # Use day=3 to make sure timezone won't affect this hacks' # month value. There are UTC+14 and UTC-12 timezones. {'date': datetime.date(year=year, month=month, day=3), 'count': date_counter[year, month]} for year, month in dates] return months
a268a886e0e3cab1057810f488feb6c2227414d3
users/serializers.py
users/serializers.py
from django.conf import settings from rest_framework import serializers from rest_framework.exceptions import ValidationError from users.models import User class UserSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = User fields = ( 'username', 'email', 'gravatar', 'password', 'password_repeat', settings.DRF_URL_FIELD_NAME, ) extra_kwargs = { settings.DRF_URL_FIELD_NAME: { "view_name": "users:user-detail", }, } password = serializers.CharField( write_only=True, required=True, allow_blank=False, min_length=6, max_length=32, ) password_repeat = serializers.CharField( write_only=True, required=True, allow_blank=False, min_length=6, max_length=32, ) def create(self, validated_data): if validated_data['password'] != validated_data['password']: raise ValidationError( detail={ "password_repeat": "Tow password doesn't match", } ) validated_data.pop('password_repeat') password = validated_data.pop('password') user = super(UserSerializer, self).create( validated_data, ) user.set_password(password) user.save() return user
from copy import deepcopy from django.conf import settings from django.contrib.auth import login from rest_framework import serializers from rest_framework.exceptions import ValidationError from users.models import User class UserSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = User fields = ( 'username', 'email', 'gravatar', 'password', 'password_repeat', settings.DRF_URL_FIELD_NAME, ) extra_kwargs = { settings.DRF_URL_FIELD_NAME: { "view_name": "users:user-detail", }, } password = serializers.CharField( write_only=True, required=True, allow_blank=False, min_length=6, max_length=32, ) password_repeat = serializers.CharField( write_only=True, required=True, allow_blank=False, min_length=6, max_length=32, ) def create(self, validated_data): if validated_data['password'] != validated_data['password']: raise ValidationError( detail={ "password_repeat": "Tow password doesn't match", } ) validated_data.pop('password_repeat') password = validated_data.pop('password') user = super(UserSerializer, self).create( validated_data, ) user.set_password(password) user.save() login( self.context['request'], user=user, backend=settings.AUTHENTICATION_BACKENDS[0], ) return user
Fix bug in register aftter login
Fix: Fix bug in register aftter login
Python
bsd-2-clause
pinry/pinry,lapo-luchini/pinry,pinry/pinry,lapo-luchini/pinry,lapo-luchini/pinry,pinry/pinry,pinry/pinry,lapo-luchini/pinry
from django.conf import settings from rest_framework import serializers from rest_framework.exceptions import ValidationError from users.models import User class UserSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = User fields = ( 'username', 'email', 'gravatar', 'password', 'password_repeat', settings.DRF_URL_FIELD_NAME, ) extra_kwargs = { settings.DRF_URL_FIELD_NAME: { "view_name": "users:user-detail", }, } password = serializers.CharField( write_only=True, required=True, allow_blank=False, min_length=6, max_length=32, ) password_repeat = serializers.CharField( write_only=True, required=True, allow_blank=False, min_length=6, max_length=32, ) def create(self, validated_data): if validated_data['password'] != validated_data['password']: raise ValidationError( detail={ "password_repeat": "Tow password doesn't match", } ) validated_data.pop('password_repeat') password = validated_data.pop('password') user = super(UserSerializer, self).create( validated_data, ) user.set_password(password) user.save() return user Fix: Fix bug in register aftter login
from copy import deepcopy from django.conf import settings from django.contrib.auth import login from rest_framework import serializers from rest_framework.exceptions import ValidationError from users.models import User class UserSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = User fields = ( 'username', 'email', 'gravatar', 'password', 'password_repeat', settings.DRF_URL_FIELD_NAME, ) extra_kwargs = { settings.DRF_URL_FIELD_NAME: { "view_name": "users:user-detail", }, } password = serializers.CharField( write_only=True, required=True, allow_blank=False, min_length=6, max_length=32, ) password_repeat = serializers.CharField( write_only=True, required=True, allow_blank=False, min_length=6, max_length=32, ) def create(self, validated_data): if validated_data['password'] != validated_data['password']: raise ValidationError( detail={ "password_repeat": "Tow password doesn't match", } ) validated_data.pop('password_repeat') password = validated_data.pop('password') user = super(UserSerializer, self).create( validated_data, ) user.set_password(password) user.save() login( self.context['request'], user=user, backend=settings.AUTHENTICATION_BACKENDS[0], ) return user
<commit_before>from django.conf import settings from rest_framework import serializers from rest_framework.exceptions import ValidationError from users.models import User class UserSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = User fields = ( 'username', 'email', 'gravatar', 'password', 'password_repeat', settings.DRF_URL_FIELD_NAME, ) extra_kwargs = { settings.DRF_URL_FIELD_NAME: { "view_name": "users:user-detail", }, } password = serializers.CharField( write_only=True, required=True, allow_blank=False, min_length=6, max_length=32, ) password_repeat = serializers.CharField( write_only=True, required=True, allow_blank=False, min_length=6, max_length=32, ) def create(self, validated_data): if validated_data['password'] != validated_data['password']: raise ValidationError( detail={ "password_repeat": "Tow password doesn't match", } ) validated_data.pop('password_repeat') password = validated_data.pop('password') user = super(UserSerializer, self).create( validated_data, ) user.set_password(password) user.save() return user <commit_msg>Fix: Fix bug in register aftter login<commit_after>
from copy import deepcopy from django.conf import settings from django.contrib.auth import login from rest_framework import serializers from rest_framework.exceptions import ValidationError from users.models import User class UserSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = User fields = ( 'username', 'email', 'gravatar', 'password', 'password_repeat', settings.DRF_URL_FIELD_NAME, ) extra_kwargs = { settings.DRF_URL_FIELD_NAME: { "view_name": "users:user-detail", }, } password = serializers.CharField( write_only=True, required=True, allow_blank=False, min_length=6, max_length=32, ) password_repeat = serializers.CharField( write_only=True, required=True, allow_blank=False, min_length=6, max_length=32, ) def create(self, validated_data): if validated_data['password'] != validated_data['password']: raise ValidationError( detail={ "password_repeat": "Tow password doesn't match", } ) validated_data.pop('password_repeat') password = validated_data.pop('password') user = super(UserSerializer, self).create( validated_data, ) user.set_password(password) user.save() login( self.context['request'], user=user, backend=settings.AUTHENTICATION_BACKENDS[0], ) return user
from django.conf import settings from rest_framework import serializers from rest_framework.exceptions import ValidationError from users.models import User class UserSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = User fields = ( 'username', 'email', 'gravatar', 'password', 'password_repeat', settings.DRF_URL_FIELD_NAME, ) extra_kwargs = { settings.DRF_URL_FIELD_NAME: { "view_name": "users:user-detail", }, } password = serializers.CharField( write_only=True, required=True, allow_blank=False, min_length=6, max_length=32, ) password_repeat = serializers.CharField( write_only=True, required=True, allow_blank=False, min_length=6, max_length=32, ) def create(self, validated_data): if validated_data['password'] != validated_data['password']: raise ValidationError( detail={ "password_repeat": "Tow password doesn't match", } ) validated_data.pop('password_repeat') password = validated_data.pop('password') user = super(UserSerializer, self).create( validated_data, ) user.set_password(password) user.save() return user Fix: Fix bug in register aftter loginfrom copy import deepcopy from django.conf import settings from django.contrib.auth import login from rest_framework import serializers from rest_framework.exceptions import ValidationError from users.models import User class UserSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = User fields = ( 'username', 'email', 'gravatar', 'password', 'password_repeat', settings.DRF_URL_FIELD_NAME, ) extra_kwargs = { settings.DRF_URL_FIELD_NAME: { "view_name": "users:user-detail", }, } password = serializers.CharField( write_only=True, required=True, allow_blank=False, min_length=6, max_length=32, ) password_repeat = serializers.CharField( write_only=True, required=True, allow_blank=False, min_length=6, max_length=32, ) def create(self, validated_data): if validated_data['password'] != validated_data['password']: raise ValidationError( detail={ "password_repeat": "Tow password doesn't match", } ) validated_data.pop('password_repeat') password = validated_data.pop('password') user = super(UserSerializer, self).create( validated_data, ) user.set_password(password) user.save() login( self.context['request'], user=user, backend=settings.AUTHENTICATION_BACKENDS[0], ) return user
<commit_before>from django.conf import settings from rest_framework import serializers from rest_framework.exceptions import ValidationError from users.models import User class UserSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = User fields = ( 'username', 'email', 'gravatar', 'password', 'password_repeat', settings.DRF_URL_FIELD_NAME, ) extra_kwargs = { settings.DRF_URL_FIELD_NAME: { "view_name": "users:user-detail", }, } password = serializers.CharField( write_only=True, required=True, allow_blank=False, min_length=6, max_length=32, ) password_repeat = serializers.CharField( write_only=True, required=True, allow_blank=False, min_length=6, max_length=32, ) def create(self, validated_data): if validated_data['password'] != validated_data['password']: raise ValidationError( detail={ "password_repeat": "Tow password doesn't match", } ) validated_data.pop('password_repeat') password = validated_data.pop('password') user = super(UserSerializer, self).create( validated_data, ) user.set_password(password) user.save() return user <commit_msg>Fix: Fix bug in register aftter login<commit_after>from copy import deepcopy from django.conf import settings from django.contrib.auth import login from rest_framework import serializers from rest_framework.exceptions import ValidationError from users.models import User class UserSerializer(serializers.HyperlinkedModelSerializer): class Meta: model = User fields = ( 'username', 'email', 'gravatar', 'password', 'password_repeat', settings.DRF_URL_FIELD_NAME, ) extra_kwargs = { settings.DRF_URL_FIELD_NAME: { "view_name": "users:user-detail", }, } password = serializers.CharField( write_only=True, required=True, allow_blank=False, min_length=6, max_length=32, ) password_repeat = serializers.CharField( write_only=True, required=True, allow_blank=False, min_length=6, max_length=32, ) def create(self, validated_data): if validated_data['password'] != validated_data['password']: raise ValidationError( detail={ "password_repeat": "Tow password doesn't match", } ) validated_data.pop('password_repeat') password = validated_data.pop('password') user = super(UserSerializer, self).create( validated_data, ) user.set_password(password) user.save() login( self.context['request'], user=user, backend=settings.AUTHENTICATION_BACKENDS[0], ) return user
d8abffb340a852d69128977fa17bafdfc6babb71
tests/test_utils.py
tests/test_utils.py
from __future__ import print_function, division, absolute_import import numpy as np from .common import * from train.utils import preprocessImage, loadNetwork, predict from constants import * test_image = 234 * np.ones((*CAMERA_RESOLUTION, 3), dtype=np.uint8) def testPreprocessing(): image = preprocessImage(test_image, INPUT_WIDTH, INPUT_HEIGHT) # Check normalization assertEq(len(np.where(np.abs(image) > 1)[0]), 0) # Check resize assertEq(image.size, 3 * INPUT_WIDTH * INPUT_HEIGHT) # Check normalization on one element assertEq(image[0, 0, 0], ((234 / 255.) - 0.5) * 2) def testPredict(): model = loadNetwork(WEIGHTS_PTH, NUM_OUTPUT, MODEL_TYPE) x, y = predict(model, test_image)
from __future__ import print_function, division, absolute_import import numpy as np from .common import * from train.utils import preprocessImage, loadNetwork, predict from constants import * test_image = 234 * np.ones((MAX_WIDTH, MAX_HEIGHT, 3), dtype=np.uint8) def testPreprocessing(): image = preprocessImage(test_image, INPUT_WIDTH, INPUT_HEIGHT) # Check normalization assertEq(len(np.where(np.abs(image) > 1)[0]), 0) # Check resize assertEq(image.size, 3 * INPUT_WIDTH * INPUT_HEIGHT) # Check normalization on one element assertEq(image[0, 0, 0], ((234 / 255.) - 0.5) * 2) def testPredict(): model = loadNetwork(WEIGHTS_PTH, NUM_OUTPUT, MODEL_TYPE) x, y = predict(model, test_image)
Fix test for python 2
Fix test for python 2
Python
mit
sergionr2/RacingRobot,sergionr2/RacingRobot,sergionr2/RacingRobot,sergionr2/RacingRobot
from __future__ import print_function, division, absolute_import import numpy as np from .common import * from train.utils import preprocessImage, loadNetwork, predict from constants import * test_image = 234 * np.ones((*CAMERA_RESOLUTION, 3), dtype=np.uint8) def testPreprocessing(): image = preprocessImage(test_image, INPUT_WIDTH, INPUT_HEIGHT) # Check normalization assertEq(len(np.where(np.abs(image) > 1)[0]), 0) # Check resize assertEq(image.size, 3 * INPUT_WIDTH * INPUT_HEIGHT) # Check normalization on one element assertEq(image[0, 0, 0], ((234 / 255.) - 0.5) * 2) def testPredict(): model = loadNetwork(WEIGHTS_PTH, NUM_OUTPUT, MODEL_TYPE) x, y = predict(model, test_image) Fix test for python 2
from __future__ import print_function, division, absolute_import import numpy as np from .common import * from train.utils import preprocessImage, loadNetwork, predict from constants import * test_image = 234 * np.ones((MAX_WIDTH, MAX_HEIGHT, 3), dtype=np.uint8) def testPreprocessing(): image = preprocessImage(test_image, INPUT_WIDTH, INPUT_HEIGHT) # Check normalization assertEq(len(np.where(np.abs(image) > 1)[0]), 0) # Check resize assertEq(image.size, 3 * INPUT_WIDTH * INPUT_HEIGHT) # Check normalization on one element assertEq(image[0, 0, 0], ((234 / 255.) - 0.5) * 2) def testPredict(): model = loadNetwork(WEIGHTS_PTH, NUM_OUTPUT, MODEL_TYPE) x, y = predict(model, test_image)
<commit_before>from __future__ import print_function, division, absolute_import import numpy as np from .common import * from train.utils import preprocessImage, loadNetwork, predict from constants import * test_image = 234 * np.ones((*CAMERA_RESOLUTION, 3), dtype=np.uint8) def testPreprocessing(): image = preprocessImage(test_image, INPUT_WIDTH, INPUT_HEIGHT) # Check normalization assertEq(len(np.where(np.abs(image) > 1)[0]), 0) # Check resize assertEq(image.size, 3 * INPUT_WIDTH * INPUT_HEIGHT) # Check normalization on one element assertEq(image[0, 0, 0], ((234 / 255.) - 0.5) * 2) def testPredict(): model = loadNetwork(WEIGHTS_PTH, NUM_OUTPUT, MODEL_TYPE) x, y = predict(model, test_image) <commit_msg>Fix test for python 2<commit_after>
from __future__ import print_function, division, absolute_import import numpy as np from .common import * from train.utils import preprocessImage, loadNetwork, predict from constants import * test_image = 234 * np.ones((MAX_WIDTH, MAX_HEIGHT, 3), dtype=np.uint8) def testPreprocessing(): image = preprocessImage(test_image, INPUT_WIDTH, INPUT_HEIGHT) # Check normalization assertEq(len(np.where(np.abs(image) > 1)[0]), 0) # Check resize assertEq(image.size, 3 * INPUT_WIDTH * INPUT_HEIGHT) # Check normalization on one element assertEq(image[0, 0, 0], ((234 / 255.) - 0.5) * 2) def testPredict(): model = loadNetwork(WEIGHTS_PTH, NUM_OUTPUT, MODEL_TYPE) x, y = predict(model, test_image)
from __future__ import print_function, division, absolute_import import numpy as np from .common import * from train.utils import preprocessImage, loadNetwork, predict from constants import * test_image = 234 * np.ones((*CAMERA_RESOLUTION, 3), dtype=np.uint8) def testPreprocessing(): image = preprocessImage(test_image, INPUT_WIDTH, INPUT_HEIGHT) # Check normalization assertEq(len(np.where(np.abs(image) > 1)[0]), 0) # Check resize assertEq(image.size, 3 * INPUT_WIDTH * INPUT_HEIGHT) # Check normalization on one element assertEq(image[0, 0, 0], ((234 / 255.) - 0.5) * 2) def testPredict(): model = loadNetwork(WEIGHTS_PTH, NUM_OUTPUT, MODEL_TYPE) x, y = predict(model, test_image) Fix test for python 2from __future__ import print_function, division, absolute_import import numpy as np from .common import * from train.utils import preprocessImage, loadNetwork, predict from constants import * test_image = 234 * np.ones((MAX_WIDTH, MAX_HEIGHT, 3), dtype=np.uint8) def testPreprocessing(): image = preprocessImage(test_image, INPUT_WIDTH, INPUT_HEIGHT) # Check normalization assertEq(len(np.where(np.abs(image) > 1)[0]), 0) # Check resize assertEq(image.size, 3 * INPUT_WIDTH * INPUT_HEIGHT) # Check normalization on one element assertEq(image[0, 0, 0], ((234 / 255.) - 0.5) * 2) def testPredict(): model = loadNetwork(WEIGHTS_PTH, NUM_OUTPUT, MODEL_TYPE) x, y = predict(model, test_image)
<commit_before>from __future__ import print_function, division, absolute_import import numpy as np from .common import * from train.utils import preprocessImage, loadNetwork, predict from constants import * test_image = 234 * np.ones((*CAMERA_RESOLUTION, 3), dtype=np.uint8) def testPreprocessing(): image = preprocessImage(test_image, INPUT_WIDTH, INPUT_HEIGHT) # Check normalization assertEq(len(np.where(np.abs(image) > 1)[0]), 0) # Check resize assertEq(image.size, 3 * INPUT_WIDTH * INPUT_HEIGHT) # Check normalization on one element assertEq(image[0, 0, 0], ((234 / 255.) - 0.5) * 2) def testPredict(): model = loadNetwork(WEIGHTS_PTH, NUM_OUTPUT, MODEL_TYPE) x, y = predict(model, test_image) <commit_msg>Fix test for python 2<commit_after>from __future__ import print_function, division, absolute_import import numpy as np from .common import * from train.utils import preprocessImage, loadNetwork, predict from constants import * test_image = 234 * np.ones((MAX_WIDTH, MAX_HEIGHT, 3), dtype=np.uint8) def testPreprocessing(): image = preprocessImage(test_image, INPUT_WIDTH, INPUT_HEIGHT) # Check normalization assertEq(len(np.where(np.abs(image) > 1)[0]), 0) # Check resize assertEq(image.size, 3 * INPUT_WIDTH * INPUT_HEIGHT) # Check normalization on one element assertEq(image[0, 0, 0], ((234 / 255.) - 0.5) * 2) def testPredict(): model = loadNetwork(WEIGHTS_PTH, NUM_OUTPUT, MODEL_TYPE) x, y = predict(model, test_image)
9f216f1fdde41730b2680eed2174b1ba75d923be
dataset/dataset/spiders/dataset_spider.py
dataset/dataset/spiders/dataset_spider.py
from scrapy.contrib.spiders import CrawlSpider, Rule from scrapy.contrib.linkextractors.sgml import SgmlLinkExtractor from scrapy.selector import Selector from .. import items class DatasetSpider(CrawlSpider): name = 'dataset' allowed_domains = ['data.gc.ca'] start_urls = ['http://data.gc.ca/data/en/dataset?page=1'] rules = [Rule(SgmlLinkExtractor(allow=['/dataset/[0-9a-z]{8}-[0-9a-z]{4}-[0-9a-z]{4}-[0-9a-z]{4}-[0-9a-z]{12}']), 'parse_dataset')] def parse_dataset(self, response): sel = Selector(response) dataset = items.DatasetItem() dataset['url'] = response.url dataset['name'] = sel.xpath("//div[@class='span-6']/article/div[@class='module'][1]/section[@class='module-content indent-large'][1]/h1/text()").extract() dataset['frequency'] = sel.xpath("//div[@class='span-2']/aside[@class='secondary']/div[@class='module-related'][2]/ul[1]/li[@class='margin-bottom-medium']/text()").extract() return dataset
from scrapy.contrib.spiders import CrawlSpider, Rule from scrapy.contrib.linkextractors.sgml import SgmlLinkExtractor from scrapy.selector import Selector from .. import items class DatasetSpider(CrawlSpider): pages = 9466 name = 'dataset' allowed_domains = ['data.gc.ca'] start_urls = [] for i in range(1, pages + 1): start_urls.append('http://data.gc.ca/data/en/dataset?page=' + str(i)) rules = [Rule(SgmlLinkExtractor(allow=['/dataset/[0-9a-z]{8}-[0-9a-z]{4}-[0-9a-z]{4}-[0-9a-z]{4}-[0-9a-z]{12}']), 'parse_dataset')] def parse_dataset(self, response): sel = Selector(response) dataset = items.DatasetItem() dataset['url'] = response.url dataset['name'] = sel.xpath("//div[@class='span-6']/article/div[@class='module'][1]/section[@class='module-content indent-large'][1]/h1/text()").extract() dataset['frequency'] = sel.xpath("//div[@class='span-2']/aside[@class='secondary']/div[@class='module-related'][2]/ul[1]/li[@class='margin-bottom-medium']/text()").extract() return dataset
Add to start urls to contain all dataset pages
Add to start urls to contain all dataset pages
Python
mit
MaxLikelihood/CODE
from scrapy.contrib.spiders import CrawlSpider, Rule from scrapy.contrib.linkextractors.sgml import SgmlLinkExtractor from scrapy.selector import Selector from .. import items class DatasetSpider(CrawlSpider): name = 'dataset' allowed_domains = ['data.gc.ca'] start_urls = ['http://data.gc.ca/data/en/dataset?page=1'] rules = [Rule(SgmlLinkExtractor(allow=['/dataset/[0-9a-z]{8}-[0-9a-z]{4}-[0-9a-z]{4}-[0-9a-z]{4}-[0-9a-z]{12}']), 'parse_dataset')] def parse_dataset(self, response): sel = Selector(response) dataset = items.DatasetItem() dataset['url'] = response.url dataset['name'] = sel.xpath("//div[@class='span-6']/article/div[@class='module'][1]/section[@class='module-content indent-large'][1]/h1/text()").extract() dataset['frequency'] = sel.xpath("//div[@class='span-2']/aside[@class='secondary']/div[@class='module-related'][2]/ul[1]/li[@class='margin-bottom-medium']/text()").extract() return dataset Add to start urls to contain all dataset pages
from scrapy.contrib.spiders import CrawlSpider, Rule from scrapy.contrib.linkextractors.sgml import SgmlLinkExtractor from scrapy.selector import Selector from .. import items class DatasetSpider(CrawlSpider): pages = 9466 name = 'dataset' allowed_domains = ['data.gc.ca'] start_urls = [] for i in range(1, pages + 1): start_urls.append('http://data.gc.ca/data/en/dataset?page=' + str(i)) rules = [Rule(SgmlLinkExtractor(allow=['/dataset/[0-9a-z]{8}-[0-9a-z]{4}-[0-9a-z]{4}-[0-9a-z]{4}-[0-9a-z]{12}']), 'parse_dataset')] def parse_dataset(self, response): sel = Selector(response) dataset = items.DatasetItem() dataset['url'] = response.url dataset['name'] = sel.xpath("//div[@class='span-6']/article/div[@class='module'][1]/section[@class='module-content indent-large'][1]/h1/text()").extract() dataset['frequency'] = sel.xpath("//div[@class='span-2']/aside[@class='secondary']/div[@class='module-related'][2]/ul[1]/li[@class='margin-bottom-medium']/text()").extract() return dataset
<commit_before>from scrapy.contrib.spiders import CrawlSpider, Rule from scrapy.contrib.linkextractors.sgml import SgmlLinkExtractor from scrapy.selector import Selector from .. import items class DatasetSpider(CrawlSpider): name = 'dataset' allowed_domains = ['data.gc.ca'] start_urls = ['http://data.gc.ca/data/en/dataset?page=1'] rules = [Rule(SgmlLinkExtractor(allow=['/dataset/[0-9a-z]{8}-[0-9a-z]{4}-[0-9a-z]{4}-[0-9a-z]{4}-[0-9a-z]{12}']), 'parse_dataset')] def parse_dataset(self, response): sel = Selector(response) dataset = items.DatasetItem() dataset['url'] = response.url dataset['name'] = sel.xpath("//div[@class='span-6']/article/div[@class='module'][1]/section[@class='module-content indent-large'][1]/h1/text()").extract() dataset['frequency'] = sel.xpath("//div[@class='span-2']/aside[@class='secondary']/div[@class='module-related'][2]/ul[1]/li[@class='margin-bottom-medium']/text()").extract() return dataset <commit_msg>Add to start urls to contain all dataset pages<commit_after>
from scrapy.contrib.spiders import CrawlSpider, Rule from scrapy.contrib.linkextractors.sgml import SgmlLinkExtractor from scrapy.selector import Selector from .. import items class DatasetSpider(CrawlSpider): pages = 9466 name = 'dataset' allowed_domains = ['data.gc.ca'] start_urls = [] for i in range(1, pages + 1): start_urls.append('http://data.gc.ca/data/en/dataset?page=' + str(i)) rules = [Rule(SgmlLinkExtractor(allow=['/dataset/[0-9a-z]{8}-[0-9a-z]{4}-[0-9a-z]{4}-[0-9a-z]{4}-[0-9a-z]{12}']), 'parse_dataset')] def parse_dataset(self, response): sel = Selector(response) dataset = items.DatasetItem() dataset['url'] = response.url dataset['name'] = sel.xpath("//div[@class='span-6']/article/div[@class='module'][1]/section[@class='module-content indent-large'][1]/h1/text()").extract() dataset['frequency'] = sel.xpath("//div[@class='span-2']/aside[@class='secondary']/div[@class='module-related'][2]/ul[1]/li[@class='margin-bottom-medium']/text()").extract() return dataset
from scrapy.contrib.spiders import CrawlSpider, Rule from scrapy.contrib.linkextractors.sgml import SgmlLinkExtractor from scrapy.selector import Selector from .. import items class DatasetSpider(CrawlSpider): name = 'dataset' allowed_domains = ['data.gc.ca'] start_urls = ['http://data.gc.ca/data/en/dataset?page=1'] rules = [Rule(SgmlLinkExtractor(allow=['/dataset/[0-9a-z]{8}-[0-9a-z]{4}-[0-9a-z]{4}-[0-9a-z]{4}-[0-9a-z]{12}']), 'parse_dataset')] def parse_dataset(self, response): sel = Selector(response) dataset = items.DatasetItem() dataset['url'] = response.url dataset['name'] = sel.xpath("//div[@class='span-6']/article/div[@class='module'][1]/section[@class='module-content indent-large'][1]/h1/text()").extract() dataset['frequency'] = sel.xpath("//div[@class='span-2']/aside[@class='secondary']/div[@class='module-related'][2]/ul[1]/li[@class='margin-bottom-medium']/text()").extract() return dataset Add to start urls to contain all dataset pagesfrom scrapy.contrib.spiders import CrawlSpider, Rule from scrapy.contrib.linkextractors.sgml import SgmlLinkExtractor from scrapy.selector import Selector from .. import items class DatasetSpider(CrawlSpider): pages = 9466 name = 'dataset' allowed_domains = ['data.gc.ca'] start_urls = [] for i in range(1, pages + 1): start_urls.append('http://data.gc.ca/data/en/dataset?page=' + str(i)) rules = [Rule(SgmlLinkExtractor(allow=['/dataset/[0-9a-z]{8}-[0-9a-z]{4}-[0-9a-z]{4}-[0-9a-z]{4}-[0-9a-z]{12}']), 'parse_dataset')] def parse_dataset(self, response): sel = Selector(response) dataset = items.DatasetItem() dataset['url'] = response.url dataset['name'] = sel.xpath("//div[@class='span-6']/article/div[@class='module'][1]/section[@class='module-content indent-large'][1]/h1/text()").extract() dataset['frequency'] = sel.xpath("//div[@class='span-2']/aside[@class='secondary']/div[@class='module-related'][2]/ul[1]/li[@class='margin-bottom-medium']/text()").extract() return dataset
<commit_before>from scrapy.contrib.spiders import CrawlSpider, Rule from scrapy.contrib.linkextractors.sgml import SgmlLinkExtractor from scrapy.selector import Selector from .. import items class DatasetSpider(CrawlSpider): name = 'dataset' allowed_domains = ['data.gc.ca'] start_urls = ['http://data.gc.ca/data/en/dataset?page=1'] rules = [Rule(SgmlLinkExtractor(allow=['/dataset/[0-9a-z]{8}-[0-9a-z]{4}-[0-9a-z]{4}-[0-9a-z]{4}-[0-9a-z]{12}']), 'parse_dataset')] def parse_dataset(self, response): sel = Selector(response) dataset = items.DatasetItem() dataset['url'] = response.url dataset['name'] = sel.xpath("//div[@class='span-6']/article/div[@class='module'][1]/section[@class='module-content indent-large'][1]/h1/text()").extract() dataset['frequency'] = sel.xpath("//div[@class='span-2']/aside[@class='secondary']/div[@class='module-related'][2]/ul[1]/li[@class='margin-bottom-medium']/text()").extract() return dataset <commit_msg>Add to start urls to contain all dataset pages<commit_after>from scrapy.contrib.spiders import CrawlSpider, Rule from scrapy.contrib.linkextractors.sgml import SgmlLinkExtractor from scrapy.selector import Selector from .. import items class DatasetSpider(CrawlSpider): pages = 9466 name = 'dataset' allowed_domains = ['data.gc.ca'] start_urls = [] for i in range(1, pages + 1): start_urls.append('http://data.gc.ca/data/en/dataset?page=' + str(i)) rules = [Rule(SgmlLinkExtractor(allow=['/dataset/[0-9a-z]{8}-[0-9a-z]{4}-[0-9a-z]{4}-[0-9a-z]{4}-[0-9a-z]{12}']), 'parse_dataset')] def parse_dataset(self, response): sel = Selector(response) dataset = items.DatasetItem() dataset['url'] = response.url dataset['name'] = sel.xpath("//div[@class='span-6']/article/div[@class='module'][1]/section[@class='module-content indent-large'][1]/h1/text()").extract() dataset['frequency'] = sel.xpath("//div[@class='span-2']/aside[@class='secondary']/div[@class='module-related'][2]/ul[1]/li[@class='margin-bottom-medium']/text()").extract() return dataset
5ac88fd5a10444b8f6384d127c13216545319169
vies/__init__.py
vies/__init__.py
# -*- coding: utf-8 -*- import logging __version__ = "5.0.1" logger = logging.getLogger('vies') VIES_WSDL_URL = str('http://ec.europa.eu/taxation_customs/vies/checkVatService.wsdl') # NoQA VATIN_MAX_LENGTH = 14
# -*- coding: utf-8 -*- import logging __version__ = "5.0.2" logger = logging.getLogger('vies') VIES_WSDL_URL = str('https://ec.europa.eu/taxation_customs/vies/checkVatService.wsdl') # NoQA VATIN_MAX_LENGTH = 14
Fix vies url -- Increase version number
Fix vies url -- Increase version number
Python
mit
codingjoe/django-vies
# -*- coding: utf-8 -*- import logging __version__ = "5.0.1" logger = logging.getLogger('vies') VIES_WSDL_URL = str('http://ec.europa.eu/taxation_customs/vies/checkVatService.wsdl') # NoQA VATIN_MAX_LENGTH = 14 Fix vies url -- Increase version number
# -*- coding: utf-8 -*- import logging __version__ = "5.0.2" logger = logging.getLogger('vies') VIES_WSDL_URL = str('https://ec.europa.eu/taxation_customs/vies/checkVatService.wsdl') # NoQA VATIN_MAX_LENGTH = 14
<commit_before># -*- coding: utf-8 -*- import logging __version__ = "5.0.1" logger = logging.getLogger('vies') VIES_WSDL_URL = str('http://ec.europa.eu/taxation_customs/vies/checkVatService.wsdl') # NoQA VATIN_MAX_LENGTH = 14 <commit_msg>Fix vies url -- Increase version number<commit_after>
# -*- coding: utf-8 -*- import logging __version__ = "5.0.2" logger = logging.getLogger('vies') VIES_WSDL_URL = str('https://ec.europa.eu/taxation_customs/vies/checkVatService.wsdl') # NoQA VATIN_MAX_LENGTH = 14
# -*- coding: utf-8 -*- import logging __version__ = "5.0.1" logger = logging.getLogger('vies') VIES_WSDL_URL = str('http://ec.europa.eu/taxation_customs/vies/checkVatService.wsdl') # NoQA VATIN_MAX_LENGTH = 14 Fix vies url -- Increase version number# -*- coding: utf-8 -*- import logging __version__ = "5.0.2" logger = logging.getLogger('vies') VIES_WSDL_URL = str('https://ec.europa.eu/taxation_customs/vies/checkVatService.wsdl') # NoQA VATIN_MAX_LENGTH = 14
<commit_before># -*- coding: utf-8 -*- import logging __version__ = "5.0.1" logger = logging.getLogger('vies') VIES_WSDL_URL = str('http://ec.europa.eu/taxation_customs/vies/checkVatService.wsdl') # NoQA VATIN_MAX_LENGTH = 14 <commit_msg>Fix vies url -- Increase version number<commit_after># -*- coding: utf-8 -*- import logging __version__ = "5.0.2" logger = logging.getLogger('vies') VIES_WSDL_URL = str('https://ec.europa.eu/taxation_customs/vies/checkVatService.wsdl') # NoQA VATIN_MAX_LENGTH = 14
95e61ccdebc33c1c610d0672558cd00798c3105f
packages/grid/backend/grid/api/users/models.py
packages/grid/backend/grid/api/users/models.py
# stdlib from typing import Optional from typing import Union # third party from nacl.encoding import HexEncoder from nacl.signing import SigningKey from pydantic import BaseModel from pydantic import EmailStr class BaseUser(BaseModel): email: Optional[EmailStr] name: Optional[str] role: Union[Optional[int], Optional[str]] # TODO: Should be int in SyftUser daa_pdf: Optional[bytes] = b"" class Config: orm_mode = True class UserCreate(BaseUser): email: EmailStr role: str = "Data Scientist" name: str password: str class UserUpdate(BaseUser): password: Optional[str] budget: Optional[float] class UserCandidate(BaseUser): email: EmailStr status: str = "pending" name: str class User(BaseUser): id: int role: Union[int, str] # TODO: This should be int. Perhaps add role_name instead? budget_spent: Optional[float] institution: Optional[str] website: Optional[str] added_by: Optional[str] class UserPrivate(User): private_key: str def get_signing_key(self) -> SigningKey: return SigningKey(self.private_key.encode(), encoder=HexEncoder) class UserSyft(User): hashed_password: str salt: str verify_key: str
# stdlib from typing import Optional from typing import Union # third party from nacl.encoding import HexEncoder from nacl.signing import SigningKey from pydantic import BaseModel from pydantic import EmailStr class BaseUser(BaseModel): email: Optional[EmailStr] name: Optional[str] role: Union[Optional[int], Optional[str]] # TODO: Should be int in SyftUser daa_pdf: Optional[bytes] = b"" class Config: orm_mode = True class UserCreate(BaseUser): email: EmailStr role: str = "Data Scientist" name: str password: str institution: Optional[str] website: Optional[str] class ApplicantStatus(BaseModel): status: str class UserUpdate(BaseUser): password: Optional[str] budget: Optional[float] class UserCandidate(BaseUser): email: EmailStr status: str = "pending" name: str class User(BaseUser): id: int role: Union[int, str] # TODO: This should be int. Perhaps add role_name instead? budget_spent: Optional[float] institution: Optional[str] website: Optional[str] added_by: Optional[str] class UserPrivate(User): private_key: str def get_signing_key(self) -> SigningKey: return SigningKey(self.private_key.encode(), encoder=HexEncoder) class UserSyft(User): hashed_password: str salt: str verify_key: str
ADD institution / website as optional fields during user creation
ADD institution / website as optional fields during user creation
Python
apache-2.0
OpenMined/PySyft,OpenMined/PySyft,OpenMined/PySyft,OpenMined/PySyft
# stdlib from typing import Optional from typing import Union # third party from nacl.encoding import HexEncoder from nacl.signing import SigningKey from pydantic import BaseModel from pydantic import EmailStr class BaseUser(BaseModel): email: Optional[EmailStr] name: Optional[str] role: Union[Optional[int], Optional[str]] # TODO: Should be int in SyftUser daa_pdf: Optional[bytes] = b"" class Config: orm_mode = True class UserCreate(BaseUser): email: EmailStr role: str = "Data Scientist" name: str password: str class UserUpdate(BaseUser): password: Optional[str] budget: Optional[float] class UserCandidate(BaseUser): email: EmailStr status: str = "pending" name: str class User(BaseUser): id: int role: Union[int, str] # TODO: This should be int. Perhaps add role_name instead? budget_spent: Optional[float] institution: Optional[str] website: Optional[str] added_by: Optional[str] class UserPrivate(User): private_key: str def get_signing_key(self) -> SigningKey: return SigningKey(self.private_key.encode(), encoder=HexEncoder) class UserSyft(User): hashed_password: str salt: str verify_key: str ADD institution / website as optional fields during user creation
# stdlib from typing import Optional from typing import Union # third party from nacl.encoding import HexEncoder from nacl.signing import SigningKey from pydantic import BaseModel from pydantic import EmailStr class BaseUser(BaseModel): email: Optional[EmailStr] name: Optional[str] role: Union[Optional[int], Optional[str]] # TODO: Should be int in SyftUser daa_pdf: Optional[bytes] = b"" class Config: orm_mode = True class UserCreate(BaseUser): email: EmailStr role: str = "Data Scientist" name: str password: str institution: Optional[str] website: Optional[str] class ApplicantStatus(BaseModel): status: str class UserUpdate(BaseUser): password: Optional[str] budget: Optional[float] class UserCandidate(BaseUser): email: EmailStr status: str = "pending" name: str class User(BaseUser): id: int role: Union[int, str] # TODO: This should be int. Perhaps add role_name instead? budget_spent: Optional[float] institution: Optional[str] website: Optional[str] added_by: Optional[str] class UserPrivate(User): private_key: str def get_signing_key(self) -> SigningKey: return SigningKey(self.private_key.encode(), encoder=HexEncoder) class UserSyft(User): hashed_password: str salt: str verify_key: str
<commit_before># stdlib from typing import Optional from typing import Union # third party from nacl.encoding import HexEncoder from nacl.signing import SigningKey from pydantic import BaseModel from pydantic import EmailStr class BaseUser(BaseModel): email: Optional[EmailStr] name: Optional[str] role: Union[Optional[int], Optional[str]] # TODO: Should be int in SyftUser daa_pdf: Optional[bytes] = b"" class Config: orm_mode = True class UserCreate(BaseUser): email: EmailStr role: str = "Data Scientist" name: str password: str class UserUpdate(BaseUser): password: Optional[str] budget: Optional[float] class UserCandidate(BaseUser): email: EmailStr status: str = "pending" name: str class User(BaseUser): id: int role: Union[int, str] # TODO: This should be int. Perhaps add role_name instead? budget_spent: Optional[float] institution: Optional[str] website: Optional[str] added_by: Optional[str] class UserPrivate(User): private_key: str def get_signing_key(self) -> SigningKey: return SigningKey(self.private_key.encode(), encoder=HexEncoder) class UserSyft(User): hashed_password: str salt: str verify_key: str <commit_msg>ADD institution / website as optional fields during user creation<commit_after>
# stdlib from typing import Optional from typing import Union # third party from nacl.encoding import HexEncoder from nacl.signing import SigningKey from pydantic import BaseModel from pydantic import EmailStr class BaseUser(BaseModel): email: Optional[EmailStr] name: Optional[str] role: Union[Optional[int], Optional[str]] # TODO: Should be int in SyftUser daa_pdf: Optional[bytes] = b"" class Config: orm_mode = True class UserCreate(BaseUser): email: EmailStr role: str = "Data Scientist" name: str password: str institution: Optional[str] website: Optional[str] class ApplicantStatus(BaseModel): status: str class UserUpdate(BaseUser): password: Optional[str] budget: Optional[float] class UserCandidate(BaseUser): email: EmailStr status: str = "pending" name: str class User(BaseUser): id: int role: Union[int, str] # TODO: This should be int. Perhaps add role_name instead? budget_spent: Optional[float] institution: Optional[str] website: Optional[str] added_by: Optional[str] class UserPrivate(User): private_key: str def get_signing_key(self) -> SigningKey: return SigningKey(self.private_key.encode(), encoder=HexEncoder) class UserSyft(User): hashed_password: str salt: str verify_key: str
# stdlib from typing import Optional from typing import Union # third party from nacl.encoding import HexEncoder from nacl.signing import SigningKey from pydantic import BaseModel from pydantic import EmailStr class BaseUser(BaseModel): email: Optional[EmailStr] name: Optional[str] role: Union[Optional[int], Optional[str]] # TODO: Should be int in SyftUser daa_pdf: Optional[bytes] = b"" class Config: orm_mode = True class UserCreate(BaseUser): email: EmailStr role: str = "Data Scientist" name: str password: str class UserUpdate(BaseUser): password: Optional[str] budget: Optional[float] class UserCandidate(BaseUser): email: EmailStr status: str = "pending" name: str class User(BaseUser): id: int role: Union[int, str] # TODO: This should be int. Perhaps add role_name instead? budget_spent: Optional[float] institution: Optional[str] website: Optional[str] added_by: Optional[str] class UserPrivate(User): private_key: str def get_signing_key(self) -> SigningKey: return SigningKey(self.private_key.encode(), encoder=HexEncoder) class UserSyft(User): hashed_password: str salt: str verify_key: str ADD institution / website as optional fields during user creation# stdlib from typing import Optional from typing import Union # third party from nacl.encoding import HexEncoder from nacl.signing import SigningKey from pydantic import BaseModel from pydantic import EmailStr class BaseUser(BaseModel): email: Optional[EmailStr] name: Optional[str] role: Union[Optional[int], Optional[str]] # TODO: Should be int in SyftUser daa_pdf: Optional[bytes] = b"" class Config: orm_mode = True class UserCreate(BaseUser): email: EmailStr role: str = "Data Scientist" name: str password: str institution: Optional[str] website: Optional[str] class ApplicantStatus(BaseModel): status: str class UserUpdate(BaseUser): password: Optional[str] budget: Optional[float] class UserCandidate(BaseUser): email: EmailStr status: str = "pending" name: str class User(BaseUser): id: int role: Union[int, str] # TODO: This should be int. Perhaps add role_name instead? budget_spent: Optional[float] institution: Optional[str] website: Optional[str] added_by: Optional[str] class UserPrivate(User): private_key: str def get_signing_key(self) -> SigningKey: return SigningKey(self.private_key.encode(), encoder=HexEncoder) class UserSyft(User): hashed_password: str salt: str verify_key: str
<commit_before># stdlib from typing import Optional from typing import Union # third party from nacl.encoding import HexEncoder from nacl.signing import SigningKey from pydantic import BaseModel from pydantic import EmailStr class BaseUser(BaseModel): email: Optional[EmailStr] name: Optional[str] role: Union[Optional[int], Optional[str]] # TODO: Should be int in SyftUser daa_pdf: Optional[bytes] = b"" class Config: orm_mode = True class UserCreate(BaseUser): email: EmailStr role: str = "Data Scientist" name: str password: str class UserUpdate(BaseUser): password: Optional[str] budget: Optional[float] class UserCandidate(BaseUser): email: EmailStr status: str = "pending" name: str class User(BaseUser): id: int role: Union[int, str] # TODO: This should be int. Perhaps add role_name instead? budget_spent: Optional[float] institution: Optional[str] website: Optional[str] added_by: Optional[str] class UserPrivate(User): private_key: str def get_signing_key(self) -> SigningKey: return SigningKey(self.private_key.encode(), encoder=HexEncoder) class UserSyft(User): hashed_password: str salt: str verify_key: str <commit_msg>ADD institution / website as optional fields during user creation<commit_after># stdlib from typing import Optional from typing import Union # third party from nacl.encoding import HexEncoder from nacl.signing import SigningKey from pydantic import BaseModel from pydantic import EmailStr class BaseUser(BaseModel): email: Optional[EmailStr] name: Optional[str] role: Union[Optional[int], Optional[str]] # TODO: Should be int in SyftUser daa_pdf: Optional[bytes] = b"" class Config: orm_mode = True class UserCreate(BaseUser): email: EmailStr role: str = "Data Scientist" name: str password: str institution: Optional[str] website: Optional[str] class ApplicantStatus(BaseModel): status: str class UserUpdate(BaseUser): password: Optional[str] budget: Optional[float] class UserCandidate(BaseUser): email: EmailStr status: str = "pending" name: str class User(BaseUser): id: int role: Union[int, str] # TODO: This should be int. Perhaps add role_name instead? budget_spent: Optional[float] institution: Optional[str] website: Optional[str] added_by: Optional[str] class UserPrivate(User): private_key: str def get_signing_key(self) -> SigningKey: return SigningKey(self.private_key.encode(), encoder=HexEncoder) class UserSyft(User): hashed_password: str salt: str verify_key: str
3921c00a6eb4d0dc50bf3efaaeb9b91ff3ad7608
vcs_gutter_change.py
vcs_gutter_change.py
import sublime_plugin try: from VcsGutter.view_collection import ViewCollection except ImportError: from view_collection import ViewCollection class VcsGutterBaseChangeCommand(sublime_plugin.WindowCommand): def lines_to_blocks(self, lines): blocks = [] last_line = -2 for line in lines: if line > last_line+1: blocks.append(line) last_line = line return blocks def run(self): view = self.window.active_view() inserted, modified, deleted = ViewCollection.diff(view) inserted = self.lines_to_blocks(inserted) modified = self.lines_to_blocks(modified) all_changes = sorted(inserted + modified + deleted) if all_changes: row, col = view.rowcol(view.sel()[0].begin()) current_row = row + 1 line = self.jump(all_changes, current_row) self.window.active_view().run_command("goto_line", {"line": line}) class VcsGutterNextChangeCommand(VcsGutterBaseChangeCommand): def jump(self, all_changes, current_row): return next((change for change in all_changes if change > current_row), all_changes[0]) class VcsGutterPrevChangeCommand(VcsGutterBaseChangeCommand): def jump(self, all_changes, current_row): return next((change for change in reversed(all_changes) if change < current_row), all_changes[-1])
import sublime_plugin try: from .view_collection import ViewCollection except ImportError: from view_collection import ViewCollection class VcsGutterBaseChangeCommand(sublime_plugin.WindowCommand): def lines_to_blocks(self, lines): blocks = [] last_line = -2 for line in lines: if line > last_line+1: blocks.append(line) last_line = line return blocks def run(self): view = self.window.active_view() inserted, modified, deleted = ViewCollection.diff(view) inserted = self.lines_to_blocks(inserted) modified = self.lines_to_blocks(modified) all_changes = sorted(inserted + modified + deleted) if all_changes: row, col = view.rowcol(view.sel()[0].begin()) current_row = row + 1 line = self.jump(all_changes, current_row) self.window.active_view().run_command("goto_line", {"line": line}) class VcsGutterNextChangeCommand(VcsGutterBaseChangeCommand): def jump(self, all_changes, current_row): return next((change for change in all_changes if change > current_row), all_changes[0]) class VcsGutterPrevChangeCommand(VcsGutterBaseChangeCommand): def jump(self, all_changes, current_row): return next((change for change in reversed(all_changes) if change < current_row), all_changes[-1])
Fix ImportError when loading change_(prev|next) module on windows
Fix ImportError when loading change_(prev|next) module on windows
Python
mit
ariofrio/VcsGutter,bradsokol/VcsGutter,ariofrio/VcsGutter,bradsokol/VcsGutter
import sublime_plugin try: from VcsGutter.view_collection import ViewCollection except ImportError: from view_collection import ViewCollection class VcsGutterBaseChangeCommand(sublime_plugin.WindowCommand): def lines_to_blocks(self, lines): blocks = [] last_line = -2 for line in lines: if line > last_line+1: blocks.append(line) last_line = line return blocks def run(self): view = self.window.active_view() inserted, modified, deleted = ViewCollection.diff(view) inserted = self.lines_to_blocks(inserted) modified = self.lines_to_blocks(modified) all_changes = sorted(inserted + modified + deleted) if all_changes: row, col = view.rowcol(view.sel()[0].begin()) current_row = row + 1 line = self.jump(all_changes, current_row) self.window.active_view().run_command("goto_line", {"line": line}) class VcsGutterNextChangeCommand(VcsGutterBaseChangeCommand): def jump(self, all_changes, current_row): return next((change for change in all_changes if change > current_row), all_changes[0]) class VcsGutterPrevChangeCommand(VcsGutterBaseChangeCommand): def jump(self, all_changes, current_row): return next((change for change in reversed(all_changes) if change < current_row), all_changes[-1]) Fix ImportError when loading change_(prev|next) module on windows
import sublime_plugin try: from .view_collection import ViewCollection except ImportError: from view_collection import ViewCollection class VcsGutterBaseChangeCommand(sublime_plugin.WindowCommand): def lines_to_blocks(self, lines): blocks = [] last_line = -2 for line in lines: if line > last_line+1: blocks.append(line) last_line = line return blocks def run(self): view = self.window.active_view() inserted, modified, deleted = ViewCollection.diff(view) inserted = self.lines_to_blocks(inserted) modified = self.lines_to_blocks(modified) all_changes = sorted(inserted + modified + deleted) if all_changes: row, col = view.rowcol(view.sel()[0].begin()) current_row = row + 1 line = self.jump(all_changes, current_row) self.window.active_view().run_command("goto_line", {"line": line}) class VcsGutterNextChangeCommand(VcsGutterBaseChangeCommand): def jump(self, all_changes, current_row): return next((change for change in all_changes if change > current_row), all_changes[0]) class VcsGutterPrevChangeCommand(VcsGutterBaseChangeCommand): def jump(self, all_changes, current_row): return next((change for change in reversed(all_changes) if change < current_row), all_changes[-1])
<commit_before>import sublime_plugin try: from VcsGutter.view_collection import ViewCollection except ImportError: from view_collection import ViewCollection class VcsGutterBaseChangeCommand(sublime_plugin.WindowCommand): def lines_to_blocks(self, lines): blocks = [] last_line = -2 for line in lines: if line > last_line+1: blocks.append(line) last_line = line return blocks def run(self): view = self.window.active_view() inserted, modified, deleted = ViewCollection.diff(view) inserted = self.lines_to_blocks(inserted) modified = self.lines_to_blocks(modified) all_changes = sorted(inserted + modified + deleted) if all_changes: row, col = view.rowcol(view.sel()[0].begin()) current_row = row + 1 line = self.jump(all_changes, current_row) self.window.active_view().run_command("goto_line", {"line": line}) class VcsGutterNextChangeCommand(VcsGutterBaseChangeCommand): def jump(self, all_changes, current_row): return next((change for change in all_changes if change > current_row), all_changes[0]) class VcsGutterPrevChangeCommand(VcsGutterBaseChangeCommand): def jump(self, all_changes, current_row): return next((change for change in reversed(all_changes) if change < current_row), all_changes[-1]) <commit_msg>Fix ImportError when loading change_(prev|next) module on windows<commit_after>
import sublime_plugin try: from .view_collection import ViewCollection except ImportError: from view_collection import ViewCollection class VcsGutterBaseChangeCommand(sublime_plugin.WindowCommand): def lines_to_blocks(self, lines): blocks = [] last_line = -2 for line in lines: if line > last_line+1: blocks.append(line) last_line = line return blocks def run(self): view = self.window.active_view() inserted, modified, deleted = ViewCollection.diff(view) inserted = self.lines_to_blocks(inserted) modified = self.lines_to_blocks(modified) all_changes = sorted(inserted + modified + deleted) if all_changes: row, col = view.rowcol(view.sel()[0].begin()) current_row = row + 1 line = self.jump(all_changes, current_row) self.window.active_view().run_command("goto_line", {"line": line}) class VcsGutterNextChangeCommand(VcsGutterBaseChangeCommand): def jump(self, all_changes, current_row): return next((change for change in all_changes if change > current_row), all_changes[0]) class VcsGutterPrevChangeCommand(VcsGutterBaseChangeCommand): def jump(self, all_changes, current_row): return next((change for change in reversed(all_changes) if change < current_row), all_changes[-1])
import sublime_plugin try: from VcsGutter.view_collection import ViewCollection except ImportError: from view_collection import ViewCollection class VcsGutterBaseChangeCommand(sublime_plugin.WindowCommand): def lines_to_blocks(self, lines): blocks = [] last_line = -2 for line in lines: if line > last_line+1: blocks.append(line) last_line = line return blocks def run(self): view = self.window.active_view() inserted, modified, deleted = ViewCollection.diff(view) inserted = self.lines_to_blocks(inserted) modified = self.lines_to_blocks(modified) all_changes = sorted(inserted + modified + deleted) if all_changes: row, col = view.rowcol(view.sel()[0].begin()) current_row = row + 1 line = self.jump(all_changes, current_row) self.window.active_view().run_command("goto_line", {"line": line}) class VcsGutterNextChangeCommand(VcsGutterBaseChangeCommand): def jump(self, all_changes, current_row): return next((change for change in all_changes if change > current_row), all_changes[0]) class VcsGutterPrevChangeCommand(VcsGutterBaseChangeCommand): def jump(self, all_changes, current_row): return next((change for change in reversed(all_changes) if change < current_row), all_changes[-1]) Fix ImportError when loading change_(prev|next) module on windowsimport sublime_plugin try: from .view_collection import ViewCollection except ImportError: from view_collection import ViewCollection class VcsGutterBaseChangeCommand(sublime_plugin.WindowCommand): def lines_to_blocks(self, lines): blocks = [] last_line = -2 for line in lines: if line > last_line+1: blocks.append(line) last_line = line return blocks def run(self): view = self.window.active_view() inserted, modified, deleted = ViewCollection.diff(view) inserted = self.lines_to_blocks(inserted) modified = self.lines_to_blocks(modified) all_changes = sorted(inserted + modified + deleted) if all_changes: row, col = view.rowcol(view.sel()[0].begin()) current_row = row + 1 line = self.jump(all_changes, current_row) self.window.active_view().run_command("goto_line", {"line": line}) class VcsGutterNextChangeCommand(VcsGutterBaseChangeCommand): def jump(self, all_changes, current_row): return next((change for change in all_changes if change > current_row), all_changes[0]) class VcsGutterPrevChangeCommand(VcsGutterBaseChangeCommand): def jump(self, all_changes, current_row): return next((change for change in reversed(all_changes) if change < current_row), all_changes[-1])
<commit_before>import sublime_plugin try: from VcsGutter.view_collection import ViewCollection except ImportError: from view_collection import ViewCollection class VcsGutterBaseChangeCommand(sublime_plugin.WindowCommand): def lines_to_blocks(self, lines): blocks = [] last_line = -2 for line in lines: if line > last_line+1: blocks.append(line) last_line = line return blocks def run(self): view = self.window.active_view() inserted, modified, deleted = ViewCollection.diff(view) inserted = self.lines_to_blocks(inserted) modified = self.lines_to_blocks(modified) all_changes = sorted(inserted + modified + deleted) if all_changes: row, col = view.rowcol(view.sel()[0].begin()) current_row = row + 1 line = self.jump(all_changes, current_row) self.window.active_view().run_command("goto_line", {"line": line}) class VcsGutterNextChangeCommand(VcsGutterBaseChangeCommand): def jump(self, all_changes, current_row): return next((change for change in all_changes if change > current_row), all_changes[0]) class VcsGutterPrevChangeCommand(VcsGutterBaseChangeCommand): def jump(self, all_changes, current_row): return next((change for change in reversed(all_changes) if change < current_row), all_changes[-1]) <commit_msg>Fix ImportError when loading change_(prev|next) module on windows<commit_after>import sublime_plugin try: from .view_collection import ViewCollection except ImportError: from view_collection import ViewCollection class VcsGutterBaseChangeCommand(sublime_plugin.WindowCommand): def lines_to_blocks(self, lines): blocks = [] last_line = -2 for line in lines: if line > last_line+1: blocks.append(line) last_line = line return blocks def run(self): view = self.window.active_view() inserted, modified, deleted = ViewCollection.diff(view) inserted = self.lines_to_blocks(inserted) modified = self.lines_to_blocks(modified) all_changes = sorted(inserted + modified + deleted) if all_changes: row, col = view.rowcol(view.sel()[0].begin()) current_row = row + 1 line = self.jump(all_changes, current_row) self.window.active_view().run_command("goto_line", {"line": line}) class VcsGutterNextChangeCommand(VcsGutterBaseChangeCommand): def jump(self, all_changes, current_row): return next((change for change in all_changes if change > current_row), all_changes[0]) class VcsGutterPrevChangeCommand(VcsGutterBaseChangeCommand): def jump(self, all_changes, current_row): return next((change for change in reversed(all_changes) if change < current_row), all_changes[-1])
affad020348ca8aa6a7b9431811d707ab8f6d99a
pyramid/__init__.py
pyramid/__init__.py
# -*- coding: utf-8 -*- # # Author: Taylor Smith <taylor.smith@alkaline-ml.com> # # The pyramid module __version__ = "0.7.0-dev" try: # this var is injected in the setup build to enable # the retrieval of the version number without actually # importing the un-built submodules. __PYRAMID_SETUP__ except NameError: __PYRAMID_SETUP__ = False if __PYRAMID_SETUP__: import sys import os sys.stderr.write('Partial import of pyramid during the build process.' + os.linesep) else: # check that the build completed properly. This prints an informative # message in the case that any of the C code was not properly compiled. from . import __check_build __all__ = [ 'arima', 'compat', 'datasets', 'utils' ] def setup_module(module): import numpy as np import random _random_seed = int(np.random.uniform() * (2 ** 31 - 1)) np.random.seed(_random_seed) random.seed(_random_seed)
# -*- coding: utf-8 -*- # # Author: Taylor Smith <taylor.smith@alkaline-ml.com> # # The pyramid module __version__ = "0.7.0" try: # this var is injected in the setup build to enable # the retrieval of the version number without actually # importing the un-built submodules. __PYRAMID_SETUP__ except NameError: __PYRAMID_SETUP__ = False if __PYRAMID_SETUP__: import sys import os sys.stderr.write('Partial import of pyramid during the build process.' + os.linesep) else: # check that the build completed properly. This prints an informative # message in the case that any of the C code was not properly compiled. from . import __check_build __all__ = [ 'arima', 'compat', 'datasets', 'utils' ] def setup_module(module): import numpy as np import random _random_seed = int(np.random.uniform() * (2 ** 31 - 1)) np.random.seed(_random_seed) random.seed(_random_seed)
Bump version for v0.7.0 release
Bump version for v0.7.0 release
Python
mit
tgsmith61591/pyramid,alkaline-ml/pmdarima,tgsmith61591/pyramid,tgsmith61591/pyramid,alkaline-ml/pmdarima,alkaline-ml/pmdarima
# -*- coding: utf-8 -*- # # Author: Taylor Smith <taylor.smith@alkaline-ml.com> # # The pyramid module __version__ = "0.7.0-dev" try: # this var is injected in the setup build to enable # the retrieval of the version number without actually # importing the un-built submodules. __PYRAMID_SETUP__ except NameError: __PYRAMID_SETUP__ = False if __PYRAMID_SETUP__: import sys import os sys.stderr.write('Partial import of pyramid during the build process.' + os.linesep) else: # check that the build completed properly. This prints an informative # message in the case that any of the C code was not properly compiled. from . import __check_build __all__ = [ 'arima', 'compat', 'datasets', 'utils' ] def setup_module(module): import numpy as np import random _random_seed = int(np.random.uniform() * (2 ** 31 - 1)) np.random.seed(_random_seed) random.seed(_random_seed) Bump version for v0.7.0 release
# -*- coding: utf-8 -*- # # Author: Taylor Smith <taylor.smith@alkaline-ml.com> # # The pyramid module __version__ = "0.7.0" try: # this var is injected in the setup build to enable # the retrieval of the version number without actually # importing the un-built submodules. __PYRAMID_SETUP__ except NameError: __PYRAMID_SETUP__ = False if __PYRAMID_SETUP__: import sys import os sys.stderr.write('Partial import of pyramid during the build process.' + os.linesep) else: # check that the build completed properly. This prints an informative # message in the case that any of the C code was not properly compiled. from . import __check_build __all__ = [ 'arima', 'compat', 'datasets', 'utils' ] def setup_module(module): import numpy as np import random _random_seed = int(np.random.uniform() * (2 ** 31 - 1)) np.random.seed(_random_seed) random.seed(_random_seed)
<commit_before># -*- coding: utf-8 -*- # # Author: Taylor Smith <taylor.smith@alkaline-ml.com> # # The pyramid module __version__ = "0.7.0-dev" try: # this var is injected in the setup build to enable # the retrieval of the version number without actually # importing the un-built submodules. __PYRAMID_SETUP__ except NameError: __PYRAMID_SETUP__ = False if __PYRAMID_SETUP__: import sys import os sys.stderr.write('Partial import of pyramid during the build process.' + os.linesep) else: # check that the build completed properly. This prints an informative # message in the case that any of the C code was not properly compiled. from . import __check_build __all__ = [ 'arima', 'compat', 'datasets', 'utils' ] def setup_module(module): import numpy as np import random _random_seed = int(np.random.uniform() * (2 ** 31 - 1)) np.random.seed(_random_seed) random.seed(_random_seed) <commit_msg>Bump version for v0.7.0 release<commit_after>
# -*- coding: utf-8 -*- # # Author: Taylor Smith <taylor.smith@alkaline-ml.com> # # The pyramid module __version__ = "0.7.0" try: # this var is injected in the setup build to enable # the retrieval of the version number without actually # importing the un-built submodules. __PYRAMID_SETUP__ except NameError: __PYRAMID_SETUP__ = False if __PYRAMID_SETUP__: import sys import os sys.stderr.write('Partial import of pyramid during the build process.' + os.linesep) else: # check that the build completed properly. This prints an informative # message in the case that any of the C code was not properly compiled. from . import __check_build __all__ = [ 'arima', 'compat', 'datasets', 'utils' ] def setup_module(module): import numpy as np import random _random_seed = int(np.random.uniform() * (2 ** 31 - 1)) np.random.seed(_random_seed) random.seed(_random_seed)
# -*- coding: utf-8 -*- # # Author: Taylor Smith <taylor.smith@alkaline-ml.com> # # The pyramid module __version__ = "0.7.0-dev" try: # this var is injected in the setup build to enable # the retrieval of the version number without actually # importing the un-built submodules. __PYRAMID_SETUP__ except NameError: __PYRAMID_SETUP__ = False if __PYRAMID_SETUP__: import sys import os sys.stderr.write('Partial import of pyramid during the build process.' + os.linesep) else: # check that the build completed properly. This prints an informative # message in the case that any of the C code was not properly compiled. from . import __check_build __all__ = [ 'arima', 'compat', 'datasets', 'utils' ] def setup_module(module): import numpy as np import random _random_seed = int(np.random.uniform() * (2 ** 31 - 1)) np.random.seed(_random_seed) random.seed(_random_seed) Bump version for v0.7.0 release# -*- coding: utf-8 -*- # # Author: Taylor Smith <taylor.smith@alkaline-ml.com> # # The pyramid module __version__ = "0.7.0" try: # this var is injected in the setup build to enable # the retrieval of the version number without actually # importing the un-built submodules. __PYRAMID_SETUP__ except NameError: __PYRAMID_SETUP__ = False if __PYRAMID_SETUP__: import sys import os sys.stderr.write('Partial import of pyramid during the build process.' + os.linesep) else: # check that the build completed properly. This prints an informative # message in the case that any of the C code was not properly compiled. from . import __check_build __all__ = [ 'arima', 'compat', 'datasets', 'utils' ] def setup_module(module): import numpy as np import random _random_seed = int(np.random.uniform() * (2 ** 31 - 1)) np.random.seed(_random_seed) random.seed(_random_seed)
<commit_before># -*- coding: utf-8 -*- # # Author: Taylor Smith <taylor.smith@alkaline-ml.com> # # The pyramid module __version__ = "0.7.0-dev" try: # this var is injected in the setup build to enable # the retrieval of the version number without actually # importing the un-built submodules. __PYRAMID_SETUP__ except NameError: __PYRAMID_SETUP__ = False if __PYRAMID_SETUP__: import sys import os sys.stderr.write('Partial import of pyramid during the build process.' + os.linesep) else: # check that the build completed properly. This prints an informative # message in the case that any of the C code was not properly compiled. from . import __check_build __all__ = [ 'arima', 'compat', 'datasets', 'utils' ] def setup_module(module): import numpy as np import random _random_seed = int(np.random.uniform() * (2 ** 31 - 1)) np.random.seed(_random_seed) random.seed(_random_seed) <commit_msg>Bump version for v0.7.0 release<commit_after># -*- coding: utf-8 -*- # # Author: Taylor Smith <taylor.smith@alkaline-ml.com> # # The pyramid module __version__ = "0.7.0" try: # this var is injected in the setup build to enable # the retrieval of the version number without actually # importing the un-built submodules. __PYRAMID_SETUP__ except NameError: __PYRAMID_SETUP__ = False if __PYRAMID_SETUP__: import sys import os sys.stderr.write('Partial import of pyramid during the build process.' + os.linesep) else: # check that the build completed properly. This prints an informative # message in the case that any of the C code was not properly compiled. from . import __check_build __all__ = [ 'arima', 'compat', 'datasets', 'utils' ] def setup_module(module): import numpy as np import random _random_seed = int(np.random.uniform() * (2 ** 31 - 1)) np.random.seed(_random_seed) random.seed(_random_seed)
194b72c7d3f0f60e73d1e11af716bfbdd9d4f08d
drupdates/plugins/slack/__init__.py
drupdates/plugins/slack/__init__.py
from drupdates.utils import * from drupdates.constructors.reports import * import json class slack(Reports): def __init__(self): self.currentDir = os.path.dirname(os.path.realpath(__file__)) self.settings = Settings(self.currentDir) def sendMessage(self, reportText): """ Post the report to a Slack channel or DM a specific user.""" url = self.settings.get('slackURL') user = self.settings.get('slackUser') payload = {} payload['text'] = reportText payload['new-bot-name'] = user dm = self.settings.get('slackRecipient') if dm: payload['channel'] = '@' + dm response = utils.apiCall(url, 'Slack', 'post', data = json.dumps(payload))
from drupdates.utils import * from drupdates.constructors.reports import * import json class slack(Reports): def __init__(self): self.currentDir = os.path.dirname(os.path.realpath(__file__)) self.settings = Settings(self.currentDir) def sendMessage(self, reportText): """ Post the report to a Slack channel or DM a specific user.""" url = self.settings.get('slackURL') user = self.settings.get('slackUser') payload = {} payload['text'] = reportText payload['new-bot-name'] = user dm = self.settings.get('slackRecipient') channel = self.settings.get('slackChannel') if dm: payload['channel'] = '@' + dm elif channel: payload['channel'] = '#' + dm response = utils.apiCall(url, 'Slack', 'post', data = json.dumps(payload))
Add Slack channels the Slack Plugin
Add Slack channels the Slack Plugin
Python
mit
jalama/drupdates
from drupdates.utils import * from drupdates.constructors.reports import * import json class slack(Reports): def __init__(self): self.currentDir = os.path.dirname(os.path.realpath(__file__)) self.settings = Settings(self.currentDir) def sendMessage(self, reportText): """ Post the report to a Slack channel or DM a specific user.""" url = self.settings.get('slackURL') user = self.settings.get('slackUser') payload = {} payload['text'] = reportText payload['new-bot-name'] = user dm = self.settings.get('slackRecipient') if dm: payload['channel'] = '@' + dm response = utils.apiCall(url, 'Slack', 'post', data = json.dumps(payload)) Add Slack channels the Slack Plugin
from drupdates.utils import * from drupdates.constructors.reports import * import json class slack(Reports): def __init__(self): self.currentDir = os.path.dirname(os.path.realpath(__file__)) self.settings = Settings(self.currentDir) def sendMessage(self, reportText): """ Post the report to a Slack channel or DM a specific user.""" url = self.settings.get('slackURL') user = self.settings.get('slackUser') payload = {} payload['text'] = reportText payload['new-bot-name'] = user dm = self.settings.get('slackRecipient') channel = self.settings.get('slackChannel') if dm: payload['channel'] = '@' + dm elif channel: payload['channel'] = '#' + dm response = utils.apiCall(url, 'Slack', 'post', data = json.dumps(payload))
<commit_before>from drupdates.utils import * from drupdates.constructors.reports import * import json class slack(Reports): def __init__(self): self.currentDir = os.path.dirname(os.path.realpath(__file__)) self.settings = Settings(self.currentDir) def sendMessage(self, reportText): """ Post the report to a Slack channel or DM a specific user.""" url = self.settings.get('slackURL') user = self.settings.get('slackUser') payload = {} payload['text'] = reportText payload['new-bot-name'] = user dm = self.settings.get('slackRecipient') if dm: payload['channel'] = '@' + dm response = utils.apiCall(url, 'Slack', 'post', data = json.dumps(payload)) <commit_msg>Add Slack channels the Slack Plugin<commit_after>
from drupdates.utils import * from drupdates.constructors.reports import * import json class slack(Reports): def __init__(self): self.currentDir = os.path.dirname(os.path.realpath(__file__)) self.settings = Settings(self.currentDir) def sendMessage(self, reportText): """ Post the report to a Slack channel or DM a specific user.""" url = self.settings.get('slackURL') user = self.settings.get('slackUser') payload = {} payload['text'] = reportText payload['new-bot-name'] = user dm = self.settings.get('slackRecipient') channel = self.settings.get('slackChannel') if dm: payload['channel'] = '@' + dm elif channel: payload['channel'] = '#' + dm response = utils.apiCall(url, 'Slack', 'post', data = json.dumps(payload))
from drupdates.utils import * from drupdates.constructors.reports import * import json class slack(Reports): def __init__(self): self.currentDir = os.path.dirname(os.path.realpath(__file__)) self.settings = Settings(self.currentDir) def sendMessage(self, reportText): """ Post the report to a Slack channel or DM a specific user.""" url = self.settings.get('slackURL') user = self.settings.get('slackUser') payload = {} payload['text'] = reportText payload['new-bot-name'] = user dm = self.settings.get('slackRecipient') if dm: payload['channel'] = '@' + dm response = utils.apiCall(url, 'Slack', 'post', data = json.dumps(payload)) Add Slack channels the Slack Pluginfrom drupdates.utils import * from drupdates.constructors.reports import * import json class slack(Reports): def __init__(self): self.currentDir = os.path.dirname(os.path.realpath(__file__)) self.settings = Settings(self.currentDir) def sendMessage(self, reportText): """ Post the report to a Slack channel or DM a specific user.""" url = self.settings.get('slackURL') user = self.settings.get('slackUser') payload = {} payload['text'] = reportText payload['new-bot-name'] = user dm = self.settings.get('slackRecipient') channel = self.settings.get('slackChannel') if dm: payload['channel'] = '@' + dm elif channel: payload['channel'] = '#' + dm response = utils.apiCall(url, 'Slack', 'post', data = json.dumps(payload))
<commit_before>from drupdates.utils import * from drupdates.constructors.reports import * import json class slack(Reports): def __init__(self): self.currentDir = os.path.dirname(os.path.realpath(__file__)) self.settings = Settings(self.currentDir) def sendMessage(self, reportText): """ Post the report to a Slack channel or DM a specific user.""" url = self.settings.get('slackURL') user = self.settings.get('slackUser') payload = {} payload['text'] = reportText payload['new-bot-name'] = user dm = self.settings.get('slackRecipient') if dm: payload['channel'] = '@' + dm response = utils.apiCall(url, 'Slack', 'post', data = json.dumps(payload)) <commit_msg>Add Slack channels the Slack Plugin<commit_after>from drupdates.utils import * from drupdates.constructors.reports import * import json class slack(Reports): def __init__(self): self.currentDir = os.path.dirname(os.path.realpath(__file__)) self.settings = Settings(self.currentDir) def sendMessage(self, reportText): """ Post the report to a Slack channel or DM a specific user.""" url = self.settings.get('slackURL') user = self.settings.get('slackUser') payload = {} payload['text'] = reportText payload['new-bot-name'] = user dm = self.settings.get('slackRecipient') channel = self.settings.get('slackChannel') if dm: payload['channel'] = '@' + dm elif channel: payload['channel'] = '#' + dm response = utils.apiCall(url, 'Slack', 'post', data = json.dumps(payload))
988f4655a96076acd3bfb906d240bd4601fbe535
getalltext.py
getalltext.py
#!/usr/bin/env python3 """ A program to extract raw text from Telegram chat log """ import argparse from json import loads def main(): parser = argparse.ArgumentParser( description="Extract all raw text from a specific Telegram chat") parser.add_argument('filepath', help='the json chatlog file to analyse') args=parser.parse_args() filepath = args.filepath with open(filepath, 'r') as jsonfile: events = (loads(line) for line in jsonfile) for event in events: #check the event is the sort we're looking for if "from" in event and "text" in event: #do i need the "from" here? print(event["text"]) if __name__ == "__main__": main()
#!/usr/bin/env python3 """ A program to extract raw text from Telegram chat log """ import argparse from json import loads def main(): parser = argparse.ArgumentParser( description="Extract all raw text from a specific Telegram chat") parser.add_argument('filepath', help='the json chatlog file to analyse') parser.add_argument('-u','--usernames', help='Show usernames before messages',action='store_true') args=parser.parse_args() filepath = args.filepath with open(filepath, 'r') as jsonfile: events = (loads(line) for line in jsonfile) for event in events: #check the event is the sort we're looking for if "from" in event and "text" in event: if args.usernames: print(event['from']['username'],end=': ') #do i need the "from" here? print(event["text"]) if __name__ == "__main__": main()
Add argument to print usernames
Add argument to print usernames
Python
mit
expectocode/telegram-analysis,expectocode/telegramAnalysis
#!/usr/bin/env python3 """ A program to extract raw text from Telegram chat log """ import argparse from json import loads def main(): parser = argparse.ArgumentParser( description="Extract all raw text from a specific Telegram chat") parser.add_argument('filepath', help='the json chatlog file to analyse') args=parser.parse_args() filepath = args.filepath with open(filepath, 'r') as jsonfile: events = (loads(line) for line in jsonfile) for event in events: #check the event is the sort we're looking for if "from" in event and "text" in event: #do i need the "from" here? print(event["text"]) if __name__ == "__main__": main() Add argument to print usernames
#!/usr/bin/env python3 """ A program to extract raw text from Telegram chat log """ import argparse from json import loads def main(): parser = argparse.ArgumentParser( description="Extract all raw text from a specific Telegram chat") parser.add_argument('filepath', help='the json chatlog file to analyse') parser.add_argument('-u','--usernames', help='Show usernames before messages',action='store_true') args=parser.parse_args() filepath = args.filepath with open(filepath, 'r') as jsonfile: events = (loads(line) for line in jsonfile) for event in events: #check the event is the sort we're looking for if "from" in event and "text" in event: if args.usernames: print(event['from']['username'],end=': ') #do i need the "from" here? print(event["text"]) if __name__ == "__main__": main()
<commit_before>#!/usr/bin/env python3 """ A program to extract raw text from Telegram chat log """ import argparse from json import loads def main(): parser = argparse.ArgumentParser( description="Extract all raw text from a specific Telegram chat") parser.add_argument('filepath', help='the json chatlog file to analyse') args=parser.parse_args() filepath = args.filepath with open(filepath, 'r') as jsonfile: events = (loads(line) for line in jsonfile) for event in events: #check the event is the sort we're looking for if "from" in event and "text" in event: #do i need the "from" here? print(event["text"]) if __name__ == "__main__": main() <commit_msg>Add argument to print usernames<commit_after>
#!/usr/bin/env python3 """ A program to extract raw text from Telegram chat log """ import argparse from json import loads def main(): parser = argparse.ArgumentParser( description="Extract all raw text from a specific Telegram chat") parser.add_argument('filepath', help='the json chatlog file to analyse') parser.add_argument('-u','--usernames', help='Show usernames before messages',action='store_true') args=parser.parse_args() filepath = args.filepath with open(filepath, 'r') as jsonfile: events = (loads(line) for line in jsonfile) for event in events: #check the event is the sort we're looking for if "from" in event and "text" in event: if args.usernames: print(event['from']['username'],end=': ') #do i need the "from" here? print(event["text"]) if __name__ == "__main__": main()
#!/usr/bin/env python3 """ A program to extract raw text from Telegram chat log """ import argparse from json import loads def main(): parser = argparse.ArgumentParser( description="Extract all raw text from a specific Telegram chat") parser.add_argument('filepath', help='the json chatlog file to analyse') args=parser.parse_args() filepath = args.filepath with open(filepath, 'r') as jsonfile: events = (loads(line) for line in jsonfile) for event in events: #check the event is the sort we're looking for if "from" in event and "text" in event: #do i need the "from" here? print(event["text"]) if __name__ == "__main__": main() Add argument to print usernames#!/usr/bin/env python3 """ A program to extract raw text from Telegram chat log """ import argparse from json import loads def main(): parser = argparse.ArgumentParser( description="Extract all raw text from a specific Telegram chat") parser.add_argument('filepath', help='the json chatlog file to analyse') parser.add_argument('-u','--usernames', help='Show usernames before messages',action='store_true') args=parser.parse_args() filepath = args.filepath with open(filepath, 'r') as jsonfile: events = (loads(line) for line in jsonfile) for event in events: #check the event is the sort we're looking for if "from" in event and "text" in event: if args.usernames: print(event['from']['username'],end=': ') #do i need the "from" here? print(event["text"]) if __name__ == "__main__": main()
<commit_before>#!/usr/bin/env python3 """ A program to extract raw text from Telegram chat log """ import argparse from json import loads def main(): parser = argparse.ArgumentParser( description="Extract all raw text from a specific Telegram chat") parser.add_argument('filepath', help='the json chatlog file to analyse') args=parser.parse_args() filepath = args.filepath with open(filepath, 'r') as jsonfile: events = (loads(line) for line in jsonfile) for event in events: #check the event is the sort we're looking for if "from" in event and "text" in event: #do i need the "from" here? print(event["text"]) if __name__ == "__main__": main() <commit_msg>Add argument to print usernames<commit_after>#!/usr/bin/env python3 """ A program to extract raw text from Telegram chat log """ import argparse from json import loads def main(): parser = argparse.ArgumentParser( description="Extract all raw text from a specific Telegram chat") parser.add_argument('filepath', help='the json chatlog file to analyse') parser.add_argument('-u','--usernames', help='Show usernames before messages',action='store_true') args=parser.parse_args() filepath = args.filepath with open(filepath, 'r') as jsonfile: events = (loads(line) for line in jsonfile) for event in events: #check the event is the sort we're looking for if "from" in event and "text" in event: if args.usernames: print(event['from']['username'],end=': ') #do i need the "from" here? print(event["text"]) if __name__ == "__main__": main()
9056746db7406e6640607210bea9e00a12c63926
ci/fix_paths.py
ci/fix_paths.py
import distutils.sysconfig from glob import glob import os from os.path import join as pjoin, basename from shutil import copy from sys import platform def main(): """ Copy HDF5 DLLs into installed h5py package """ # This is the function Tox also uses to locate site-packages (Apr 2019) sitepackagesdir = distutils.sysconfig.get_python_lib(plat_specific=True) print("site packages dir:", sitepackagesdir) hdf5_path = os.environ.get("HDF5_DIR") print("HDF5_DIR", hdf5_path) # HDF5_DIR is not set when we're testing wheels; these should already have # the necessary libraries bundled in. if platform.startswith('win') and hdf5_path is not None: for f in glob(pjoin(hdf5_path, 'lib/*.dll')): copy(f, pjoin(sitepackagesdir, 'h5py', basename(f))) print("Copied", f) zlib_root = os.environ.get("ZLIB_ROOT") if zlib_root: f = pjoin(zlib_root, 'bin_release', 'zlib.dll') copy(f, pjoin(sitepackagesdir, 'h5py', 'zlib.dll')) print("Copied", f) print("In installed h5py:", os.listdir(pjoin(sitepackagesdir, 'h5py'))) if __name__ == '__main__': main()
import distutils.sysconfig from glob import glob import os from os.path import join as pjoin, basename from shutil import copy from sys import platform def main(): """ Copy HDF5 DLLs into installed h5py package """ # This is the function Tox also uses to locate site-packages (Apr 2019) sitepackagesdir = distutils.sysconfig.get_python_lib(plat_specific=True) print("site packages dir:", sitepackagesdir) hdf5_path = os.environ.get("HDF5_DIR") print("HDF5_DIR", hdf5_path) # HDF5_DIR is not set when we're testing wheels; these should already have # the necessary libraries bundled in. if platform.startswith('win') and hdf5_path is not None: for f in glob(pjoin(hdf5_path, 'lib/*.dll')): copy(f, pjoin(sitepackagesdir, 'h5py', basename(f))) print("Copied", f) zlib_root = os.environ.get("ZLIB_ROOT") if zlib_root: f = pjoin(zlib_root, 'bin_release', 'zlib.dll') copy(f, pjoin(sitepackagesdir, 'h5py', 'zlib.dll')) print("Copied", f) print("In installed h5py:", sorted(os.listdir(pjoin(sitepackagesdir, 'h5py')))) if __name__ == '__main__': main()
Sort list of files inside h5py
Sort list of files inside h5py
Python
bsd-3-clause
h5py/h5py,h5py/h5py,h5py/h5py
import distutils.sysconfig from glob import glob import os from os.path import join as pjoin, basename from shutil import copy from sys import platform def main(): """ Copy HDF5 DLLs into installed h5py package """ # This is the function Tox also uses to locate site-packages (Apr 2019) sitepackagesdir = distutils.sysconfig.get_python_lib(plat_specific=True) print("site packages dir:", sitepackagesdir) hdf5_path = os.environ.get("HDF5_DIR") print("HDF5_DIR", hdf5_path) # HDF5_DIR is not set when we're testing wheels; these should already have # the necessary libraries bundled in. if platform.startswith('win') and hdf5_path is not None: for f in glob(pjoin(hdf5_path, 'lib/*.dll')): copy(f, pjoin(sitepackagesdir, 'h5py', basename(f))) print("Copied", f) zlib_root = os.environ.get("ZLIB_ROOT") if zlib_root: f = pjoin(zlib_root, 'bin_release', 'zlib.dll') copy(f, pjoin(sitepackagesdir, 'h5py', 'zlib.dll')) print("Copied", f) print("In installed h5py:", os.listdir(pjoin(sitepackagesdir, 'h5py'))) if __name__ == '__main__': main() Sort list of files inside h5py
import distutils.sysconfig from glob import glob import os from os.path import join as pjoin, basename from shutil import copy from sys import platform def main(): """ Copy HDF5 DLLs into installed h5py package """ # This is the function Tox also uses to locate site-packages (Apr 2019) sitepackagesdir = distutils.sysconfig.get_python_lib(plat_specific=True) print("site packages dir:", sitepackagesdir) hdf5_path = os.environ.get("HDF5_DIR") print("HDF5_DIR", hdf5_path) # HDF5_DIR is not set when we're testing wheels; these should already have # the necessary libraries bundled in. if platform.startswith('win') and hdf5_path is not None: for f in glob(pjoin(hdf5_path, 'lib/*.dll')): copy(f, pjoin(sitepackagesdir, 'h5py', basename(f))) print("Copied", f) zlib_root = os.environ.get("ZLIB_ROOT") if zlib_root: f = pjoin(zlib_root, 'bin_release', 'zlib.dll') copy(f, pjoin(sitepackagesdir, 'h5py', 'zlib.dll')) print("Copied", f) print("In installed h5py:", sorted(os.listdir(pjoin(sitepackagesdir, 'h5py')))) if __name__ == '__main__': main()
<commit_before>import distutils.sysconfig from glob import glob import os from os.path import join as pjoin, basename from shutil import copy from sys import platform def main(): """ Copy HDF5 DLLs into installed h5py package """ # This is the function Tox also uses to locate site-packages (Apr 2019) sitepackagesdir = distutils.sysconfig.get_python_lib(plat_specific=True) print("site packages dir:", sitepackagesdir) hdf5_path = os.environ.get("HDF5_DIR") print("HDF5_DIR", hdf5_path) # HDF5_DIR is not set when we're testing wheels; these should already have # the necessary libraries bundled in. if platform.startswith('win') and hdf5_path is not None: for f in glob(pjoin(hdf5_path, 'lib/*.dll')): copy(f, pjoin(sitepackagesdir, 'h5py', basename(f))) print("Copied", f) zlib_root = os.environ.get("ZLIB_ROOT") if zlib_root: f = pjoin(zlib_root, 'bin_release', 'zlib.dll') copy(f, pjoin(sitepackagesdir, 'h5py', 'zlib.dll')) print("Copied", f) print("In installed h5py:", os.listdir(pjoin(sitepackagesdir, 'h5py'))) if __name__ == '__main__': main() <commit_msg>Sort list of files inside h5py<commit_after>
import distutils.sysconfig from glob import glob import os from os.path import join as pjoin, basename from shutil import copy from sys import platform def main(): """ Copy HDF5 DLLs into installed h5py package """ # This is the function Tox also uses to locate site-packages (Apr 2019) sitepackagesdir = distutils.sysconfig.get_python_lib(plat_specific=True) print("site packages dir:", sitepackagesdir) hdf5_path = os.environ.get("HDF5_DIR") print("HDF5_DIR", hdf5_path) # HDF5_DIR is not set when we're testing wheels; these should already have # the necessary libraries bundled in. if platform.startswith('win') and hdf5_path is not None: for f in glob(pjoin(hdf5_path, 'lib/*.dll')): copy(f, pjoin(sitepackagesdir, 'h5py', basename(f))) print("Copied", f) zlib_root = os.environ.get("ZLIB_ROOT") if zlib_root: f = pjoin(zlib_root, 'bin_release', 'zlib.dll') copy(f, pjoin(sitepackagesdir, 'h5py', 'zlib.dll')) print("Copied", f) print("In installed h5py:", sorted(os.listdir(pjoin(sitepackagesdir, 'h5py')))) if __name__ == '__main__': main()
import distutils.sysconfig from glob import glob import os from os.path import join as pjoin, basename from shutil import copy from sys import platform def main(): """ Copy HDF5 DLLs into installed h5py package """ # This is the function Tox also uses to locate site-packages (Apr 2019) sitepackagesdir = distutils.sysconfig.get_python_lib(plat_specific=True) print("site packages dir:", sitepackagesdir) hdf5_path = os.environ.get("HDF5_DIR") print("HDF5_DIR", hdf5_path) # HDF5_DIR is not set when we're testing wheels; these should already have # the necessary libraries bundled in. if platform.startswith('win') and hdf5_path is not None: for f in glob(pjoin(hdf5_path, 'lib/*.dll')): copy(f, pjoin(sitepackagesdir, 'h5py', basename(f))) print("Copied", f) zlib_root = os.environ.get("ZLIB_ROOT") if zlib_root: f = pjoin(zlib_root, 'bin_release', 'zlib.dll') copy(f, pjoin(sitepackagesdir, 'h5py', 'zlib.dll')) print("Copied", f) print("In installed h5py:", os.listdir(pjoin(sitepackagesdir, 'h5py'))) if __name__ == '__main__': main() Sort list of files inside h5pyimport distutils.sysconfig from glob import glob import os from os.path import join as pjoin, basename from shutil import copy from sys import platform def main(): """ Copy HDF5 DLLs into installed h5py package """ # This is the function Tox also uses to locate site-packages (Apr 2019) sitepackagesdir = distutils.sysconfig.get_python_lib(plat_specific=True) print("site packages dir:", sitepackagesdir) hdf5_path = os.environ.get("HDF5_DIR") print("HDF5_DIR", hdf5_path) # HDF5_DIR is not set when we're testing wheels; these should already have # the necessary libraries bundled in. if platform.startswith('win') and hdf5_path is not None: for f in glob(pjoin(hdf5_path, 'lib/*.dll')): copy(f, pjoin(sitepackagesdir, 'h5py', basename(f))) print("Copied", f) zlib_root = os.environ.get("ZLIB_ROOT") if zlib_root: f = pjoin(zlib_root, 'bin_release', 'zlib.dll') copy(f, pjoin(sitepackagesdir, 'h5py', 'zlib.dll')) print("Copied", f) print("In installed h5py:", sorted(os.listdir(pjoin(sitepackagesdir, 'h5py')))) if __name__ == '__main__': main()
<commit_before>import distutils.sysconfig from glob import glob import os from os.path import join as pjoin, basename from shutil import copy from sys import platform def main(): """ Copy HDF5 DLLs into installed h5py package """ # This is the function Tox also uses to locate site-packages (Apr 2019) sitepackagesdir = distutils.sysconfig.get_python_lib(plat_specific=True) print("site packages dir:", sitepackagesdir) hdf5_path = os.environ.get("HDF5_DIR") print("HDF5_DIR", hdf5_path) # HDF5_DIR is not set when we're testing wheels; these should already have # the necessary libraries bundled in. if platform.startswith('win') and hdf5_path is not None: for f in glob(pjoin(hdf5_path, 'lib/*.dll')): copy(f, pjoin(sitepackagesdir, 'h5py', basename(f))) print("Copied", f) zlib_root = os.environ.get("ZLIB_ROOT") if zlib_root: f = pjoin(zlib_root, 'bin_release', 'zlib.dll') copy(f, pjoin(sitepackagesdir, 'h5py', 'zlib.dll')) print("Copied", f) print("In installed h5py:", os.listdir(pjoin(sitepackagesdir, 'h5py'))) if __name__ == '__main__': main() <commit_msg>Sort list of files inside h5py<commit_after>import distutils.sysconfig from glob import glob import os from os.path import join as pjoin, basename from shutil import copy from sys import platform def main(): """ Copy HDF5 DLLs into installed h5py package """ # This is the function Tox also uses to locate site-packages (Apr 2019) sitepackagesdir = distutils.sysconfig.get_python_lib(plat_specific=True) print("site packages dir:", sitepackagesdir) hdf5_path = os.environ.get("HDF5_DIR") print("HDF5_DIR", hdf5_path) # HDF5_DIR is not set when we're testing wheels; these should already have # the necessary libraries bundled in. if platform.startswith('win') and hdf5_path is not None: for f in glob(pjoin(hdf5_path, 'lib/*.dll')): copy(f, pjoin(sitepackagesdir, 'h5py', basename(f))) print("Copied", f) zlib_root = os.environ.get("ZLIB_ROOT") if zlib_root: f = pjoin(zlib_root, 'bin_release', 'zlib.dll') copy(f, pjoin(sitepackagesdir, 'h5py', 'zlib.dll')) print("Copied", f) print("In installed h5py:", sorted(os.listdir(pjoin(sitepackagesdir, 'h5py')))) if __name__ == '__main__': main()
cd8d7235190b4040fd135df6bb45e984c341b568
cities/admin.py
cities/admin.py
from django.contrib.gis import admin from cities.models import * class CityAdmin(admin.OSMGeoAdmin): list_display = ('__unicode__', 'population_2000') list_filter = ('pop_range',) search_fields = ('name',) admin.site.register(City, CityAdmin)
from django.contrib.gis import admin from cities.models import * class CityAdmin(admin.OSMGeoAdmin): list_display = ('__unicode__', 'population_2000') list_filter = ('pop_range', 'state') search_fields = ('name',) admin.site.register(City, CityAdmin)
Allow filtering by state code
Allow filtering by state code
Python
bsd-3-clause
adamfast/usgsdata-citiesx020
from django.contrib.gis import admin from cities.models import * class CityAdmin(admin.OSMGeoAdmin): list_display = ('__unicode__', 'population_2000') list_filter = ('pop_range',) search_fields = ('name',) admin.site.register(City, CityAdmin) Allow filtering by state code
from django.contrib.gis import admin from cities.models import * class CityAdmin(admin.OSMGeoAdmin): list_display = ('__unicode__', 'population_2000') list_filter = ('pop_range', 'state') search_fields = ('name',) admin.site.register(City, CityAdmin)
<commit_before>from django.contrib.gis import admin from cities.models import * class CityAdmin(admin.OSMGeoAdmin): list_display = ('__unicode__', 'population_2000') list_filter = ('pop_range',) search_fields = ('name',) admin.site.register(City, CityAdmin) <commit_msg>Allow filtering by state code<commit_after>
from django.contrib.gis import admin from cities.models import * class CityAdmin(admin.OSMGeoAdmin): list_display = ('__unicode__', 'population_2000') list_filter = ('pop_range', 'state') search_fields = ('name',) admin.site.register(City, CityAdmin)
from django.contrib.gis import admin from cities.models import * class CityAdmin(admin.OSMGeoAdmin): list_display = ('__unicode__', 'population_2000') list_filter = ('pop_range',) search_fields = ('name',) admin.site.register(City, CityAdmin) Allow filtering by state codefrom django.contrib.gis import admin from cities.models import * class CityAdmin(admin.OSMGeoAdmin): list_display = ('__unicode__', 'population_2000') list_filter = ('pop_range', 'state') search_fields = ('name',) admin.site.register(City, CityAdmin)
<commit_before>from django.contrib.gis import admin from cities.models import * class CityAdmin(admin.OSMGeoAdmin): list_display = ('__unicode__', 'population_2000') list_filter = ('pop_range',) search_fields = ('name',) admin.site.register(City, CityAdmin) <commit_msg>Allow filtering by state code<commit_after>from django.contrib.gis import admin from cities.models import * class CityAdmin(admin.OSMGeoAdmin): list_display = ('__unicode__', 'population_2000') list_filter = ('pop_range', 'state') search_fields = ('name',) admin.site.register(City, CityAdmin)
743f5a72840a5f829720899fe0febcdd7466be3e
bika/lims/upgrade/to1101.py
bika/lims/upgrade/to1101.py
import logging from Acquisition import aq_base from Acquisition import aq_inner from Acquisition import aq_parent from Products.CMFCore import permissions from bika.lims.permissions import * from Products.CMFCore.utils import getToolByName def upgrade(tool): """ issue #615: missing configuration for some add permissions """ portal = aq_parent(aq_inner(tool)) # batch permission defaults mp = portal.manage_permission mp(AddAnalysisSpec, ['Manager', 'Owner', 'LabManager', 'LabClerk'], 1) mp(AddSamplingDeviation, ['Manager', 'Owner', 'LabManager', 'LabClerk'], 1) mp(AddSamplingMatrix, ['Manager', 'Owner', 'LabManager', 'LabClerk'], 1) portal.reindexObject() return True
import logging from Acquisition import aq_base from Acquisition import aq_inner from Acquisition import aq_parent from Products.CMFCore import permissions from bika.lims.permissions import * from Products.CMFCore.utils import getToolByName def upgrade(tool): """ issue #615: missing configuration for some add permissions """ portal = aq_parent(aq_inner(tool)) # batch permission defaults mp = portal.manage_permission mp(AddAnalysisSpec, ['Manager', 'Owner', 'LabManager', 'LabClerk'], 1) mp(AddSamplingDeviation, ['Manager', 'Owner', 'LabManager', 'LabClerk'], 1) mp(AddSampleMatrix, ['Manager', 'Owner', 'LabManager', 'LabClerk'], 1) portal.reindexObject() return True
Fix permission name in upgrade-1101
Fix permission name in upgrade-1101
Python
agpl-3.0
rockfruit/bika.lims,veroc/Bika-LIMS,labsanmartin/Bika-LIMS,anneline/Bika-LIMS,labsanmartin/Bika-LIMS,DeBortoliWines/Bika-LIMS,labsanmartin/Bika-LIMS,anneline/Bika-LIMS,veroc/Bika-LIMS,anneline/Bika-LIMS,DeBortoliWines/Bika-LIMS,rockfruit/bika.lims,DeBortoliWines/Bika-LIMS,veroc/Bika-LIMS
import logging from Acquisition import aq_base from Acquisition import aq_inner from Acquisition import aq_parent from Products.CMFCore import permissions from bika.lims.permissions import * from Products.CMFCore.utils import getToolByName def upgrade(tool): """ issue #615: missing configuration for some add permissions """ portal = aq_parent(aq_inner(tool)) # batch permission defaults mp = portal.manage_permission mp(AddAnalysisSpec, ['Manager', 'Owner', 'LabManager', 'LabClerk'], 1) mp(AddSamplingDeviation, ['Manager', 'Owner', 'LabManager', 'LabClerk'], 1) mp(AddSamplingMatrix, ['Manager', 'Owner', 'LabManager', 'LabClerk'], 1) portal.reindexObject() return True Fix permission name in upgrade-1101
import logging from Acquisition import aq_base from Acquisition import aq_inner from Acquisition import aq_parent from Products.CMFCore import permissions from bika.lims.permissions import * from Products.CMFCore.utils import getToolByName def upgrade(tool): """ issue #615: missing configuration for some add permissions """ portal = aq_parent(aq_inner(tool)) # batch permission defaults mp = portal.manage_permission mp(AddAnalysisSpec, ['Manager', 'Owner', 'LabManager', 'LabClerk'], 1) mp(AddSamplingDeviation, ['Manager', 'Owner', 'LabManager', 'LabClerk'], 1) mp(AddSampleMatrix, ['Manager', 'Owner', 'LabManager', 'LabClerk'], 1) portal.reindexObject() return True
<commit_before>import logging from Acquisition import aq_base from Acquisition import aq_inner from Acquisition import aq_parent from Products.CMFCore import permissions from bika.lims.permissions import * from Products.CMFCore.utils import getToolByName def upgrade(tool): """ issue #615: missing configuration for some add permissions """ portal = aq_parent(aq_inner(tool)) # batch permission defaults mp = portal.manage_permission mp(AddAnalysisSpec, ['Manager', 'Owner', 'LabManager', 'LabClerk'], 1) mp(AddSamplingDeviation, ['Manager', 'Owner', 'LabManager', 'LabClerk'], 1) mp(AddSamplingMatrix, ['Manager', 'Owner', 'LabManager', 'LabClerk'], 1) portal.reindexObject() return True <commit_msg>Fix permission name in upgrade-1101<commit_after>
import logging from Acquisition import aq_base from Acquisition import aq_inner from Acquisition import aq_parent from Products.CMFCore import permissions from bika.lims.permissions import * from Products.CMFCore.utils import getToolByName def upgrade(tool): """ issue #615: missing configuration for some add permissions """ portal = aq_parent(aq_inner(tool)) # batch permission defaults mp = portal.manage_permission mp(AddAnalysisSpec, ['Manager', 'Owner', 'LabManager', 'LabClerk'], 1) mp(AddSamplingDeviation, ['Manager', 'Owner', 'LabManager', 'LabClerk'], 1) mp(AddSampleMatrix, ['Manager', 'Owner', 'LabManager', 'LabClerk'], 1) portal.reindexObject() return True
import logging from Acquisition import aq_base from Acquisition import aq_inner from Acquisition import aq_parent from Products.CMFCore import permissions from bika.lims.permissions import * from Products.CMFCore.utils import getToolByName def upgrade(tool): """ issue #615: missing configuration for some add permissions """ portal = aq_parent(aq_inner(tool)) # batch permission defaults mp = portal.manage_permission mp(AddAnalysisSpec, ['Manager', 'Owner', 'LabManager', 'LabClerk'], 1) mp(AddSamplingDeviation, ['Manager', 'Owner', 'LabManager', 'LabClerk'], 1) mp(AddSamplingMatrix, ['Manager', 'Owner', 'LabManager', 'LabClerk'], 1) portal.reindexObject() return True Fix permission name in upgrade-1101import logging from Acquisition import aq_base from Acquisition import aq_inner from Acquisition import aq_parent from Products.CMFCore import permissions from bika.lims.permissions import * from Products.CMFCore.utils import getToolByName def upgrade(tool): """ issue #615: missing configuration for some add permissions """ portal = aq_parent(aq_inner(tool)) # batch permission defaults mp = portal.manage_permission mp(AddAnalysisSpec, ['Manager', 'Owner', 'LabManager', 'LabClerk'], 1) mp(AddSamplingDeviation, ['Manager', 'Owner', 'LabManager', 'LabClerk'], 1) mp(AddSampleMatrix, ['Manager', 'Owner', 'LabManager', 'LabClerk'], 1) portal.reindexObject() return True
<commit_before>import logging from Acquisition import aq_base from Acquisition import aq_inner from Acquisition import aq_parent from Products.CMFCore import permissions from bika.lims.permissions import * from Products.CMFCore.utils import getToolByName def upgrade(tool): """ issue #615: missing configuration for some add permissions """ portal = aq_parent(aq_inner(tool)) # batch permission defaults mp = portal.manage_permission mp(AddAnalysisSpec, ['Manager', 'Owner', 'LabManager', 'LabClerk'], 1) mp(AddSamplingDeviation, ['Manager', 'Owner', 'LabManager', 'LabClerk'], 1) mp(AddSamplingMatrix, ['Manager', 'Owner', 'LabManager', 'LabClerk'], 1) portal.reindexObject() return True <commit_msg>Fix permission name in upgrade-1101<commit_after>import logging from Acquisition import aq_base from Acquisition import aq_inner from Acquisition import aq_parent from Products.CMFCore import permissions from bika.lims.permissions import * from Products.CMFCore.utils import getToolByName def upgrade(tool): """ issue #615: missing configuration for some add permissions """ portal = aq_parent(aq_inner(tool)) # batch permission defaults mp = portal.manage_permission mp(AddAnalysisSpec, ['Manager', 'Owner', 'LabManager', 'LabClerk'], 1) mp(AddSamplingDeviation, ['Manager', 'Owner', 'LabManager', 'LabClerk'], 1) mp(AddSampleMatrix, ['Manager', 'Owner', 'LabManager', 'LabClerk'], 1) portal.reindexObject() return True
3b091fba819f1ad69d0ce9e9038ccf5d14fea215
tests/core/tests/test_mixins.py
tests/core/tests/test_mixins.py
from core.models import Category from django.test.testcases import TestCase from django.urls import reverse class ExportViewMixinTest(TestCase): def setUp(self): self.url = reverse('export-category') self.cat1 = Category.objects.create(name='Cat 1') self.cat2 = Category.objects.create(name='Cat 2') def test_get(self): response = self.client.get(self.url) self.assertContains(response, self.cat1.name, status_code=200) self.assertTrue(response['Content-Type'], 'text/html') def test_post(self): data = { 'file_format': '0', } response = self.client.post(self.url, data) self.assertContains(response, self.cat1.name, status_code=200) self.assertTrue(response.has_header("Content-Disposition")) self.assertTrue(response['Content-Type'], 'text/csv')
from core.models import Category from django.test.testcases import TestCase from django.urls import reverse class ExportViewMixinTest(TestCase): def setUp(self): self.url = reverse('export-category') self.cat1 = Category.objects.create(name='Cat 1') self.cat2 = Category.objects.create(name='Cat 2') def test_get(self): response = self.client.get(self.url) self.assertContains(response, self.cat1.name, status_code=200) self.assertEquals(response['Content-Type'], 'text/html; charset=utf-8') def test_post(self): data = { 'file_format': '0', } response = self.client.post(self.url, data) self.assertContains(response, self.cat1.name, status_code=200) self.assertTrue(response.has_header("Content-Disposition")) self.assertEquals(response['Content-Type'], 'text/csv')
Correct mistaken assertTrue() -> assertEquals()
Correct mistaken assertTrue() -> assertEquals()
Python
bsd-2-clause
bmihelac/django-import-export,bmihelac/django-import-export,bmihelac/django-import-export,jnns/django-import-export,jnns/django-import-export,jnns/django-import-export,django-import-export/django-import-export,django-import-export/django-import-export,django-import-export/django-import-export,django-import-export/django-import-export,jnns/django-import-export,bmihelac/django-import-export
from core.models import Category from django.test.testcases import TestCase from django.urls import reverse class ExportViewMixinTest(TestCase): def setUp(self): self.url = reverse('export-category') self.cat1 = Category.objects.create(name='Cat 1') self.cat2 = Category.objects.create(name='Cat 2') def test_get(self): response = self.client.get(self.url) self.assertContains(response, self.cat1.name, status_code=200) self.assertTrue(response['Content-Type'], 'text/html') def test_post(self): data = { 'file_format': '0', } response = self.client.post(self.url, data) self.assertContains(response, self.cat1.name, status_code=200) self.assertTrue(response.has_header("Content-Disposition")) self.assertTrue(response['Content-Type'], 'text/csv') Correct mistaken assertTrue() -> assertEquals()
from core.models import Category from django.test.testcases import TestCase from django.urls import reverse class ExportViewMixinTest(TestCase): def setUp(self): self.url = reverse('export-category') self.cat1 = Category.objects.create(name='Cat 1') self.cat2 = Category.objects.create(name='Cat 2') def test_get(self): response = self.client.get(self.url) self.assertContains(response, self.cat1.name, status_code=200) self.assertEquals(response['Content-Type'], 'text/html; charset=utf-8') def test_post(self): data = { 'file_format': '0', } response = self.client.post(self.url, data) self.assertContains(response, self.cat1.name, status_code=200) self.assertTrue(response.has_header("Content-Disposition")) self.assertEquals(response['Content-Type'], 'text/csv')
<commit_before>from core.models import Category from django.test.testcases import TestCase from django.urls import reverse class ExportViewMixinTest(TestCase): def setUp(self): self.url = reverse('export-category') self.cat1 = Category.objects.create(name='Cat 1') self.cat2 = Category.objects.create(name='Cat 2') def test_get(self): response = self.client.get(self.url) self.assertContains(response, self.cat1.name, status_code=200) self.assertTrue(response['Content-Type'], 'text/html') def test_post(self): data = { 'file_format': '0', } response = self.client.post(self.url, data) self.assertContains(response, self.cat1.name, status_code=200) self.assertTrue(response.has_header("Content-Disposition")) self.assertTrue(response['Content-Type'], 'text/csv') <commit_msg>Correct mistaken assertTrue() -> assertEquals()<commit_after>
from core.models import Category from django.test.testcases import TestCase from django.urls import reverse class ExportViewMixinTest(TestCase): def setUp(self): self.url = reverse('export-category') self.cat1 = Category.objects.create(name='Cat 1') self.cat2 = Category.objects.create(name='Cat 2') def test_get(self): response = self.client.get(self.url) self.assertContains(response, self.cat1.name, status_code=200) self.assertEquals(response['Content-Type'], 'text/html; charset=utf-8') def test_post(self): data = { 'file_format': '0', } response = self.client.post(self.url, data) self.assertContains(response, self.cat1.name, status_code=200) self.assertTrue(response.has_header("Content-Disposition")) self.assertEquals(response['Content-Type'], 'text/csv')
from core.models import Category from django.test.testcases import TestCase from django.urls import reverse class ExportViewMixinTest(TestCase): def setUp(self): self.url = reverse('export-category') self.cat1 = Category.objects.create(name='Cat 1') self.cat2 = Category.objects.create(name='Cat 2') def test_get(self): response = self.client.get(self.url) self.assertContains(response, self.cat1.name, status_code=200) self.assertTrue(response['Content-Type'], 'text/html') def test_post(self): data = { 'file_format': '0', } response = self.client.post(self.url, data) self.assertContains(response, self.cat1.name, status_code=200) self.assertTrue(response.has_header("Content-Disposition")) self.assertTrue(response['Content-Type'], 'text/csv') Correct mistaken assertTrue() -> assertEquals()from core.models import Category from django.test.testcases import TestCase from django.urls import reverse class ExportViewMixinTest(TestCase): def setUp(self): self.url = reverse('export-category') self.cat1 = Category.objects.create(name='Cat 1') self.cat2 = Category.objects.create(name='Cat 2') def test_get(self): response = self.client.get(self.url) self.assertContains(response, self.cat1.name, status_code=200) self.assertEquals(response['Content-Type'], 'text/html; charset=utf-8') def test_post(self): data = { 'file_format': '0', } response = self.client.post(self.url, data) self.assertContains(response, self.cat1.name, status_code=200) self.assertTrue(response.has_header("Content-Disposition")) self.assertEquals(response['Content-Type'], 'text/csv')
<commit_before>from core.models import Category from django.test.testcases import TestCase from django.urls import reverse class ExportViewMixinTest(TestCase): def setUp(self): self.url = reverse('export-category') self.cat1 = Category.objects.create(name='Cat 1') self.cat2 = Category.objects.create(name='Cat 2') def test_get(self): response = self.client.get(self.url) self.assertContains(response, self.cat1.name, status_code=200) self.assertTrue(response['Content-Type'], 'text/html') def test_post(self): data = { 'file_format': '0', } response = self.client.post(self.url, data) self.assertContains(response, self.cat1.name, status_code=200) self.assertTrue(response.has_header("Content-Disposition")) self.assertTrue(response['Content-Type'], 'text/csv') <commit_msg>Correct mistaken assertTrue() -> assertEquals()<commit_after>from core.models import Category from django.test.testcases import TestCase from django.urls import reverse class ExportViewMixinTest(TestCase): def setUp(self): self.url = reverse('export-category') self.cat1 = Category.objects.create(name='Cat 1') self.cat2 = Category.objects.create(name='Cat 2') def test_get(self): response = self.client.get(self.url) self.assertContains(response, self.cat1.name, status_code=200) self.assertEquals(response['Content-Type'], 'text/html; charset=utf-8') def test_post(self): data = { 'file_format': '0', } response = self.client.post(self.url, data) self.assertContains(response, self.cat1.name, status_code=200) self.assertTrue(response.has_header("Content-Disposition")) self.assertEquals(response['Content-Type'], 'text/csv')
eb0d6d2c1d13e4dc7f84ca602ba95c2ebc31431a
src/lambda_function.py
src/lambda_function.py
"""lambda_function.py main entry point for aws lambda""" import json import urllib.request from typing import Dict, Optional import settings from logger import logger from spot import to_msw_id def close(fulfillment_state: str, message: Dict[str, str]) -> dict: """Close dialog generator""" return { 'dialogAction': { 'type': 'Close', 'fulfillmentState': fulfillment_state, 'message': message } } def lambda_handler(event: dict, context: dict) -> Optional[dict]: """Lambda Handler Entry point for every lambda function call """ if not settings.MSW_API: logger.error("Couldn't read SF_MSW_API env variable") return None spot = event.get("slots", {}).get("SurfSpot", None) spot = to_msw_id(spot) if spot is not None else None if not spot: logger.error("Couldn't parse spot id") return close('Fulfilled', {'contentType': 'PlainText', 'content': 'Surf spot not found :('}) logger.debug('event=%s context=%s', event, context) res = json.loads(urllib.request.urlopen( "http://magicseaweed.com/api/{}/forecast/?spot_id={}".format( settings.MSW_API, spot)).read()) logger.debug('json.response=%s', res) return close('Fulfilled', {'contentType': 'PlainText', 'content': 'Number of stars: {}'.format( res[0]['solidRating'])})
"""lambda_function.py main entry point for aws lambda""" import json import urllib.request from typing import Dict, Optional import settings from logger import logger from spot import to_msw_id def close(fulfillment_state: str, message: Dict[str, str]) -> dict: """Close dialog generator""" return { 'dialogAction': { 'type': 'Close', 'fulfillmentState': fulfillment_state, 'message': message } } def lambda_handler(event: dict, context: dict) -> Optional[dict]: """Lambda Handler Entry point for every lambda function call """ logger.debug('event=%s context=%s', event, context) if not settings.MSW_API: logger.error("Couldn't read SF_MSW_API env variable") return None spot = event.get("slots", {}).get("SurfSpot", None) spot = to_msw_id(spot) if spot is not None else None if not spot: logger.error("Couldn't parse spot id") return close('Fulfilled', {'contentType': 'PlainText', 'content': 'Surf spot not found :('}) res = json.loads(urllib.request.urlopen( "http://magicseaweed.com/api/{}/forecast/?spot_id={}".format( settings.MSW_API, spot)).read()) logger.debug('json.response=%s', res) return close('Fulfilled', {'contentType': 'PlainText', 'content': 'Number of stars: {}'.format( res[0]['solidRating'])})
Move debugger line to the top
Move debugger line to the top
Python
mit
Smotko/surfbot,Smotko/surfbot
"""lambda_function.py main entry point for aws lambda""" import json import urllib.request from typing import Dict, Optional import settings from logger import logger from spot import to_msw_id def close(fulfillment_state: str, message: Dict[str, str]) -> dict: """Close dialog generator""" return { 'dialogAction': { 'type': 'Close', 'fulfillmentState': fulfillment_state, 'message': message } } def lambda_handler(event: dict, context: dict) -> Optional[dict]: """Lambda Handler Entry point for every lambda function call """ if not settings.MSW_API: logger.error("Couldn't read SF_MSW_API env variable") return None spot = event.get("slots", {}).get("SurfSpot", None) spot = to_msw_id(spot) if spot is not None else None if not spot: logger.error("Couldn't parse spot id") return close('Fulfilled', {'contentType': 'PlainText', 'content': 'Surf spot not found :('}) logger.debug('event=%s context=%s', event, context) res = json.loads(urllib.request.urlopen( "http://magicseaweed.com/api/{}/forecast/?spot_id={}".format( settings.MSW_API, spot)).read()) logger.debug('json.response=%s', res) return close('Fulfilled', {'contentType': 'PlainText', 'content': 'Number of stars: {}'.format( res[0]['solidRating'])}) Move debugger line to the top
"""lambda_function.py main entry point for aws lambda""" import json import urllib.request from typing import Dict, Optional import settings from logger import logger from spot import to_msw_id def close(fulfillment_state: str, message: Dict[str, str]) -> dict: """Close dialog generator""" return { 'dialogAction': { 'type': 'Close', 'fulfillmentState': fulfillment_state, 'message': message } } def lambda_handler(event: dict, context: dict) -> Optional[dict]: """Lambda Handler Entry point for every lambda function call """ logger.debug('event=%s context=%s', event, context) if not settings.MSW_API: logger.error("Couldn't read SF_MSW_API env variable") return None spot = event.get("slots", {}).get("SurfSpot", None) spot = to_msw_id(spot) if spot is not None else None if not spot: logger.error("Couldn't parse spot id") return close('Fulfilled', {'contentType': 'PlainText', 'content': 'Surf spot not found :('}) res = json.loads(urllib.request.urlopen( "http://magicseaweed.com/api/{}/forecast/?spot_id={}".format( settings.MSW_API, spot)).read()) logger.debug('json.response=%s', res) return close('Fulfilled', {'contentType': 'PlainText', 'content': 'Number of stars: {}'.format( res[0]['solidRating'])})
<commit_before>"""lambda_function.py main entry point for aws lambda""" import json import urllib.request from typing import Dict, Optional import settings from logger import logger from spot import to_msw_id def close(fulfillment_state: str, message: Dict[str, str]) -> dict: """Close dialog generator""" return { 'dialogAction': { 'type': 'Close', 'fulfillmentState': fulfillment_state, 'message': message } } def lambda_handler(event: dict, context: dict) -> Optional[dict]: """Lambda Handler Entry point for every lambda function call """ if not settings.MSW_API: logger.error("Couldn't read SF_MSW_API env variable") return None spot = event.get("slots", {}).get("SurfSpot", None) spot = to_msw_id(spot) if spot is not None else None if not spot: logger.error("Couldn't parse spot id") return close('Fulfilled', {'contentType': 'PlainText', 'content': 'Surf spot not found :('}) logger.debug('event=%s context=%s', event, context) res = json.loads(urllib.request.urlopen( "http://magicseaweed.com/api/{}/forecast/?spot_id={}".format( settings.MSW_API, spot)).read()) logger.debug('json.response=%s', res) return close('Fulfilled', {'contentType': 'PlainText', 'content': 'Number of stars: {}'.format( res[0]['solidRating'])}) <commit_msg>Move debugger line to the top<commit_after>
"""lambda_function.py main entry point for aws lambda""" import json import urllib.request from typing import Dict, Optional import settings from logger import logger from spot import to_msw_id def close(fulfillment_state: str, message: Dict[str, str]) -> dict: """Close dialog generator""" return { 'dialogAction': { 'type': 'Close', 'fulfillmentState': fulfillment_state, 'message': message } } def lambda_handler(event: dict, context: dict) -> Optional[dict]: """Lambda Handler Entry point for every lambda function call """ logger.debug('event=%s context=%s', event, context) if not settings.MSW_API: logger.error("Couldn't read SF_MSW_API env variable") return None spot = event.get("slots", {}).get("SurfSpot", None) spot = to_msw_id(spot) if spot is not None else None if not spot: logger.error("Couldn't parse spot id") return close('Fulfilled', {'contentType': 'PlainText', 'content': 'Surf spot not found :('}) res = json.loads(urllib.request.urlopen( "http://magicseaweed.com/api/{}/forecast/?spot_id={}".format( settings.MSW_API, spot)).read()) logger.debug('json.response=%s', res) return close('Fulfilled', {'contentType': 'PlainText', 'content': 'Number of stars: {}'.format( res[0]['solidRating'])})
"""lambda_function.py main entry point for aws lambda""" import json import urllib.request from typing import Dict, Optional import settings from logger import logger from spot import to_msw_id def close(fulfillment_state: str, message: Dict[str, str]) -> dict: """Close dialog generator""" return { 'dialogAction': { 'type': 'Close', 'fulfillmentState': fulfillment_state, 'message': message } } def lambda_handler(event: dict, context: dict) -> Optional[dict]: """Lambda Handler Entry point for every lambda function call """ if not settings.MSW_API: logger.error("Couldn't read SF_MSW_API env variable") return None spot = event.get("slots", {}).get("SurfSpot", None) spot = to_msw_id(spot) if spot is not None else None if not spot: logger.error("Couldn't parse spot id") return close('Fulfilled', {'contentType': 'PlainText', 'content': 'Surf spot not found :('}) logger.debug('event=%s context=%s', event, context) res = json.loads(urllib.request.urlopen( "http://magicseaweed.com/api/{}/forecast/?spot_id={}".format( settings.MSW_API, spot)).read()) logger.debug('json.response=%s', res) return close('Fulfilled', {'contentType': 'PlainText', 'content': 'Number of stars: {}'.format( res[0]['solidRating'])}) Move debugger line to the top"""lambda_function.py main entry point for aws lambda""" import json import urllib.request from typing import Dict, Optional import settings from logger import logger from spot import to_msw_id def close(fulfillment_state: str, message: Dict[str, str]) -> dict: """Close dialog generator""" return { 'dialogAction': { 'type': 'Close', 'fulfillmentState': fulfillment_state, 'message': message } } def lambda_handler(event: dict, context: dict) -> Optional[dict]: """Lambda Handler Entry point for every lambda function call """ logger.debug('event=%s context=%s', event, context) if not settings.MSW_API: logger.error("Couldn't read SF_MSW_API env variable") return None spot = event.get("slots", {}).get("SurfSpot", None) spot = to_msw_id(spot) if spot is not None else None if not spot: logger.error("Couldn't parse spot id") return close('Fulfilled', {'contentType': 'PlainText', 'content': 'Surf spot not found :('}) res = json.loads(urllib.request.urlopen( "http://magicseaweed.com/api/{}/forecast/?spot_id={}".format( settings.MSW_API, spot)).read()) logger.debug('json.response=%s', res) return close('Fulfilled', {'contentType': 'PlainText', 'content': 'Number of stars: {}'.format( res[0]['solidRating'])})
<commit_before>"""lambda_function.py main entry point for aws lambda""" import json import urllib.request from typing import Dict, Optional import settings from logger import logger from spot import to_msw_id def close(fulfillment_state: str, message: Dict[str, str]) -> dict: """Close dialog generator""" return { 'dialogAction': { 'type': 'Close', 'fulfillmentState': fulfillment_state, 'message': message } } def lambda_handler(event: dict, context: dict) -> Optional[dict]: """Lambda Handler Entry point for every lambda function call """ if not settings.MSW_API: logger.error("Couldn't read SF_MSW_API env variable") return None spot = event.get("slots", {}).get("SurfSpot", None) spot = to_msw_id(spot) if spot is not None else None if not spot: logger.error("Couldn't parse spot id") return close('Fulfilled', {'contentType': 'PlainText', 'content': 'Surf spot not found :('}) logger.debug('event=%s context=%s', event, context) res = json.loads(urllib.request.urlopen( "http://magicseaweed.com/api/{}/forecast/?spot_id={}".format( settings.MSW_API, spot)).read()) logger.debug('json.response=%s', res) return close('Fulfilled', {'contentType': 'PlainText', 'content': 'Number of stars: {}'.format( res[0]['solidRating'])}) <commit_msg>Move debugger line to the top<commit_after>"""lambda_function.py main entry point for aws lambda""" import json import urllib.request from typing import Dict, Optional import settings from logger import logger from spot import to_msw_id def close(fulfillment_state: str, message: Dict[str, str]) -> dict: """Close dialog generator""" return { 'dialogAction': { 'type': 'Close', 'fulfillmentState': fulfillment_state, 'message': message } } def lambda_handler(event: dict, context: dict) -> Optional[dict]: """Lambda Handler Entry point for every lambda function call """ logger.debug('event=%s context=%s', event, context) if not settings.MSW_API: logger.error("Couldn't read SF_MSW_API env variable") return None spot = event.get("slots", {}).get("SurfSpot", None) spot = to_msw_id(spot) if spot is not None else None if not spot: logger.error("Couldn't parse spot id") return close('Fulfilled', {'contentType': 'PlainText', 'content': 'Surf spot not found :('}) res = json.loads(urllib.request.urlopen( "http://magicseaweed.com/api/{}/forecast/?spot_id={}".format( settings.MSW_API, spot)).read()) logger.debug('json.response=%s', res) return close('Fulfilled', {'contentType': 'PlainText', 'content': 'Number of stars: {}'.format( res[0]['solidRating'])})
1ba3536e214e283f503db0a9bf0d1ac4aa64f771
tcconfig/_tc_command_helper.py
tcconfig/_tc_command_helper.py
# encoding: utf-8 """ .. codeauthor:: Tsuyoshi Hombashi <tsuyoshi.hombashi@gmail.com> """ from __future__ import absolute_import, unicode_literals import errno import sys import subprocrunner as spr from ._common import find_bin_path from ._const import Tc, TcSubCommand from ._error import NetworkInterfaceNotFoundError from ._logger import logger def check_tc_command_installation(): try: spr.Which("tc").verify() except spr.CommandNotFoundError as e: logger.error("{:s}: {}".format(e.__class__.__name__, e)) sys.exit(errno.ENOENT) def get_tc_base_command(tc_subcommand): if tc_subcommand not in TcSubCommand: raise ValueError("the argument must be a TcSubCommand value") return "{:s} {:s}".format(find_bin_path("tc"), tc_subcommand.value) def run_tc_show(subcommand, device): from ._network import verify_network_interface verify_network_interface(device) runner = spr.SubprocessRunner( "{:s} show dev {:s}".format(get_tc_base_command(subcommand), device)) if runner.run() != 0 and runner.stderr.find("Cannot find device") != -1: # reach here if the device does not exist at the system and netiface # not installed. raise NetworkInterfaceNotFoundError(device=device) return runner.stdout
# encoding: utf-8 """ .. codeauthor:: Tsuyoshi Hombashi <tsuyoshi.hombashi@gmail.com> """ from __future__ import absolute_import, unicode_literals import errno import sys import subprocrunner as spr from ._common import find_bin_path from ._const import Tc, TcSubCommand from ._error import NetworkInterfaceNotFoundError from ._logger import logger def check_tc_command_installation(): if find_bin_path("tc"): return logger.error("command not found: tc") sys.exit(errno.ENOENT) def get_tc_base_command(tc_subcommand): if tc_subcommand not in TcSubCommand: raise ValueError("the argument must be a TcSubCommand value") return "{:s} {:s}".format(find_bin_path("tc"), tc_subcommand.value) def run_tc_show(subcommand, device): from ._network import verify_network_interface verify_network_interface(device) runner = spr.SubprocessRunner( "{:s} show dev {:s}".format(get_tc_base_command(subcommand), device)) if runner.run() != 0 and runner.stderr.find("Cannot find device") != -1: # reach here if the device does not exist at the system and netiface # not installed. raise NetworkInterfaceNotFoundError(device=device) return runner.stdout
Change command installation check process
Change command installation check process To properly check even if the user is not root.
Python
mit
thombashi/tcconfig,thombashi/tcconfig
# encoding: utf-8 """ .. codeauthor:: Tsuyoshi Hombashi <tsuyoshi.hombashi@gmail.com> """ from __future__ import absolute_import, unicode_literals import errno import sys import subprocrunner as spr from ._common import find_bin_path from ._const import Tc, TcSubCommand from ._error import NetworkInterfaceNotFoundError from ._logger import logger def check_tc_command_installation(): try: spr.Which("tc").verify() except spr.CommandNotFoundError as e: logger.error("{:s}: {}".format(e.__class__.__name__, e)) sys.exit(errno.ENOENT) def get_tc_base_command(tc_subcommand): if tc_subcommand not in TcSubCommand: raise ValueError("the argument must be a TcSubCommand value") return "{:s} {:s}".format(find_bin_path("tc"), tc_subcommand.value) def run_tc_show(subcommand, device): from ._network import verify_network_interface verify_network_interface(device) runner = spr.SubprocessRunner( "{:s} show dev {:s}".format(get_tc_base_command(subcommand), device)) if runner.run() != 0 and runner.stderr.find("Cannot find device") != -1: # reach here if the device does not exist at the system and netiface # not installed. raise NetworkInterfaceNotFoundError(device=device) return runner.stdout Change command installation check process To properly check even if the user is not root.
# encoding: utf-8 """ .. codeauthor:: Tsuyoshi Hombashi <tsuyoshi.hombashi@gmail.com> """ from __future__ import absolute_import, unicode_literals import errno import sys import subprocrunner as spr from ._common import find_bin_path from ._const import Tc, TcSubCommand from ._error import NetworkInterfaceNotFoundError from ._logger import logger def check_tc_command_installation(): if find_bin_path("tc"): return logger.error("command not found: tc") sys.exit(errno.ENOENT) def get_tc_base_command(tc_subcommand): if tc_subcommand not in TcSubCommand: raise ValueError("the argument must be a TcSubCommand value") return "{:s} {:s}".format(find_bin_path("tc"), tc_subcommand.value) def run_tc_show(subcommand, device): from ._network import verify_network_interface verify_network_interface(device) runner = spr.SubprocessRunner( "{:s} show dev {:s}".format(get_tc_base_command(subcommand), device)) if runner.run() != 0 and runner.stderr.find("Cannot find device") != -1: # reach here if the device does not exist at the system and netiface # not installed. raise NetworkInterfaceNotFoundError(device=device) return runner.stdout
<commit_before># encoding: utf-8 """ .. codeauthor:: Tsuyoshi Hombashi <tsuyoshi.hombashi@gmail.com> """ from __future__ import absolute_import, unicode_literals import errno import sys import subprocrunner as spr from ._common import find_bin_path from ._const import Tc, TcSubCommand from ._error import NetworkInterfaceNotFoundError from ._logger import logger def check_tc_command_installation(): try: spr.Which("tc").verify() except spr.CommandNotFoundError as e: logger.error("{:s}: {}".format(e.__class__.__name__, e)) sys.exit(errno.ENOENT) def get_tc_base_command(tc_subcommand): if tc_subcommand not in TcSubCommand: raise ValueError("the argument must be a TcSubCommand value") return "{:s} {:s}".format(find_bin_path("tc"), tc_subcommand.value) def run_tc_show(subcommand, device): from ._network import verify_network_interface verify_network_interface(device) runner = spr.SubprocessRunner( "{:s} show dev {:s}".format(get_tc_base_command(subcommand), device)) if runner.run() != 0 and runner.stderr.find("Cannot find device") != -1: # reach here if the device does not exist at the system and netiface # not installed. raise NetworkInterfaceNotFoundError(device=device) return runner.stdout <commit_msg>Change command installation check process To properly check even if the user is not root.<commit_after>
# encoding: utf-8 """ .. codeauthor:: Tsuyoshi Hombashi <tsuyoshi.hombashi@gmail.com> """ from __future__ import absolute_import, unicode_literals import errno import sys import subprocrunner as spr from ._common import find_bin_path from ._const import Tc, TcSubCommand from ._error import NetworkInterfaceNotFoundError from ._logger import logger def check_tc_command_installation(): if find_bin_path("tc"): return logger.error("command not found: tc") sys.exit(errno.ENOENT) def get_tc_base_command(tc_subcommand): if tc_subcommand not in TcSubCommand: raise ValueError("the argument must be a TcSubCommand value") return "{:s} {:s}".format(find_bin_path("tc"), tc_subcommand.value) def run_tc_show(subcommand, device): from ._network import verify_network_interface verify_network_interface(device) runner = spr.SubprocessRunner( "{:s} show dev {:s}".format(get_tc_base_command(subcommand), device)) if runner.run() != 0 and runner.stderr.find("Cannot find device") != -1: # reach here if the device does not exist at the system and netiface # not installed. raise NetworkInterfaceNotFoundError(device=device) return runner.stdout
# encoding: utf-8 """ .. codeauthor:: Tsuyoshi Hombashi <tsuyoshi.hombashi@gmail.com> """ from __future__ import absolute_import, unicode_literals import errno import sys import subprocrunner as spr from ._common import find_bin_path from ._const import Tc, TcSubCommand from ._error import NetworkInterfaceNotFoundError from ._logger import logger def check_tc_command_installation(): try: spr.Which("tc").verify() except spr.CommandNotFoundError as e: logger.error("{:s}: {}".format(e.__class__.__name__, e)) sys.exit(errno.ENOENT) def get_tc_base_command(tc_subcommand): if tc_subcommand not in TcSubCommand: raise ValueError("the argument must be a TcSubCommand value") return "{:s} {:s}".format(find_bin_path("tc"), tc_subcommand.value) def run_tc_show(subcommand, device): from ._network import verify_network_interface verify_network_interface(device) runner = spr.SubprocessRunner( "{:s} show dev {:s}".format(get_tc_base_command(subcommand), device)) if runner.run() != 0 and runner.stderr.find("Cannot find device") != -1: # reach here if the device does not exist at the system and netiface # not installed. raise NetworkInterfaceNotFoundError(device=device) return runner.stdout Change command installation check process To properly check even if the user is not root.# encoding: utf-8 """ .. codeauthor:: Tsuyoshi Hombashi <tsuyoshi.hombashi@gmail.com> """ from __future__ import absolute_import, unicode_literals import errno import sys import subprocrunner as spr from ._common import find_bin_path from ._const import Tc, TcSubCommand from ._error import NetworkInterfaceNotFoundError from ._logger import logger def check_tc_command_installation(): if find_bin_path("tc"): return logger.error("command not found: tc") sys.exit(errno.ENOENT) def get_tc_base_command(tc_subcommand): if tc_subcommand not in TcSubCommand: raise ValueError("the argument must be a TcSubCommand value") return "{:s} {:s}".format(find_bin_path("tc"), tc_subcommand.value) def run_tc_show(subcommand, device): from ._network import verify_network_interface verify_network_interface(device) runner = spr.SubprocessRunner( "{:s} show dev {:s}".format(get_tc_base_command(subcommand), device)) if runner.run() != 0 and runner.stderr.find("Cannot find device") != -1: # reach here if the device does not exist at the system and netiface # not installed. raise NetworkInterfaceNotFoundError(device=device) return runner.stdout
<commit_before># encoding: utf-8 """ .. codeauthor:: Tsuyoshi Hombashi <tsuyoshi.hombashi@gmail.com> """ from __future__ import absolute_import, unicode_literals import errno import sys import subprocrunner as spr from ._common import find_bin_path from ._const import Tc, TcSubCommand from ._error import NetworkInterfaceNotFoundError from ._logger import logger def check_tc_command_installation(): try: spr.Which("tc").verify() except spr.CommandNotFoundError as e: logger.error("{:s}: {}".format(e.__class__.__name__, e)) sys.exit(errno.ENOENT) def get_tc_base_command(tc_subcommand): if tc_subcommand not in TcSubCommand: raise ValueError("the argument must be a TcSubCommand value") return "{:s} {:s}".format(find_bin_path("tc"), tc_subcommand.value) def run_tc_show(subcommand, device): from ._network import verify_network_interface verify_network_interface(device) runner = spr.SubprocessRunner( "{:s} show dev {:s}".format(get_tc_base_command(subcommand), device)) if runner.run() != 0 and runner.stderr.find("Cannot find device") != -1: # reach here if the device does not exist at the system and netiface # not installed. raise NetworkInterfaceNotFoundError(device=device) return runner.stdout <commit_msg>Change command installation check process To properly check even if the user is not root.<commit_after># encoding: utf-8 """ .. codeauthor:: Tsuyoshi Hombashi <tsuyoshi.hombashi@gmail.com> """ from __future__ import absolute_import, unicode_literals import errno import sys import subprocrunner as spr from ._common import find_bin_path from ._const import Tc, TcSubCommand from ._error import NetworkInterfaceNotFoundError from ._logger import logger def check_tc_command_installation(): if find_bin_path("tc"): return logger.error("command not found: tc") sys.exit(errno.ENOENT) def get_tc_base_command(tc_subcommand): if tc_subcommand not in TcSubCommand: raise ValueError("the argument must be a TcSubCommand value") return "{:s} {:s}".format(find_bin_path("tc"), tc_subcommand.value) def run_tc_show(subcommand, device): from ._network import verify_network_interface verify_network_interface(device) runner = spr.SubprocessRunner( "{:s} show dev {:s}".format(get_tc_base_command(subcommand), device)) if runner.run() != 0 and runner.stderr.find("Cannot find device") != -1: # reach here if the device does not exist at the system and netiface # not installed. raise NetworkInterfaceNotFoundError(device=device) return runner.stdout
c8291c72b21be4b06a834449d89b2e4f91c1bc2b
dthm4kaiako/config/__init__.py
dthm4kaiako/config/__init__.py
"""Configuration for Django system.""" __version__ = "0.16.4" __version_info__ = tuple( [ int(num) if num.isdigit() else num for num in __version__.replace("-", ".", 1).split(".") ] )
"""Configuration for Django system.""" __version__ = "0.16.5" __version_info__ = tuple( [ int(num) if num.isdigit() else num for num in __version__.replace("-", ".", 1).split(".") ] )
Increment version number to 0.16.5
Increment version number to 0.16.5
Python
mit
uccser/cs4teachers,uccser/cs4teachers,uccser/cs4teachers,uccser/cs4teachers
"""Configuration for Django system.""" __version__ = "0.16.4" __version_info__ = tuple( [ int(num) if num.isdigit() else num for num in __version__.replace("-", ".", 1).split(".") ] ) Increment version number to 0.16.5
"""Configuration for Django system.""" __version__ = "0.16.5" __version_info__ = tuple( [ int(num) if num.isdigit() else num for num in __version__.replace("-", ".", 1).split(".") ] )
<commit_before>"""Configuration for Django system.""" __version__ = "0.16.4" __version_info__ = tuple( [ int(num) if num.isdigit() else num for num in __version__.replace("-", ".", 1).split(".") ] ) <commit_msg>Increment version number to 0.16.5<commit_after>
"""Configuration for Django system.""" __version__ = "0.16.5" __version_info__ = tuple( [ int(num) if num.isdigit() else num for num in __version__.replace("-", ".", 1).split(".") ] )
"""Configuration for Django system.""" __version__ = "0.16.4" __version_info__ = tuple( [ int(num) if num.isdigit() else num for num in __version__.replace("-", ".", 1).split(".") ] ) Increment version number to 0.16.5"""Configuration for Django system.""" __version__ = "0.16.5" __version_info__ = tuple( [ int(num) if num.isdigit() else num for num in __version__.replace("-", ".", 1).split(".") ] )
<commit_before>"""Configuration for Django system.""" __version__ = "0.16.4" __version_info__ = tuple( [ int(num) if num.isdigit() else num for num in __version__.replace("-", ".", 1).split(".") ] ) <commit_msg>Increment version number to 0.16.5<commit_after>"""Configuration for Django system.""" __version__ = "0.16.5" __version_info__ = tuple( [ int(num) if num.isdigit() else num for num in __version__.replace("-", ".", 1).split(".") ] )
abd2df6436d7a4a1304bf521c0f0a6c8922e5826
src/benchmark_rank_filter.py
src/benchmark_rank_filter.py
#!/usr/bin/env python import sys import timeit import numpy from rank_filter import lineRankOrderFilter def benchmark(): input_array = numpy.random.normal(size=(100, 101, 102)) output_array = numpy.empty_like(input_array) lineRankOrderFilter(input_array, 25, 0.5, 0, output_array) def main(*argv): times = 10 avg_time = timeit.timeit( "benchmark()", setup="from __main__ import benchmark", number=times ) / times print( "Of " + repr(times) + " trials," + " the average runtime = " + repr(avg_time) + "s." ) return(0) if __name__ == '__main__': sys.exit(main(*sys.argv))
#!/usr/bin/env python from __future__ import division import sys import timeit import numpy from rank_filter import lineRankOrderFilter def benchmark(): input_array = numpy.random.normal(size=(100, 101, 102)) output_array = numpy.empty_like(input_array) lineRankOrderFilter(input_array, 25, 0.5, 0, output_array) def main(*argv): times = 10 avg_time = timeit.timeit( "benchmark()", setup="from __main__ import benchmark", number=times ) / times print( "Of " + repr(times) + " trials," + " the average runtime = " + repr(avg_time) + "s." ) return(0) if __name__ == '__main__': sys.exit(main(*sys.argv))
Use Python 3 division on Python 2
Use Python 3 division on Python 2 Make sure that floating point division is used on Python 2. This basically happens anyways as we have a floating point value divided by an integral value. Still it is good to ensure that our average doesn't get truncated by accident.
Python
bsd-3-clause
nanshe-org/rank_filter,DudLab/rank_filter,jakirkham/rank_filter,jakirkham/rank_filter,nanshe-org/rank_filter,nanshe-org/rank_filter,DudLab/rank_filter,DudLab/rank_filter,jakirkham/rank_filter
#!/usr/bin/env python import sys import timeit import numpy from rank_filter import lineRankOrderFilter def benchmark(): input_array = numpy.random.normal(size=(100, 101, 102)) output_array = numpy.empty_like(input_array) lineRankOrderFilter(input_array, 25, 0.5, 0, output_array) def main(*argv): times = 10 avg_time = timeit.timeit( "benchmark()", setup="from __main__ import benchmark", number=times ) / times print( "Of " + repr(times) + " trials," + " the average runtime = " + repr(avg_time) + "s." ) return(0) if __name__ == '__main__': sys.exit(main(*sys.argv)) Use Python 3 division on Python 2 Make sure that floating point division is used on Python 2. This basically happens anyways as we have a floating point value divided by an integral value. Still it is good to ensure that our average doesn't get truncated by accident.
#!/usr/bin/env python from __future__ import division import sys import timeit import numpy from rank_filter import lineRankOrderFilter def benchmark(): input_array = numpy.random.normal(size=(100, 101, 102)) output_array = numpy.empty_like(input_array) lineRankOrderFilter(input_array, 25, 0.5, 0, output_array) def main(*argv): times = 10 avg_time = timeit.timeit( "benchmark()", setup="from __main__ import benchmark", number=times ) / times print( "Of " + repr(times) + " trials," + " the average runtime = " + repr(avg_time) + "s." ) return(0) if __name__ == '__main__': sys.exit(main(*sys.argv))
<commit_before>#!/usr/bin/env python import sys import timeit import numpy from rank_filter import lineRankOrderFilter def benchmark(): input_array = numpy.random.normal(size=(100, 101, 102)) output_array = numpy.empty_like(input_array) lineRankOrderFilter(input_array, 25, 0.5, 0, output_array) def main(*argv): times = 10 avg_time = timeit.timeit( "benchmark()", setup="from __main__ import benchmark", number=times ) / times print( "Of " + repr(times) + " trials," + " the average runtime = " + repr(avg_time) + "s." ) return(0) if __name__ == '__main__': sys.exit(main(*sys.argv)) <commit_msg>Use Python 3 division on Python 2 Make sure that floating point division is used on Python 2. This basically happens anyways as we have a floating point value divided by an integral value. Still it is good to ensure that our average doesn't get truncated by accident.<commit_after>
#!/usr/bin/env python from __future__ import division import sys import timeit import numpy from rank_filter import lineRankOrderFilter def benchmark(): input_array = numpy.random.normal(size=(100, 101, 102)) output_array = numpy.empty_like(input_array) lineRankOrderFilter(input_array, 25, 0.5, 0, output_array) def main(*argv): times = 10 avg_time = timeit.timeit( "benchmark()", setup="from __main__ import benchmark", number=times ) / times print( "Of " + repr(times) + " trials," + " the average runtime = " + repr(avg_time) + "s." ) return(0) if __name__ == '__main__': sys.exit(main(*sys.argv))
#!/usr/bin/env python import sys import timeit import numpy from rank_filter import lineRankOrderFilter def benchmark(): input_array = numpy.random.normal(size=(100, 101, 102)) output_array = numpy.empty_like(input_array) lineRankOrderFilter(input_array, 25, 0.5, 0, output_array) def main(*argv): times = 10 avg_time = timeit.timeit( "benchmark()", setup="from __main__ import benchmark", number=times ) / times print( "Of " + repr(times) + " trials," + " the average runtime = " + repr(avg_time) + "s." ) return(0) if __name__ == '__main__': sys.exit(main(*sys.argv)) Use Python 3 division on Python 2 Make sure that floating point division is used on Python 2. This basically happens anyways as we have a floating point value divided by an integral value. Still it is good to ensure that our average doesn't get truncated by accident.#!/usr/bin/env python from __future__ import division import sys import timeit import numpy from rank_filter import lineRankOrderFilter def benchmark(): input_array = numpy.random.normal(size=(100, 101, 102)) output_array = numpy.empty_like(input_array) lineRankOrderFilter(input_array, 25, 0.5, 0, output_array) def main(*argv): times = 10 avg_time = timeit.timeit( "benchmark()", setup="from __main__ import benchmark", number=times ) / times print( "Of " + repr(times) + " trials," + " the average runtime = " + repr(avg_time) + "s." ) return(0) if __name__ == '__main__': sys.exit(main(*sys.argv))
<commit_before>#!/usr/bin/env python import sys import timeit import numpy from rank_filter import lineRankOrderFilter def benchmark(): input_array = numpy.random.normal(size=(100, 101, 102)) output_array = numpy.empty_like(input_array) lineRankOrderFilter(input_array, 25, 0.5, 0, output_array) def main(*argv): times = 10 avg_time = timeit.timeit( "benchmark()", setup="from __main__ import benchmark", number=times ) / times print( "Of " + repr(times) + " trials," + " the average runtime = " + repr(avg_time) + "s." ) return(0) if __name__ == '__main__': sys.exit(main(*sys.argv)) <commit_msg>Use Python 3 division on Python 2 Make sure that floating point division is used on Python 2. This basically happens anyways as we have a floating point value divided by an integral value. Still it is good to ensure that our average doesn't get truncated by accident.<commit_after>#!/usr/bin/env python from __future__ import division import sys import timeit import numpy from rank_filter import lineRankOrderFilter def benchmark(): input_array = numpy.random.normal(size=(100, 101, 102)) output_array = numpy.empty_like(input_array) lineRankOrderFilter(input_array, 25, 0.5, 0, output_array) def main(*argv): times = 10 avg_time = timeit.timeit( "benchmark()", setup="from __main__ import benchmark", number=times ) / times print( "Of " + repr(times) + " trials," + " the average runtime = " + repr(avg_time) + "s." ) return(0) if __name__ == '__main__': sys.exit(main(*sys.argv))
d608d9f474f089df8f5d6e0e899554f9324e84f3
squash/dashboard/urls.py
squash/dashboard/urls.py
from django.conf.urls import include, url from django.contrib import admin from rest_framework.authtoken.views import obtain_auth_token from rest_framework.routers import DefaultRouter from . import views api_router = DefaultRouter() api_router.register(r'jobs', views.JobViewSet) api_router.register(r'metrics', views.MetricViewSet) urlpatterns = [ url(r'^dashboard/admin', include(admin.site.urls)), url(r'^dashboard/api/', include(api_router.urls)), url(r'^dashboard/api/token/', obtain_auth_token, name='api-token'), url(r'^$', views.HomeView.as_view(), name='home'), url(r'^(?P<pk>[a-zA-Z0-9]*$)', views.ListMetricsView.as_view(), name='metrics-list'), ]
from django.conf.urls import include, url from django.contrib import admin from rest_framework.authtoken.views import obtain_auth_token from rest_framework.routers import DefaultRouter from . import views admin.site.site_header = 'SQUASH Admin' api_router = DefaultRouter() api_router.register(r'jobs', views.JobViewSet) api_router.register(r'metrics', views.MetricViewSet) urlpatterns = [ url(r'^dashboard/admin', include(admin.site.urls)), url(r'^dashboard/api/', include(api_router.urls)), url(r'^dashboard/api/token/', obtain_auth_token, name='api-token'), url(r'^$', views.HomeView.as_view(), name='home'), url(r'^(?P<pk>[a-zA-Z0-9]*$)', views.ListMetricsView.as_view(), name='metrics-list'), ]
Set title for SQUASH admin interface
Set title for SQUASH admin interface
Python
mit
lsst-sqre/qa-dashboard,lsst-sqre/qa-dashboard,lsst-sqre/qa-dashboard
from django.conf.urls import include, url from django.contrib import admin from rest_framework.authtoken.views import obtain_auth_token from rest_framework.routers import DefaultRouter from . import views api_router = DefaultRouter() api_router.register(r'jobs', views.JobViewSet) api_router.register(r'metrics', views.MetricViewSet) urlpatterns = [ url(r'^dashboard/admin', include(admin.site.urls)), url(r'^dashboard/api/', include(api_router.urls)), url(r'^dashboard/api/token/', obtain_auth_token, name='api-token'), url(r'^$', views.HomeView.as_view(), name='home'), url(r'^(?P<pk>[a-zA-Z0-9]*$)', views.ListMetricsView.as_view(), name='metrics-list'), ] Set title for SQUASH admin interface
from django.conf.urls import include, url from django.contrib import admin from rest_framework.authtoken.views import obtain_auth_token from rest_framework.routers import DefaultRouter from . import views admin.site.site_header = 'SQUASH Admin' api_router = DefaultRouter() api_router.register(r'jobs', views.JobViewSet) api_router.register(r'metrics', views.MetricViewSet) urlpatterns = [ url(r'^dashboard/admin', include(admin.site.urls)), url(r'^dashboard/api/', include(api_router.urls)), url(r'^dashboard/api/token/', obtain_auth_token, name='api-token'), url(r'^$', views.HomeView.as_view(), name='home'), url(r'^(?P<pk>[a-zA-Z0-9]*$)', views.ListMetricsView.as_view(), name='metrics-list'), ]
<commit_before>from django.conf.urls import include, url from django.contrib import admin from rest_framework.authtoken.views import obtain_auth_token from rest_framework.routers import DefaultRouter from . import views api_router = DefaultRouter() api_router.register(r'jobs', views.JobViewSet) api_router.register(r'metrics', views.MetricViewSet) urlpatterns = [ url(r'^dashboard/admin', include(admin.site.urls)), url(r'^dashboard/api/', include(api_router.urls)), url(r'^dashboard/api/token/', obtain_auth_token, name='api-token'), url(r'^$', views.HomeView.as_view(), name='home'), url(r'^(?P<pk>[a-zA-Z0-9]*$)', views.ListMetricsView.as_view(), name='metrics-list'), ] <commit_msg>Set title for SQUASH admin interface<commit_after>
from django.conf.urls import include, url from django.contrib import admin from rest_framework.authtoken.views import obtain_auth_token from rest_framework.routers import DefaultRouter from . import views admin.site.site_header = 'SQUASH Admin' api_router = DefaultRouter() api_router.register(r'jobs', views.JobViewSet) api_router.register(r'metrics', views.MetricViewSet) urlpatterns = [ url(r'^dashboard/admin', include(admin.site.urls)), url(r'^dashboard/api/', include(api_router.urls)), url(r'^dashboard/api/token/', obtain_auth_token, name='api-token'), url(r'^$', views.HomeView.as_view(), name='home'), url(r'^(?P<pk>[a-zA-Z0-9]*$)', views.ListMetricsView.as_view(), name='metrics-list'), ]
from django.conf.urls import include, url from django.contrib import admin from rest_framework.authtoken.views import obtain_auth_token from rest_framework.routers import DefaultRouter from . import views api_router = DefaultRouter() api_router.register(r'jobs', views.JobViewSet) api_router.register(r'metrics', views.MetricViewSet) urlpatterns = [ url(r'^dashboard/admin', include(admin.site.urls)), url(r'^dashboard/api/', include(api_router.urls)), url(r'^dashboard/api/token/', obtain_auth_token, name='api-token'), url(r'^$', views.HomeView.as_view(), name='home'), url(r'^(?P<pk>[a-zA-Z0-9]*$)', views.ListMetricsView.as_view(), name='metrics-list'), ] Set title for SQUASH admin interfacefrom django.conf.urls import include, url from django.contrib import admin from rest_framework.authtoken.views import obtain_auth_token from rest_framework.routers import DefaultRouter from . import views admin.site.site_header = 'SQUASH Admin' api_router = DefaultRouter() api_router.register(r'jobs', views.JobViewSet) api_router.register(r'metrics', views.MetricViewSet) urlpatterns = [ url(r'^dashboard/admin', include(admin.site.urls)), url(r'^dashboard/api/', include(api_router.urls)), url(r'^dashboard/api/token/', obtain_auth_token, name='api-token'), url(r'^$', views.HomeView.as_view(), name='home'), url(r'^(?P<pk>[a-zA-Z0-9]*$)', views.ListMetricsView.as_view(), name='metrics-list'), ]
<commit_before>from django.conf.urls import include, url from django.contrib import admin from rest_framework.authtoken.views import obtain_auth_token from rest_framework.routers import DefaultRouter from . import views api_router = DefaultRouter() api_router.register(r'jobs', views.JobViewSet) api_router.register(r'metrics', views.MetricViewSet) urlpatterns = [ url(r'^dashboard/admin', include(admin.site.urls)), url(r'^dashboard/api/', include(api_router.urls)), url(r'^dashboard/api/token/', obtain_auth_token, name='api-token'), url(r'^$', views.HomeView.as_view(), name='home'), url(r'^(?P<pk>[a-zA-Z0-9]*$)', views.ListMetricsView.as_view(), name='metrics-list'), ] <commit_msg>Set title for SQUASH admin interface<commit_after>from django.conf.urls import include, url from django.contrib import admin from rest_framework.authtoken.views import obtain_auth_token from rest_framework.routers import DefaultRouter from . import views admin.site.site_header = 'SQUASH Admin' api_router = DefaultRouter() api_router.register(r'jobs', views.JobViewSet) api_router.register(r'metrics', views.MetricViewSet) urlpatterns = [ url(r'^dashboard/admin', include(admin.site.urls)), url(r'^dashboard/api/', include(api_router.urls)), url(r'^dashboard/api/token/', obtain_auth_token, name='api-token'), url(r'^$', views.HomeView.as_view(), name='home'), url(r'^(?P<pk>[a-zA-Z0-9]*$)', views.ListMetricsView.as_view(), name='metrics-list'), ]
1dca59c4de479d2a75c34c0d1aae3948d819b84b
imagersite/imager_profile/models.py
imagersite/imager_profile/models.py
"""Models.""" from django.db import models # Create your models here. class ImagerProfile(models.Model): """Imager Profile Model.""" camera_model = models.CharField(max_length=200) photography_type = models.TextField() Friends = models.ManyToManyField('self') Region = models.CharField(max_length=200)
"""Models.""" from django.db import models # Create your models here. class ImagerProfile(models.Model): """Imager Profile Model.""" camera_model = models.CharField(max_length=200) photography_type = models.TextField() friends = models.ManyToManyField('self') region = models.CharField(max_length=200)
Change model fields to lower case letters
Change model fields to lower case letters
Python
mit
DZwell/django-imager
"""Models.""" from django.db import models # Create your models here. class ImagerProfile(models.Model): """Imager Profile Model.""" camera_model = models.CharField(max_length=200) photography_type = models.TextField() Friends = models.ManyToManyField('self') Region = models.CharField(max_length=200) Change model fields to lower case letters
"""Models.""" from django.db import models # Create your models here. class ImagerProfile(models.Model): """Imager Profile Model.""" camera_model = models.CharField(max_length=200) photography_type = models.TextField() friends = models.ManyToManyField('self') region = models.CharField(max_length=200)
<commit_before>"""Models.""" from django.db import models # Create your models here. class ImagerProfile(models.Model): """Imager Profile Model.""" camera_model = models.CharField(max_length=200) photography_type = models.TextField() Friends = models.ManyToManyField('self') Region = models.CharField(max_length=200) <commit_msg>Change model fields to lower case letters<commit_after>
"""Models.""" from django.db import models # Create your models here. class ImagerProfile(models.Model): """Imager Profile Model.""" camera_model = models.CharField(max_length=200) photography_type = models.TextField() friends = models.ManyToManyField('self') region = models.CharField(max_length=200)
"""Models.""" from django.db import models # Create your models here. class ImagerProfile(models.Model): """Imager Profile Model.""" camera_model = models.CharField(max_length=200) photography_type = models.TextField() Friends = models.ManyToManyField('self') Region = models.CharField(max_length=200) Change model fields to lower case letters"""Models.""" from django.db import models # Create your models here. class ImagerProfile(models.Model): """Imager Profile Model.""" camera_model = models.CharField(max_length=200) photography_type = models.TextField() friends = models.ManyToManyField('self') region = models.CharField(max_length=200)
<commit_before>"""Models.""" from django.db import models # Create your models here. class ImagerProfile(models.Model): """Imager Profile Model.""" camera_model = models.CharField(max_length=200) photography_type = models.TextField() Friends = models.ManyToManyField('self') Region = models.CharField(max_length=200) <commit_msg>Change model fields to lower case letters<commit_after>"""Models.""" from django.db import models # Create your models here. class ImagerProfile(models.Model): """Imager Profile Model.""" camera_model = models.CharField(max_length=200) photography_type = models.TextField() friends = models.ManyToManyField('self') region = models.CharField(max_length=200)
918568524f1c5ef7264fee76597e8a88b6e2427f
src/ussclicore/utils/ascii_bar_graph.py
src/ussclicore/utils/ascii_bar_graph.py
# To change this license header, choose License Headers in Project Properties. # To change this template file, choose Tools | Templates # and open the template in the editor. __author__="UShareSoft" def print_graph(values): max=0 for v in values: if len(v)>max: max=len(v) for v in values: if len(v)<max: newV=v+(" " * int(max-len(v))) if values[v]!=-1: print newV, values[v]*'|'+'-'*int(50-values[v]) else: print newV,20*'-'+"UNLIMITED"+21*'-' else: if values[v]!=-1: print v, values[v]*'|'+'-'*int(50-values[v]) else: print v,20*'-'+"UNLIMITED"+21*'-'
__author__="UShareSoft" def print_graph(values): max=0 for v in values: if len(v)>max: max=len(v) for v in values: value = int(values[v]) if len(v)<max: newV=v+(" " * int(max-len(v))) if value!=-1: print newV, value*'|'+'-'*int(50-value) else: print newV,20*'-'+"UNLIMITED"+21*'-' else: if value!=-1: print v, value*'|'+'-'*int(50-value) else: print v,20*'-'+"UNLIMITED"+21*'-'
Fix bug in the ascii bar graph
Fix bug in the ascii bar graph
Python
apache-2.0
yanngit/ussclicore,usharesoft/ussclicore
# To change this license header, choose License Headers in Project Properties. # To change this template file, choose Tools | Templates # and open the template in the editor. __author__="UShareSoft" def print_graph(values): max=0 for v in values: if len(v)>max: max=len(v) for v in values: if len(v)<max: newV=v+(" " * int(max-len(v))) if values[v]!=-1: print newV, values[v]*'|'+'-'*int(50-values[v]) else: print newV,20*'-'+"UNLIMITED"+21*'-' else: if values[v]!=-1: print v, values[v]*'|'+'-'*int(50-values[v]) else: print v,20*'-'+"UNLIMITED"+21*'-' Fix bug in the ascii bar graph
__author__="UShareSoft" def print_graph(values): max=0 for v in values: if len(v)>max: max=len(v) for v in values: value = int(values[v]) if len(v)<max: newV=v+(" " * int(max-len(v))) if value!=-1: print newV, value*'|'+'-'*int(50-value) else: print newV,20*'-'+"UNLIMITED"+21*'-' else: if value!=-1: print v, value*'|'+'-'*int(50-value) else: print v,20*'-'+"UNLIMITED"+21*'-'
<commit_before># To change this license header, choose License Headers in Project Properties. # To change this template file, choose Tools | Templates # and open the template in the editor. __author__="UShareSoft" def print_graph(values): max=0 for v in values: if len(v)>max: max=len(v) for v in values: if len(v)<max: newV=v+(" " * int(max-len(v))) if values[v]!=-1: print newV, values[v]*'|'+'-'*int(50-values[v]) else: print newV,20*'-'+"UNLIMITED"+21*'-' else: if values[v]!=-1: print v, values[v]*'|'+'-'*int(50-values[v]) else: print v,20*'-'+"UNLIMITED"+21*'-' <commit_msg>Fix bug in the ascii bar graph<commit_after>
__author__="UShareSoft" def print_graph(values): max=0 for v in values: if len(v)>max: max=len(v) for v in values: value = int(values[v]) if len(v)<max: newV=v+(" " * int(max-len(v))) if value!=-1: print newV, value*'|'+'-'*int(50-value) else: print newV,20*'-'+"UNLIMITED"+21*'-' else: if value!=-1: print v, value*'|'+'-'*int(50-value) else: print v,20*'-'+"UNLIMITED"+21*'-'
# To change this license header, choose License Headers in Project Properties. # To change this template file, choose Tools | Templates # and open the template in the editor. __author__="UShareSoft" def print_graph(values): max=0 for v in values: if len(v)>max: max=len(v) for v in values: if len(v)<max: newV=v+(" " * int(max-len(v))) if values[v]!=-1: print newV, values[v]*'|'+'-'*int(50-values[v]) else: print newV,20*'-'+"UNLIMITED"+21*'-' else: if values[v]!=-1: print v, values[v]*'|'+'-'*int(50-values[v]) else: print v,20*'-'+"UNLIMITED"+21*'-' Fix bug in the ascii bar graph__author__="UShareSoft" def print_graph(values): max=0 for v in values: if len(v)>max: max=len(v) for v in values: value = int(values[v]) if len(v)<max: newV=v+(" " * int(max-len(v))) if value!=-1: print newV, value*'|'+'-'*int(50-value) else: print newV,20*'-'+"UNLIMITED"+21*'-' else: if value!=-1: print v, value*'|'+'-'*int(50-value) else: print v,20*'-'+"UNLIMITED"+21*'-'
<commit_before># To change this license header, choose License Headers in Project Properties. # To change this template file, choose Tools | Templates # and open the template in the editor. __author__="UShareSoft" def print_graph(values): max=0 for v in values: if len(v)>max: max=len(v) for v in values: if len(v)<max: newV=v+(" " * int(max-len(v))) if values[v]!=-1: print newV, values[v]*'|'+'-'*int(50-values[v]) else: print newV,20*'-'+"UNLIMITED"+21*'-' else: if values[v]!=-1: print v, values[v]*'|'+'-'*int(50-values[v]) else: print v,20*'-'+"UNLIMITED"+21*'-' <commit_msg>Fix bug in the ascii bar graph<commit_after>__author__="UShareSoft" def print_graph(values): max=0 for v in values: if len(v)>max: max=len(v) for v in values: value = int(values[v]) if len(v)<max: newV=v+(" " * int(max-len(v))) if value!=-1: print newV, value*'|'+'-'*int(50-value) else: print newV,20*'-'+"UNLIMITED"+21*'-' else: if value!=-1: print v, value*'|'+'-'*int(50-value) else: print v,20*'-'+"UNLIMITED"+21*'-'
205b832c287bdd587eff7ba266a4429f6aebb277
data/models.py
data/models.py
import numpy import ast from django.db import models class DataPoint(models.Model): name = models.CharField(max_length=600) exact_name = models.CharField(max_length=1000, null=True, blank=True) decay_feature = models.CharField(max_length=1000, null=True, blank=True) created = models.DateTimeField(auto_now_add=True) options = models.CharField(max_length=100) homo = models.FloatField() lumo = models.FloatField() homo_orbital = models.IntegerField() energy = models.FloatField() dipole = models.FloatField() band_gap = models.FloatField(null=True, blank=True) def __unicode__(self): return self.exact_name @classmethod def get_all_data(cls): data = DataPoint.objects.filter(band_gap__isnull=False, exact_name__isnull=False, decay_feature__isnull=False) M = len(data) HOMO = numpy.zeros((M, 1)) LUMO = numpy.zeros((M, 1)) GAP = numpy.zeros((M, 1)) vectors = [] for i, x in enumerate(data): HOMO[i] = x.homo LUMO[i] = x.lumo GAP[i] = x.band_gap vectors.append(ast.literal_eval(x.decay_feature)) FEATURE = numpy.matrix(vectors) return FEATURE, HOMO, LUMO, GAP
import numpy import ast from django.db import models class DataPoint(models.Model): name = models.CharField(max_length=600) exact_name = models.CharField(max_length=1000, null=True, blank=True) decay_feature = models.CharField(max_length=1000, null=True, blank=True) created = models.DateTimeField(auto_now_add=True) options = models.CharField(max_length=100) homo = models.FloatField() lumo = models.FloatField() homo_orbital = models.IntegerField() energy = models.FloatField() dipole = models.FloatField() band_gap = models.FloatField(null=True, blank=True) def __unicode__(self): return unicode(self.name) @classmethod def get_all_data(cls): data = DataPoint.objects.filter(band_gap__isnull=False, exact_name__isnull=False, decay_feature__isnull=False) M = len(data) HOMO = numpy.zeros((M, 1)) LUMO = numpy.zeros((M, 1)) GAP = numpy.zeros((M, 1)) vectors = [] for i, x in enumerate(data): HOMO[i] = x.homo LUMO[i] = x.lumo GAP[i] = x.band_gap vectors.append(ast.literal_eval(x.decay_feature)) FEATURE = numpy.matrix(vectors) return FEATURE, HOMO, LUMO, GAP
Fix __unicode__ for DataPoint model
Fix __unicode__ for DataPoint model
Python
mit
crcollins/chemtools-webapp,crcollins/chemtools-webapp,crcollins/chemtools-webapp,crcollins/chemtools-webapp,crcollins/chemtools-webapp
import numpy import ast from django.db import models class DataPoint(models.Model): name = models.CharField(max_length=600) exact_name = models.CharField(max_length=1000, null=True, blank=True) decay_feature = models.CharField(max_length=1000, null=True, blank=True) created = models.DateTimeField(auto_now_add=True) options = models.CharField(max_length=100) homo = models.FloatField() lumo = models.FloatField() homo_orbital = models.IntegerField() energy = models.FloatField() dipole = models.FloatField() band_gap = models.FloatField(null=True, blank=True) def __unicode__(self): return self.exact_name @classmethod def get_all_data(cls): data = DataPoint.objects.filter(band_gap__isnull=False, exact_name__isnull=False, decay_feature__isnull=False) M = len(data) HOMO = numpy.zeros((M, 1)) LUMO = numpy.zeros((M, 1)) GAP = numpy.zeros((M, 1)) vectors = [] for i, x in enumerate(data): HOMO[i] = x.homo LUMO[i] = x.lumo GAP[i] = x.band_gap vectors.append(ast.literal_eval(x.decay_feature)) FEATURE = numpy.matrix(vectors) return FEATURE, HOMO, LUMO, GAPFix __unicode__ for DataPoint model
import numpy import ast from django.db import models class DataPoint(models.Model): name = models.CharField(max_length=600) exact_name = models.CharField(max_length=1000, null=True, blank=True) decay_feature = models.CharField(max_length=1000, null=True, blank=True) created = models.DateTimeField(auto_now_add=True) options = models.CharField(max_length=100) homo = models.FloatField() lumo = models.FloatField() homo_orbital = models.IntegerField() energy = models.FloatField() dipole = models.FloatField() band_gap = models.FloatField(null=True, blank=True) def __unicode__(self): return unicode(self.name) @classmethod def get_all_data(cls): data = DataPoint.objects.filter(band_gap__isnull=False, exact_name__isnull=False, decay_feature__isnull=False) M = len(data) HOMO = numpy.zeros((M, 1)) LUMO = numpy.zeros((M, 1)) GAP = numpy.zeros((M, 1)) vectors = [] for i, x in enumerate(data): HOMO[i] = x.homo LUMO[i] = x.lumo GAP[i] = x.band_gap vectors.append(ast.literal_eval(x.decay_feature)) FEATURE = numpy.matrix(vectors) return FEATURE, HOMO, LUMO, GAP
<commit_before>import numpy import ast from django.db import models class DataPoint(models.Model): name = models.CharField(max_length=600) exact_name = models.CharField(max_length=1000, null=True, blank=True) decay_feature = models.CharField(max_length=1000, null=True, blank=True) created = models.DateTimeField(auto_now_add=True) options = models.CharField(max_length=100) homo = models.FloatField() lumo = models.FloatField() homo_orbital = models.IntegerField() energy = models.FloatField() dipole = models.FloatField() band_gap = models.FloatField(null=True, blank=True) def __unicode__(self): return self.exact_name @classmethod def get_all_data(cls): data = DataPoint.objects.filter(band_gap__isnull=False, exact_name__isnull=False, decay_feature__isnull=False) M = len(data) HOMO = numpy.zeros((M, 1)) LUMO = numpy.zeros((M, 1)) GAP = numpy.zeros((M, 1)) vectors = [] for i, x in enumerate(data): HOMO[i] = x.homo LUMO[i] = x.lumo GAP[i] = x.band_gap vectors.append(ast.literal_eval(x.decay_feature)) FEATURE = numpy.matrix(vectors) return FEATURE, HOMO, LUMO, GAP<commit_msg>Fix __unicode__ for DataPoint model<commit_after>
import numpy import ast from django.db import models class DataPoint(models.Model): name = models.CharField(max_length=600) exact_name = models.CharField(max_length=1000, null=True, blank=True) decay_feature = models.CharField(max_length=1000, null=True, blank=True) created = models.DateTimeField(auto_now_add=True) options = models.CharField(max_length=100) homo = models.FloatField() lumo = models.FloatField() homo_orbital = models.IntegerField() energy = models.FloatField() dipole = models.FloatField() band_gap = models.FloatField(null=True, blank=True) def __unicode__(self): return unicode(self.name) @classmethod def get_all_data(cls): data = DataPoint.objects.filter(band_gap__isnull=False, exact_name__isnull=False, decay_feature__isnull=False) M = len(data) HOMO = numpy.zeros((M, 1)) LUMO = numpy.zeros((M, 1)) GAP = numpy.zeros((M, 1)) vectors = [] for i, x in enumerate(data): HOMO[i] = x.homo LUMO[i] = x.lumo GAP[i] = x.band_gap vectors.append(ast.literal_eval(x.decay_feature)) FEATURE = numpy.matrix(vectors) return FEATURE, HOMO, LUMO, GAP
import numpy import ast from django.db import models class DataPoint(models.Model): name = models.CharField(max_length=600) exact_name = models.CharField(max_length=1000, null=True, blank=True) decay_feature = models.CharField(max_length=1000, null=True, blank=True) created = models.DateTimeField(auto_now_add=True) options = models.CharField(max_length=100) homo = models.FloatField() lumo = models.FloatField() homo_orbital = models.IntegerField() energy = models.FloatField() dipole = models.FloatField() band_gap = models.FloatField(null=True, blank=True) def __unicode__(self): return self.exact_name @classmethod def get_all_data(cls): data = DataPoint.objects.filter(band_gap__isnull=False, exact_name__isnull=False, decay_feature__isnull=False) M = len(data) HOMO = numpy.zeros((M, 1)) LUMO = numpy.zeros((M, 1)) GAP = numpy.zeros((M, 1)) vectors = [] for i, x in enumerate(data): HOMO[i] = x.homo LUMO[i] = x.lumo GAP[i] = x.band_gap vectors.append(ast.literal_eval(x.decay_feature)) FEATURE = numpy.matrix(vectors) return FEATURE, HOMO, LUMO, GAPFix __unicode__ for DataPoint modelimport numpy import ast from django.db import models class DataPoint(models.Model): name = models.CharField(max_length=600) exact_name = models.CharField(max_length=1000, null=True, blank=True) decay_feature = models.CharField(max_length=1000, null=True, blank=True) created = models.DateTimeField(auto_now_add=True) options = models.CharField(max_length=100) homo = models.FloatField() lumo = models.FloatField() homo_orbital = models.IntegerField() energy = models.FloatField() dipole = models.FloatField() band_gap = models.FloatField(null=True, blank=True) def __unicode__(self): return unicode(self.name) @classmethod def get_all_data(cls): data = DataPoint.objects.filter(band_gap__isnull=False, exact_name__isnull=False, decay_feature__isnull=False) M = len(data) HOMO = numpy.zeros((M, 1)) LUMO = numpy.zeros((M, 1)) GAP = numpy.zeros((M, 1)) vectors = [] for i, x in enumerate(data): HOMO[i] = x.homo LUMO[i] = x.lumo GAP[i] = x.band_gap vectors.append(ast.literal_eval(x.decay_feature)) FEATURE = numpy.matrix(vectors) return FEATURE, HOMO, LUMO, GAP
<commit_before>import numpy import ast from django.db import models class DataPoint(models.Model): name = models.CharField(max_length=600) exact_name = models.CharField(max_length=1000, null=True, blank=True) decay_feature = models.CharField(max_length=1000, null=True, blank=True) created = models.DateTimeField(auto_now_add=True) options = models.CharField(max_length=100) homo = models.FloatField() lumo = models.FloatField() homo_orbital = models.IntegerField() energy = models.FloatField() dipole = models.FloatField() band_gap = models.FloatField(null=True, blank=True) def __unicode__(self): return self.exact_name @classmethod def get_all_data(cls): data = DataPoint.objects.filter(band_gap__isnull=False, exact_name__isnull=False, decay_feature__isnull=False) M = len(data) HOMO = numpy.zeros((M, 1)) LUMO = numpy.zeros((M, 1)) GAP = numpy.zeros((M, 1)) vectors = [] for i, x in enumerate(data): HOMO[i] = x.homo LUMO[i] = x.lumo GAP[i] = x.band_gap vectors.append(ast.literal_eval(x.decay_feature)) FEATURE = numpy.matrix(vectors) return FEATURE, HOMO, LUMO, GAP<commit_msg>Fix __unicode__ for DataPoint model<commit_after>import numpy import ast from django.db import models class DataPoint(models.Model): name = models.CharField(max_length=600) exact_name = models.CharField(max_length=1000, null=True, blank=True) decay_feature = models.CharField(max_length=1000, null=True, blank=True) created = models.DateTimeField(auto_now_add=True) options = models.CharField(max_length=100) homo = models.FloatField() lumo = models.FloatField() homo_orbital = models.IntegerField() energy = models.FloatField() dipole = models.FloatField() band_gap = models.FloatField(null=True, blank=True) def __unicode__(self): return unicode(self.name) @classmethod def get_all_data(cls): data = DataPoint.objects.filter(band_gap__isnull=False, exact_name__isnull=False, decay_feature__isnull=False) M = len(data) HOMO = numpy.zeros((M, 1)) LUMO = numpy.zeros((M, 1)) GAP = numpy.zeros((M, 1)) vectors = [] for i, x in enumerate(data): HOMO[i] = x.homo LUMO[i] = x.lumo GAP[i] = x.band_gap vectors.append(ast.literal_eval(x.decay_feature)) FEATURE = numpy.matrix(vectors) return FEATURE, HOMO, LUMO, GAP
f8df781d4f2d96496f59942646718e9b30c9337f
tests/test_construct_policy.py
tests/test_construct_policy.py
"""Test IAM Policies for correctness.""" import json from foremast.iam.construct_policy import construct_policy ANSWER1 = { 'Version': '2012-10-17', 'Statement': [ { 'Effect': 'Allow', 'Action': [ 's3:GetObject', 's3:ListObject' ], 'Resource': [ 'arn:aws:s3:::archaius-stage/forrest/unicornforrest', 'arn:aws:s3:::archaius-stage/forrest/unicornforrest/*' ] } ] } def test_main(): """Check general assemblage.""" settings = {'services': {'s3': True}} policy_json = construct_policy(app='unicornforrest', env='stage', group='forrest', pipeline_settings=settings) assert json.loads(policy_json) == ANSWER1 # TODO: Test other services besides S3 settings.update({'services': {'dynamodb': ['coreforrest', 'edgeforrest', 'attendantdevops']}}) policy_json = construct_policy(pipeline_settings=settings) policy = json.loads(policy_json)
"""Test IAM Policies for correctness.""" import json from foremast.iam.construct_policy import construct_policy ANSWER1 = { 'Version': '2012-10-17', 'Statement': [ { 'Effect': 'Allow', 'Action': [ 's3:GetObject', 's3:ListObject' ], 'Resource': [ 'arn:aws:s3:::archaius-stage/forrest/unicornforrest', 'arn:aws:s3:::archaius-stage/forrest/unicornforrest/*' ] } ] } def test_main(): """Check general assemblage.""" settings = {} policy_json = construct_policy(pipeline_settings=settings) assert json.loads(policy_json) == {} settings = {'services': {'s3': True}} policy_json = construct_policy(app='unicornforrest', env='stage', group='forrest', pipeline_settings=settings) assert json.loads(policy_json) == ANSWER1 # TODO: Test other services besides S3 settings.update({'services': {'dynamodb': ['coreforrest', 'edgeforrest', 'attendantdevops']}}) policy_json = construct_policy(pipeline_settings=settings) policy = json.loads(policy_json)
Test IAM Policy with no "services"
tests: Test IAM Policy with no "services" See also: PSOBAT-1482
Python
apache-2.0
gogoair/foremast,gogoair/foremast
"""Test IAM Policies for correctness.""" import json from foremast.iam.construct_policy import construct_policy ANSWER1 = { 'Version': '2012-10-17', 'Statement': [ { 'Effect': 'Allow', 'Action': [ 's3:GetObject', 's3:ListObject' ], 'Resource': [ 'arn:aws:s3:::archaius-stage/forrest/unicornforrest', 'arn:aws:s3:::archaius-stage/forrest/unicornforrest/*' ] } ] } def test_main(): """Check general assemblage.""" settings = {'services': {'s3': True}} policy_json = construct_policy(app='unicornforrest', env='stage', group='forrest', pipeline_settings=settings) assert json.loads(policy_json) == ANSWER1 # TODO: Test other services besides S3 settings.update({'services': {'dynamodb': ['coreforrest', 'edgeforrest', 'attendantdevops']}}) policy_json = construct_policy(pipeline_settings=settings) policy = json.loads(policy_json) tests: Test IAM Policy with no "services" See also: PSOBAT-1482
"""Test IAM Policies for correctness.""" import json from foremast.iam.construct_policy import construct_policy ANSWER1 = { 'Version': '2012-10-17', 'Statement': [ { 'Effect': 'Allow', 'Action': [ 's3:GetObject', 's3:ListObject' ], 'Resource': [ 'arn:aws:s3:::archaius-stage/forrest/unicornforrest', 'arn:aws:s3:::archaius-stage/forrest/unicornforrest/*' ] } ] } def test_main(): """Check general assemblage.""" settings = {} policy_json = construct_policy(pipeline_settings=settings) assert json.loads(policy_json) == {} settings = {'services': {'s3': True}} policy_json = construct_policy(app='unicornforrest', env='stage', group='forrest', pipeline_settings=settings) assert json.loads(policy_json) == ANSWER1 # TODO: Test other services besides S3 settings.update({'services': {'dynamodb': ['coreforrest', 'edgeforrest', 'attendantdevops']}}) policy_json = construct_policy(pipeline_settings=settings) policy = json.loads(policy_json)
<commit_before>"""Test IAM Policies for correctness.""" import json from foremast.iam.construct_policy import construct_policy ANSWER1 = { 'Version': '2012-10-17', 'Statement': [ { 'Effect': 'Allow', 'Action': [ 's3:GetObject', 's3:ListObject' ], 'Resource': [ 'arn:aws:s3:::archaius-stage/forrest/unicornforrest', 'arn:aws:s3:::archaius-stage/forrest/unicornforrest/*' ] } ] } def test_main(): """Check general assemblage.""" settings = {'services': {'s3': True}} policy_json = construct_policy(app='unicornforrest', env='stage', group='forrest', pipeline_settings=settings) assert json.loads(policy_json) == ANSWER1 # TODO: Test other services besides S3 settings.update({'services': {'dynamodb': ['coreforrest', 'edgeforrest', 'attendantdevops']}}) policy_json = construct_policy(pipeline_settings=settings) policy = json.loads(policy_json) <commit_msg>tests: Test IAM Policy with no "services" See also: PSOBAT-1482<commit_after>
"""Test IAM Policies for correctness.""" import json from foremast.iam.construct_policy import construct_policy ANSWER1 = { 'Version': '2012-10-17', 'Statement': [ { 'Effect': 'Allow', 'Action': [ 's3:GetObject', 's3:ListObject' ], 'Resource': [ 'arn:aws:s3:::archaius-stage/forrest/unicornforrest', 'arn:aws:s3:::archaius-stage/forrest/unicornforrest/*' ] } ] } def test_main(): """Check general assemblage.""" settings = {} policy_json = construct_policy(pipeline_settings=settings) assert json.loads(policy_json) == {} settings = {'services': {'s3': True}} policy_json = construct_policy(app='unicornforrest', env='stage', group='forrest', pipeline_settings=settings) assert json.loads(policy_json) == ANSWER1 # TODO: Test other services besides S3 settings.update({'services': {'dynamodb': ['coreforrest', 'edgeforrest', 'attendantdevops']}}) policy_json = construct_policy(pipeline_settings=settings) policy = json.loads(policy_json)
"""Test IAM Policies for correctness.""" import json from foremast.iam.construct_policy import construct_policy ANSWER1 = { 'Version': '2012-10-17', 'Statement': [ { 'Effect': 'Allow', 'Action': [ 's3:GetObject', 's3:ListObject' ], 'Resource': [ 'arn:aws:s3:::archaius-stage/forrest/unicornforrest', 'arn:aws:s3:::archaius-stage/forrest/unicornforrest/*' ] } ] } def test_main(): """Check general assemblage.""" settings = {'services': {'s3': True}} policy_json = construct_policy(app='unicornforrest', env='stage', group='forrest', pipeline_settings=settings) assert json.loads(policy_json) == ANSWER1 # TODO: Test other services besides S3 settings.update({'services': {'dynamodb': ['coreforrest', 'edgeforrest', 'attendantdevops']}}) policy_json = construct_policy(pipeline_settings=settings) policy = json.loads(policy_json) tests: Test IAM Policy with no "services" See also: PSOBAT-1482"""Test IAM Policies for correctness.""" import json from foremast.iam.construct_policy import construct_policy ANSWER1 = { 'Version': '2012-10-17', 'Statement': [ { 'Effect': 'Allow', 'Action': [ 's3:GetObject', 's3:ListObject' ], 'Resource': [ 'arn:aws:s3:::archaius-stage/forrest/unicornforrest', 'arn:aws:s3:::archaius-stage/forrest/unicornforrest/*' ] } ] } def test_main(): """Check general assemblage.""" settings = {} policy_json = construct_policy(pipeline_settings=settings) assert json.loads(policy_json) == {} settings = {'services': {'s3': True}} policy_json = construct_policy(app='unicornforrest', env='stage', group='forrest', pipeline_settings=settings) assert json.loads(policy_json) == ANSWER1 # TODO: Test other services besides S3 settings.update({'services': {'dynamodb': ['coreforrest', 'edgeforrest', 'attendantdevops']}}) policy_json = construct_policy(pipeline_settings=settings) policy = json.loads(policy_json)
<commit_before>"""Test IAM Policies for correctness.""" import json from foremast.iam.construct_policy import construct_policy ANSWER1 = { 'Version': '2012-10-17', 'Statement': [ { 'Effect': 'Allow', 'Action': [ 's3:GetObject', 's3:ListObject' ], 'Resource': [ 'arn:aws:s3:::archaius-stage/forrest/unicornforrest', 'arn:aws:s3:::archaius-stage/forrest/unicornforrest/*' ] } ] } def test_main(): """Check general assemblage.""" settings = {'services': {'s3': True}} policy_json = construct_policy(app='unicornforrest', env='stage', group='forrest', pipeline_settings=settings) assert json.loads(policy_json) == ANSWER1 # TODO: Test other services besides S3 settings.update({'services': {'dynamodb': ['coreforrest', 'edgeforrest', 'attendantdevops']}}) policy_json = construct_policy(pipeline_settings=settings) policy = json.loads(policy_json) <commit_msg>tests: Test IAM Policy with no "services" See also: PSOBAT-1482<commit_after>"""Test IAM Policies for correctness.""" import json from foremast.iam.construct_policy import construct_policy ANSWER1 = { 'Version': '2012-10-17', 'Statement': [ { 'Effect': 'Allow', 'Action': [ 's3:GetObject', 's3:ListObject' ], 'Resource': [ 'arn:aws:s3:::archaius-stage/forrest/unicornforrest', 'arn:aws:s3:::archaius-stage/forrest/unicornforrest/*' ] } ] } def test_main(): """Check general assemblage.""" settings = {} policy_json = construct_policy(pipeline_settings=settings) assert json.loads(policy_json) == {} settings = {'services': {'s3': True}} policy_json = construct_policy(app='unicornforrest', env='stage', group='forrest', pipeline_settings=settings) assert json.loads(policy_json) == ANSWER1 # TODO: Test other services besides S3 settings.update({'services': {'dynamodb': ['coreforrest', 'edgeforrest', 'attendantdevops']}}) policy_json = construct_policy(pipeline_settings=settings) policy = json.loads(policy_json)
260daaad18e4889c0e468befd46c38d02bb1316a
tests/test_py35/test_client.py
tests/test_py35/test_client.py
import aiohttp async def test_async_with_session(loop): async with aiohttp.ClientSession(loop=loop) as session: pass assert session.closed
from contextlib import suppress import aiohttp from aiohttp import web async def test_async_with_session(loop): async with aiohttp.ClientSession(loop=loop) as session: pass assert session.closed async def test_close_resp_on_error_async_with_session(loop, test_server): async def handler(request): return web.Response() app = web.Application(loop=loop) app.router.add_get('/', handler) server = await test_server(app) async with aiohttp.ClientSession(loop=loop) as session: with suppress(RuntimeError): async with session.get(server.make_url('/')) as resp: resp.content.set_exception(RuntimeError()) await resp.read() assert len(session._connector._conns) == 0 async def test_release_resp_on_normal_exit_from_cm(loop, test_server): async def handler(request): return web.Response() app = web.Application(loop=loop) app.router.add_get('/', handler) server = await test_server(app) async with aiohttp.ClientSession(loop=loop) as session: async with session.get(server.make_url('/')) as resp: await resp.read() assert len(session._connector._conns) == 1
Add tests on closing connection by error
Add tests on closing connection by error
Python
apache-2.0
AraHaanOrg/aiohttp,rutsky/aiohttp,juliatem/aiohttp,singulared/aiohttp,moden-py/aiohttp,singulared/aiohttp,hellysmile/aiohttp,z2v/aiohttp,alex-eri/aiohttp-1,z2v/aiohttp,moden-py/aiohttp,alex-eri/aiohttp-1,z2v/aiohttp,arthurdarcet/aiohttp,KeepSafe/aiohttp,arthurdarcet/aiohttp,KeepSafe/aiohttp,rutsky/aiohttp,arthurdarcet/aiohttp,hellysmile/aiohttp,singulared/aiohttp,pfreixes/aiohttp,KeepSafe/aiohttp,Eyepea/aiohttp,alex-eri/aiohttp-1,juliatem/aiohttp,moden-py/aiohttp,AraHaanOrg/aiohttp,pfreixes/aiohttp,playpauseandstop/aiohttp,rutsky/aiohttp
import aiohttp async def test_async_with_session(loop): async with aiohttp.ClientSession(loop=loop) as session: pass assert session.closed Add tests on closing connection by error
from contextlib import suppress import aiohttp from aiohttp import web async def test_async_with_session(loop): async with aiohttp.ClientSession(loop=loop) as session: pass assert session.closed async def test_close_resp_on_error_async_with_session(loop, test_server): async def handler(request): return web.Response() app = web.Application(loop=loop) app.router.add_get('/', handler) server = await test_server(app) async with aiohttp.ClientSession(loop=loop) as session: with suppress(RuntimeError): async with session.get(server.make_url('/')) as resp: resp.content.set_exception(RuntimeError()) await resp.read() assert len(session._connector._conns) == 0 async def test_release_resp_on_normal_exit_from_cm(loop, test_server): async def handler(request): return web.Response() app = web.Application(loop=loop) app.router.add_get('/', handler) server = await test_server(app) async with aiohttp.ClientSession(loop=loop) as session: async with session.get(server.make_url('/')) as resp: await resp.read() assert len(session._connector._conns) == 1
<commit_before>import aiohttp async def test_async_with_session(loop): async with aiohttp.ClientSession(loop=loop) as session: pass assert session.closed <commit_msg>Add tests on closing connection by error<commit_after>
from contextlib import suppress import aiohttp from aiohttp import web async def test_async_with_session(loop): async with aiohttp.ClientSession(loop=loop) as session: pass assert session.closed async def test_close_resp_on_error_async_with_session(loop, test_server): async def handler(request): return web.Response() app = web.Application(loop=loop) app.router.add_get('/', handler) server = await test_server(app) async with aiohttp.ClientSession(loop=loop) as session: with suppress(RuntimeError): async with session.get(server.make_url('/')) as resp: resp.content.set_exception(RuntimeError()) await resp.read() assert len(session._connector._conns) == 0 async def test_release_resp_on_normal_exit_from_cm(loop, test_server): async def handler(request): return web.Response() app = web.Application(loop=loop) app.router.add_get('/', handler) server = await test_server(app) async with aiohttp.ClientSession(loop=loop) as session: async with session.get(server.make_url('/')) as resp: await resp.read() assert len(session._connector._conns) == 1
import aiohttp async def test_async_with_session(loop): async with aiohttp.ClientSession(loop=loop) as session: pass assert session.closed Add tests on closing connection by errorfrom contextlib import suppress import aiohttp from aiohttp import web async def test_async_with_session(loop): async with aiohttp.ClientSession(loop=loop) as session: pass assert session.closed async def test_close_resp_on_error_async_with_session(loop, test_server): async def handler(request): return web.Response() app = web.Application(loop=loop) app.router.add_get('/', handler) server = await test_server(app) async with aiohttp.ClientSession(loop=loop) as session: with suppress(RuntimeError): async with session.get(server.make_url('/')) as resp: resp.content.set_exception(RuntimeError()) await resp.read() assert len(session._connector._conns) == 0 async def test_release_resp_on_normal_exit_from_cm(loop, test_server): async def handler(request): return web.Response() app = web.Application(loop=loop) app.router.add_get('/', handler) server = await test_server(app) async with aiohttp.ClientSession(loop=loop) as session: async with session.get(server.make_url('/')) as resp: await resp.read() assert len(session._connector._conns) == 1
<commit_before>import aiohttp async def test_async_with_session(loop): async with aiohttp.ClientSession(loop=loop) as session: pass assert session.closed <commit_msg>Add tests on closing connection by error<commit_after>from contextlib import suppress import aiohttp from aiohttp import web async def test_async_with_session(loop): async with aiohttp.ClientSession(loop=loop) as session: pass assert session.closed async def test_close_resp_on_error_async_with_session(loop, test_server): async def handler(request): return web.Response() app = web.Application(loop=loop) app.router.add_get('/', handler) server = await test_server(app) async with aiohttp.ClientSession(loop=loop) as session: with suppress(RuntimeError): async with session.get(server.make_url('/')) as resp: resp.content.set_exception(RuntimeError()) await resp.read() assert len(session._connector._conns) == 0 async def test_release_resp_on_normal_exit_from_cm(loop, test_server): async def handler(request): return web.Response() app = web.Application(loop=loop) app.router.add_get('/', handler) server = await test_server(app) async with aiohttp.ClientSession(loop=loop) as session: async with session.get(server.make_url('/')) as resp: await resp.read() assert len(session._connector._conns) == 1
69d0cf6cc0d19f1669f56a361447935e375ac05c
indico/modules/events/logs/views.py
indico/modules/events/logs/views.py
# This file is part of Indico. # Copyright (C) 2002 - 2018 European Organization for Nuclear Research (CERN). # # Indico is free software; you can redistribute it and/or # modify it under the terms of the GNU General Public License as # published by the Free Software Foundation; either version 3 of the # License, or (at your option) any later version. # # Indico is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # General Public License for more details. # # You should have received a copy of the GNU General Public License # along with Indico; if not, see <http://www.gnu.org/licenses/>. from __future__ import unicode_literals from indico.modules.events.management.views import WPEventManagement class WPEventLogs(WPEventManagement): bundles = ('react.js', 'module_events.logs.js', 'module_events.logs.css') template_prefix = 'events/logs/' sidemenu_option = 'logs'
# This file is part of Indico. # Copyright (C) 2002 - 2018 European Organization for Nuclear Research (CERN). # # Indico is free software; you can redistribute it and/or # modify it under the terms of the GNU General Public License as # published by the Free Software Foundation; either version 3 of the # License, or (at your option) any later version. # # Indico is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # General Public License for more details. # # You should have received a copy of the GNU General Public License # along with Indico; if not, see <http://www.gnu.org/licenses/>. from __future__ import unicode_literals from indico.modules.events.management.views import WPEventManagement class WPEventLogs(WPEventManagement): bundles = ('react.js', 'semantic-ui.js', 'module_events.logs.js', 'module_events.logs.css') template_prefix = 'events/logs/' sidemenu_option = 'logs'
Include SUIR JS on logs page
Include SUIR JS on logs page This is not pretty, as we don't even use SUIR there, but indico/utils/redux imports a module that imports SUIR and thus breaks the logs page if SUIR is not included.
Python
mit
DirkHoffmann/indico,indico/indico,OmeGak/indico,pferreir/indico,ThiefMaster/indico,ThiefMaster/indico,DirkHoffmann/indico,mvidalgarcia/indico,mvidalgarcia/indico,pferreir/indico,indico/indico,OmeGak/indico,mvidalgarcia/indico,indico/indico,pferreir/indico,ThiefMaster/indico,mic4ael/indico,indico/indico,DirkHoffmann/indico,mic4ael/indico,pferreir/indico,OmeGak/indico,ThiefMaster/indico,DirkHoffmann/indico,OmeGak/indico,mvidalgarcia/indico,mic4ael/indico,mic4ael/indico
# This file is part of Indico. # Copyright (C) 2002 - 2018 European Organization for Nuclear Research (CERN). # # Indico is free software; you can redistribute it and/or # modify it under the terms of the GNU General Public License as # published by the Free Software Foundation; either version 3 of the # License, or (at your option) any later version. # # Indico is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # General Public License for more details. # # You should have received a copy of the GNU General Public License # along with Indico; if not, see <http://www.gnu.org/licenses/>. from __future__ import unicode_literals from indico.modules.events.management.views import WPEventManagement class WPEventLogs(WPEventManagement): bundles = ('react.js', 'module_events.logs.js', 'module_events.logs.css') template_prefix = 'events/logs/' sidemenu_option = 'logs' Include SUIR JS on logs page This is not pretty, as we don't even use SUIR there, but indico/utils/redux imports a module that imports SUIR and thus breaks the logs page if SUIR is not included.
# This file is part of Indico. # Copyright (C) 2002 - 2018 European Organization for Nuclear Research (CERN). # # Indico is free software; you can redistribute it and/or # modify it under the terms of the GNU General Public License as # published by the Free Software Foundation; either version 3 of the # License, or (at your option) any later version. # # Indico is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # General Public License for more details. # # You should have received a copy of the GNU General Public License # along with Indico; if not, see <http://www.gnu.org/licenses/>. from __future__ import unicode_literals from indico.modules.events.management.views import WPEventManagement class WPEventLogs(WPEventManagement): bundles = ('react.js', 'semantic-ui.js', 'module_events.logs.js', 'module_events.logs.css') template_prefix = 'events/logs/' sidemenu_option = 'logs'
<commit_before># This file is part of Indico. # Copyright (C) 2002 - 2018 European Organization for Nuclear Research (CERN). # # Indico is free software; you can redistribute it and/or # modify it under the terms of the GNU General Public License as # published by the Free Software Foundation; either version 3 of the # License, or (at your option) any later version. # # Indico is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # General Public License for more details. # # You should have received a copy of the GNU General Public License # along with Indico; if not, see <http://www.gnu.org/licenses/>. from __future__ import unicode_literals from indico.modules.events.management.views import WPEventManagement class WPEventLogs(WPEventManagement): bundles = ('react.js', 'module_events.logs.js', 'module_events.logs.css') template_prefix = 'events/logs/' sidemenu_option = 'logs' <commit_msg>Include SUIR JS on logs page This is not pretty, as we don't even use SUIR there, but indico/utils/redux imports a module that imports SUIR and thus breaks the logs page if SUIR is not included.<commit_after>
# This file is part of Indico. # Copyright (C) 2002 - 2018 European Organization for Nuclear Research (CERN). # # Indico is free software; you can redistribute it and/or # modify it under the terms of the GNU General Public License as # published by the Free Software Foundation; either version 3 of the # License, or (at your option) any later version. # # Indico is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # General Public License for more details. # # You should have received a copy of the GNU General Public License # along with Indico; if not, see <http://www.gnu.org/licenses/>. from __future__ import unicode_literals from indico.modules.events.management.views import WPEventManagement class WPEventLogs(WPEventManagement): bundles = ('react.js', 'semantic-ui.js', 'module_events.logs.js', 'module_events.logs.css') template_prefix = 'events/logs/' sidemenu_option = 'logs'
# This file is part of Indico. # Copyright (C) 2002 - 2018 European Organization for Nuclear Research (CERN). # # Indico is free software; you can redistribute it and/or # modify it under the terms of the GNU General Public License as # published by the Free Software Foundation; either version 3 of the # License, or (at your option) any later version. # # Indico is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # General Public License for more details. # # You should have received a copy of the GNU General Public License # along with Indico; if not, see <http://www.gnu.org/licenses/>. from __future__ import unicode_literals from indico.modules.events.management.views import WPEventManagement class WPEventLogs(WPEventManagement): bundles = ('react.js', 'module_events.logs.js', 'module_events.logs.css') template_prefix = 'events/logs/' sidemenu_option = 'logs' Include SUIR JS on logs page This is not pretty, as we don't even use SUIR there, but indico/utils/redux imports a module that imports SUIR and thus breaks the logs page if SUIR is not included.# This file is part of Indico. # Copyright (C) 2002 - 2018 European Organization for Nuclear Research (CERN). # # Indico is free software; you can redistribute it and/or # modify it under the terms of the GNU General Public License as # published by the Free Software Foundation; either version 3 of the # License, or (at your option) any later version. # # Indico is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # General Public License for more details. # # You should have received a copy of the GNU General Public License # along with Indico; if not, see <http://www.gnu.org/licenses/>. from __future__ import unicode_literals from indico.modules.events.management.views import WPEventManagement class WPEventLogs(WPEventManagement): bundles = ('react.js', 'semantic-ui.js', 'module_events.logs.js', 'module_events.logs.css') template_prefix = 'events/logs/' sidemenu_option = 'logs'
<commit_before># This file is part of Indico. # Copyright (C) 2002 - 2018 European Organization for Nuclear Research (CERN). # # Indico is free software; you can redistribute it and/or # modify it under the terms of the GNU General Public License as # published by the Free Software Foundation; either version 3 of the # License, or (at your option) any later version. # # Indico is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # General Public License for more details. # # You should have received a copy of the GNU General Public License # along with Indico; if not, see <http://www.gnu.org/licenses/>. from __future__ import unicode_literals from indico.modules.events.management.views import WPEventManagement class WPEventLogs(WPEventManagement): bundles = ('react.js', 'module_events.logs.js', 'module_events.logs.css') template_prefix = 'events/logs/' sidemenu_option = 'logs' <commit_msg>Include SUIR JS on logs page This is not pretty, as we don't even use SUIR there, but indico/utils/redux imports a module that imports SUIR and thus breaks the logs page if SUIR is not included.<commit_after># This file is part of Indico. # Copyright (C) 2002 - 2018 European Organization for Nuclear Research (CERN). # # Indico is free software; you can redistribute it and/or # modify it under the terms of the GNU General Public License as # published by the Free Software Foundation; either version 3 of the # License, or (at your option) any later version. # # Indico is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # General Public License for more details. # # You should have received a copy of the GNU General Public License # along with Indico; if not, see <http://www.gnu.org/licenses/>. from __future__ import unicode_literals from indico.modules.events.management.views import WPEventManagement class WPEventLogs(WPEventManagement): bundles = ('react.js', 'semantic-ui.js', 'module_events.logs.js', 'module_events.logs.css') template_prefix = 'events/logs/' sidemenu_option = 'logs'
839f9edc811776b8898cdf1fa7116eec9aef50a7
tests/xmlsec/test_templates.py
tests/xmlsec/test_templates.py
import xmlsec def test_create_signature_template(): node = xmlsec.create_signature_template() assert node.tag.endswith('Signature') assert node.xpath('*[local-name() = "SignatureValue"]') assert node.xpath('*[local-name() = "SignedInfo"]') return node def test_add_reference(): node = test_create_signature_template() ref = xmlsec.add_reference(node, uri=b'#_34275907093489075620748690') assert ref.tag.endswith('Reference') assert node.xpath('.//*[local-name() = "Reference"]')
import xmlsec def test_create_signature_template(): node = xmlsec.create_signature_template() assert node.tag.endswith('Signature') assert node.xpath('*[local-name() = "SignatureValue"]') assert node.xpath('*[local-name() = "SignedInfo"]') def test_add_reference(): node = xmlsec.create_signature_template() ref = xmlsec.add_reference(node, uri=b'#_34275907093489075620748690') assert ref.tag.endswith('Reference') assert node.xpath('.//*[local-name() = "Reference"]') def test_add_transform(): node = xmlsec.create_signature_template() ref = xmlsec.add_reference(node, uri=b'#_34275907093489075620748690') xmlsec.add_transform(ref, xmlsec.method.ENVELOPED) assert ref.xpath('.//*[local-name() = "Transform"]') def test_ensure_key_info(): node = xmlsec.create_signature_template() xmlsec.ensure_key_info(node) assert node.xpath('.//*[local-name() = "KeyInfo"]') def test_add_x509_data(): node = xmlsec.create_signature_template() info = xmlsec.ensure_key_info(node) xmlsec.add_x509_data(info) assert node.xpath('.//*[local-name() = "X509Data"]') def test_add_key_name(): node = xmlsec.create_signature_template() info = xmlsec.ensure_key_info(node) xmlsec.add_key_name(info, b'bob.pem') assert node.xpath('.//*[local-name() = "KeyName" and text() = "bob.pem"]')
Add additional tests for templates.
Add additional tests for templates.
Python
mit
devsisters/python-xmlsec,concordusapps/python-xmlsec,mehcode/python-xmlsec,devsisters/python-xmlsec,mehcode/python-xmlsec,concordusapps/python-xmlsec
import xmlsec def test_create_signature_template(): node = xmlsec.create_signature_template() assert node.tag.endswith('Signature') assert node.xpath('*[local-name() = "SignatureValue"]') assert node.xpath('*[local-name() = "SignedInfo"]') return node def test_add_reference(): node = test_create_signature_template() ref = xmlsec.add_reference(node, uri=b'#_34275907093489075620748690') assert ref.tag.endswith('Reference') assert node.xpath('.//*[local-name() = "Reference"]') Add additional tests for templates.
import xmlsec def test_create_signature_template(): node = xmlsec.create_signature_template() assert node.tag.endswith('Signature') assert node.xpath('*[local-name() = "SignatureValue"]') assert node.xpath('*[local-name() = "SignedInfo"]') def test_add_reference(): node = xmlsec.create_signature_template() ref = xmlsec.add_reference(node, uri=b'#_34275907093489075620748690') assert ref.tag.endswith('Reference') assert node.xpath('.//*[local-name() = "Reference"]') def test_add_transform(): node = xmlsec.create_signature_template() ref = xmlsec.add_reference(node, uri=b'#_34275907093489075620748690') xmlsec.add_transform(ref, xmlsec.method.ENVELOPED) assert ref.xpath('.//*[local-name() = "Transform"]') def test_ensure_key_info(): node = xmlsec.create_signature_template() xmlsec.ensure_key_info(node) assert node.xpath('.//*[local-name() = "KeyInfo"]') def test_add_x509_data(): node = xmlsec.create_signature_template() info = xmlsec.ensure_key_info(node) xmlsec.add_x509_data(info) assert node.xpath('.//*[local-name() = "X509Data"]') def test_add_key_name(): node = xmlsec.create_signature_template() info = xmlsec.ensure_key_info(node) xmlsec.add_key_name(info, b'bob.pem') assert node.xpath('.//*[local-name() = "KeyName" and text() = "bob.pem"]')
<commit_before>import xmlsec def test_create_signature_template(): node = xmlsec.create_signature_template() assert node.tag.endswith('Signature') assert node.xpath('*[local-name() = "SignatureValue"]') assert node.xpath('*[local-name() = "SignedInfo"]') return node def test_add_reference(): node = test_create_signature_template() ref = xmlsec.add_reference(node, uri=b'#_34275907093489075620748690') assert ref.tag.endswith('Reference') assert node.xpath('.//*[local-name() = "Reference"]') <commit_msg>Add additional tests for templates.<commit_after>
import xmlsec def test_create_signature_template(): node = xmlsec.create_signature_template() assert node.tag.endswith('Signature') assert node.xpath('*[local-name() = "SignatureValue"]') assert node.xpath('*[local-name() = "SignedInfo"]') def test_add_reference(): node = xmlsec.create_signature_template() ref = xmlsec.add_reference(node, uri=b'#_34275907093489075620748690') assert ref.tag.endswith('Reference') assert node.xpath('.//*[local-name() = "Reference"]') def test_add_transform(): node = xmlsec.create_signature_template() ref = xmlsec.add_reference(node, uri=b'#_34275907093489075620748690') xmlsec.add_transform(ref, xmlsec.method.ENVELOPED) assert ref.xpath('.//*[local-name() = "Transform"]') def test_ensure_key_info(): node = xmlsec.create_signature_template() xmlsec.ensure_key_info(node) assert node.xpath('.//*[local-name() = "KeyInfo"]') def test_add_x509_data(): node = xmlsec.create_signature_template() info = xmlsec.ensure_key_info(node) xmlsec.add_x509_data(info) assert node.xpath('.//*[local-name() = "X509Data"]') def test_add_key_name(): node = xmlsec.create_signature_template() info = xmlsec.ensure_key_info(node) xmlsec.add_key_name(info, b'bob.pem') assert node.xpath('.//*[local-name() = "KeyName" and text() = "bob.pem"]')
import xmlsec def test_create_signature_template(): node = xmlsec.create_signature_template() assert node.tag.endswith('Signature') assert node.xpath('*[local-name() = "SignatureValue"]') assert node.xpath('*[local-name() = "SignedInfo"]') return node def test_add_reference(): node = test_create_signature_template() ref = xmlsec.add_reference(node, uri=b'#_34275907093489075620748690') assert ref.tag.endswith('Reference') assert node.xpath('.//*[local-name() = "Reference"]') Add additional tests for templates.import xmlsec def test_create_signature_template(): node = xmlsec.create_signature_template() assert node.tag.endswith('Signature') assert node.xpath('*[local-name() = "SignatureValue"]') assert node.xpath('*[local-name() = "SignedInfo"]') def test_add_reference(): node = xmlsec.create_signature_template() ref = xmlsec.add_reference(node, uri=b'#_34275907093489075620748690') assert ref.tag.endswith('Reference') assert node.xpath('.//*[local-name() = "Reference"]') def test_add_transform(): node = xmlsec.create_signature_template() ref = xmlsec.add_reference(node, uri=b'#_34275907093489075620748690') xmlsec.add_transform(ref, xmlsec.method.ENVELOPED) assert ref.xpath('.//*[local-name() = "Transform"]') def test_ensure_key_info(): node = xmlsec.create_signature_template() xmlsec.ensure_key_info(node) assert node.xpath('.//*[local-name() = "KeyInfo"]') def test_add_x509_data(): node = xmlsec.create_signature_template() info = xmlsec.ensure_key_info(node) xmlsec.add_x509_data(info) assert node.xpath('.//*[local-name() = "X509Data"]') def test_add_key_name(): node = xmlsec.create_signature_template() info = xmlsec.ensure_key_info(node) xmlsec.add_key_name(info, b'bob.pem') assert node.xpath('.//*[local-name() = "KeyName" and text() = "bob.pem"]')
<commit_before>import xmlsec def test_create_signature_template(): node = xmlsec.create_signature_template() assert node.tag.endswith('Signature') assert node.xpath('*[local-name() = "SignatureValue"]') assert node.xpath('*[local-name() = "SignedInfo"]') return node def test_add_reference(): node = test_create_signature_template() ref = xmlsec.add_reference(node, uri=b'#_34275907093489075620748690') assert ref.tag.endswith('Reference') assert node.xpath('.//*[local-name() = "Reference"]') <commit_msg>Add additional tests for templates.<commit_after>import xmlsec def test_create_signature_template(): node = xmlsec.create_signature_template() assert node.tag.endswith('Signature') assert node.xpath('*[local-name() = "SignatureValue"]') assert node.xpath('*[local-name() = "SignedInfo"]') def test_add_reference(): node = xmlsec.create_signature_template() ref = xmlsec.add_reference(node, uri=b'#_34275907093489075620748690') assert ref.tag.endswith('Reference') assert node.xpath('.//*[local-name() = "Reference"]') def test_add_transform(): node = xmlsec.create_signature_template() ref = xmlsec.add_reference(node, uri=b'#_34275907093489075620748690') xmlsec.add_transform(ref, xmlsec.method.ENVELOPED) assert ref.xpath('.//*[local-name() = "Transform"]') def test_ensure_key_info(): node = xmlsec.create_signature_template() xmlsec.ensure_key_info(node) assert node.xpath('.//*[local-name() = "KeyInfo"]') def test_add_x509_data(): node = xmlsec.create_signature_template() info = xmlsec.ensure_key_info(node) xmlsec.add_x509_data(info) assert node.xpath('.//*[local-name() = "X509Data"]') def test_add_key_name(): node = xmlsec.create_signature_template() info = xmlsec.ensure_key_info(node) xmlsec.add_key_name(info, b'bob.pem') assert node.xpath('.//*[local-name() = "KeyName" and text() = "bob.pem"]')
0d48a418787e5724078d3821d28c639f0a76c8b2
src/robot/htmldata/normaltemplate.py
src/robot/htmldata/normaltemplate.py
# Copyright 2008-2015 Nokia Networks # Copyright 2016- Robot Framework Foundation # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import os from os.path import abspath, dirname, join, normpath class HtmlTemplate(object): _base_dir = join(dirname(abspath(__file__)), '..', 'htmldata') def __init__(self, filename): self._path = normpath(join(self._base_dir, filename.replace('/', os.sep))) def __iter__(self): with open(self._path) as file: for line in file: yield line.rstrip()
# Copyright 2008-2015 Nokia Networks # Copyright 2016- Robot Framework Foundation # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import codecs import os from os.path import abspath, dirname, join, normpath class HtmlTemplate(object): _base_dir = join(dirname(abspath(__file__)), '..', 'htmldata') def __init__(self, filename): self._path = normpath(join(self._base_dir, filename.replace('/', os.sep))) def __iter__(self): with codecs.open(self._path, encoding='UTF-8') as file: for line in file: yield line.rstrip()
Fix reading log/report templates/dependencies in UTF-8.
Fix reading log/report templates/dependencies in UTF-8. Some of the updated js dependencies (#2419) had non-ASCII data in UTF-8 format but our code reading these files didn't take that into account.
Python
apache-2.0
robotframework/robotframework,robotframework/robotframework,HelioGuilherme66/robotframework,HelioGuilherme66/robotframework,robotframework/robotframework,HelioGuilherme66/robotframework
# Copyright 2008-2015 Nokia Networks # Copyright 2016- Robot Framework Foundation # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import os from os.path import abspath, dirname, join, normpath class HtmlTemplate(object): _base_dir = join(dirname(abspath(__file__)), '..', 'htmldata') def __init__(self, filename): self._path = normpath(join(self._base_dir, filename.replace('/', os.sep))) def __iter__(self): with open(self._path) as file: for line in file: yield line.rstrip() Fix reading log/report templates/dependencies in UTF-8. Some of the updated js dependencies (#2419) had non-ASCII data in UTF-8 format but our code reading these files didn't take that into account.
# Copyright 2008-2015 Nokia Networks # Copyright 2016- Robot Framework Foundation # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import codecs import os from os.path import abspath, dirname, join, normpath class HtmlTemplate(object): _base_dir = join(dirname(abspath(__file__)), '..', 'htmldata') def __init__(self, filename): self._path = normpath(join(self._base_dir, filename.replace('/', os.sep))) def __iter__(self): with codecs.open(self._path, encoding='UTF-8') as file: for line in file: yield line.rstrip()
<commit_before># Copyright 2008-2015 Nokia Networks # Copyright 2016- Robot Framework Foundation # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import os from os.path import abspath, dirname, join, normpath class HtmlTemplate(object): _base_dir = join(dirname(abspath(__file__)), '..', 'htmldata') def __init__(self, filename): self._path = normpath(join(self._base_dir, filename.replace('/', os.sep))) def __iter__(self): with open(self._path) as file: for line in file: yield line.rstrip() <commit_msg>Fix reading log/report templates/dependencies in UTF-8. Some of the updated js dependencies (#2419) had non-ASCII data in UTF-8 format but our code reading these files didn't take that into account.<commit_after>
# Copyright 2008-2015 Nokia Networks # Copyright 2016- Robot Framework Foundation # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import codecs import os from os.path import abspath, dirname, join, normpath class HtmlTemplate(object): _base_dir = join(dirname(abspath(__file__)), '..', 'htmldata') def __init__(self, filename): self._path = normpath(join(self._base_dir, filename.replace('/', os.sep))) def __iter__(self): with codecs.open(self._path, encoding='UTF-8') as file: for line in file: yield line.rstrip()
# Copyright 2008-2015 Nokia Networks # Copyright 2016- Robot Framework Foundation # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import os from os.path import abspath, dirname, join, normpath class HtmlTemplate(object): _base_dir = join(dirname(abspath(__file__)), '..', 'htmldata') def __init__(self, filename): self._path = normpath(join(self._base_dir, filename.replace('/', os.sep))) def __iter__(self): with open(self._path) as file: for line in file: yield line.rstrip() Fix reading log/report templates/dependencies in UTF-8. Some of the updated js dependencies (#2419) had non-ASCII data in UTF-8 format but our code reading these files didn't take that into account.# Copyright 2008-2015 Nokia Networks # Copyright 2016- Robot Framework Foundation # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import codecs import os from os.path import abspath, dirname, join, normpath class HtmlTemplate(object): _base_dir = join(dirname(abspath(__file__)), '..', 'htmldata') def __init__(self, filename): self._path = normpath(join(self._base_dir, filename.replace('/', os.sep))) def __iter__(self): with codecs.open(self._path, encoding='UTF-8') as file: for line in file: yield line.rstrip()
<commit_before># Copyright 2008-2015 Nokia Networks # Copyright 2016- Robot Framework Foundation # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import os from os.path import abspath, dirname, join, normpath class HtmlTemplate(object): _base_dir = join(dirname(abspath(__file__)), '..', 'htmldata') def __init__(self, filename): self._path = normpath(join(self._base_dir, filename.replace('/', os.sep))) def __iter__(self): with open(self._path) as file: for line in file: yield line.rstrip() <commit_msg>Fix reading log/report templates/dependencies in UTF-8. Some of the updated js dependencies (#2419) had non-ASCII data in UTF-8 format but our code reading these files didn't take that into account.<commit_after># Copyright 2008-2015 Nokia Networks # Copyright 2016- Robot Framework Foundation # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. import codecs import os from os.path import abspath, dirname, join, normpath class HtmlTemplate(object): _base_dir = join(dirname(abspath(__file__)), '..', 'htmldata') def __init__(self, filename): self._path = normpath(join(self._base_dir, filename.replace('/', os.sep))) def __iter__(self): with codecs.open(self._path, encoding='UTF-8') as file: for line in file: yield line.rstrip()
93fa014ea7a34834ae6bb85ea802879ae0026941
keystone/common/policies/revoke_event.py
keystone/common/policies/revoke_event.py
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from oslo_policy import policy from keystone.common.policies import base revoke_event_policies = [ policy.DocumentedRuleDefault( name=base.IDENTITY % 'list_revoke_events', check_str=base.RULE_SERVICE_OR_ADMIN, description='List revocation events.', operations=[{'path': '/v3/OS-REVOKE/events', 'method': 'GET'}]) ] def list_rules(): return revoke_event_policies
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from oslo_policy import policy from keystone.common.policies import base revoke_event_policies = [ policy.DocumentedRuleDefault( name=base.IDENTITY % 'list_revoke_events', check_str=base.RULE_SERVICE_OR_ADMIN, # NOTE(lbragstad): This API was originally introduced so that services # could invalidate tokens based on revocation events. This is system # specific so it make sense to associate `system` as the scope type # required for this policy. scope_types=['system'], description='List revocation events.', operations=[{'path': '/v3/OS-REVOKE/events', 'method': 'GET'}]) ] def list_rules(): return revoke_event_policies
Add scope_types for revoke event policies
Add scope_types for revoke event policies This commit associates `system` to revoke event policies, since these policies were developed to assist the system in offline token validation. From now on, a warning will be logged when a project-scoped token is used to get revocation events. Operators can opt into requiring system-scoped tokens for these policies by enabling oslo.policy's `enforce_scope` configuration option, which will result in an HTTP Forbidden exception when mismatching scope is used. Change-Id: I1dddeb216b2523b8471e5f2d5370921bb7a45e7f
Python
apache-2.0
mahak/keystone,openstack/keystone,mahak/keystone,mahak/keystone,openstack/keystone,openstack/keystone
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from oslo_policy import policy from keystone.common.policies import base revoke_event_policies = [ policy.DocumentedRuleDefault( name=base.IDENTITY % 'list_revoke_events', check_str=base.RULE_SERVICE_OR_ADMIN, description='List revocation events.', operations=[{'path': '/v3/OS-REVOKE/events', 'method': 'GET'}]) ] def list_rules(): return revoke_event_policies Add scope_types for revoke event policies This commit associates `system` to revoke event policies, since these policies were developed to assist the system in offline token validation. From now on, a warning will be logged when a project-scoped token is used to get revocation events. Operators can opt into requiring system-scoped tokens for these policies by enabling oslo.policy's `enforce_scope` configuration option, which will result in an HTTP Forbidden exception when mismatching scope is used. Change-Id: I1dddeb216b2523b8471e5f2d5370921bb7a45e7f
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from oslo_policy import policy from keystone.common.policies import base revoke_event_policies = [ policy.DocumentedRuleDefault( name=base.IDENTITY % 'list_revoke_events', check_str=base.RULE_SERVICE_OR_ADMIN, # NOTE(lbragstad): This API was originally introduced so that services # could invalidate tokens based on revocation events. This is system # specific so it make sense to associate `system` as the scope type # required for this policy. scope_types=['system'], description='List revocation events.', operations=[{'path': '/v3/OS-REVOKE/events', 'method': 'GET'}]) ] def list_rules(): return revoke_event_policies
<commit_before># Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from oslo_policy import policy from keystone.common.policies import base revoke_event_policies = [ policy.DocumentedRuleDefault( name=base.IDENTITY % 'list_revoke_events', check_str=base.RULE_SERVICE_OR_ADMIN, description='List revocation events.', operations=[{'path': '/v3/OS-REVOKE/events', 'method': 'GET'}]) ] def list_rules(): return revoke_event_policies <commit_msg>Add scope_types for revoke event policies This commit associates `system` to revoke event policies, since these policies were developed to assist the system in offline token validation. From now on, a warning will be logged when a project-scoped token is used to get revocation events. Operators can opt into requiring system-scoped tokens for these policies by enabling oslo.policy's `enforce_scope` configuration option, which will result in an HTTP Forbidden exception when mismatching scope is used. Change-Id: I1dddeb216b2523b8471e5f2d5370921bb7a45e7f<commit_after>
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from oslo_policy import policy from keystone.common.policies import base revoke_event_policies = [ policy.DocumentedRuleDefault( name=base.IDENTITY % 'list_revoke_events', check_str=base.RULE_SERVICE_OR_ADMIN, # NOTE(lbragstad): This API was originally introduced so that services # could invalidate tokens based on revocation events. This is system # specific so it make sense to associate `system` as the scope type # required for this policy. scope_types=['system'], description='List revocation events.', operations=[{'path': '/v3/OS-REVOKE/events', 'method': 'GET'}]) ] def list_rules(): return revoke_event_policies
# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from oslo_policy import policy from keystone.common.policies import base revoke_event_policies = [ policy.DocumentedRuleDefault( name=base.IDENTITY % 'list_revoke_events', check_str=base.RULE_SERVICE_OR_ADMIN, description='List revocation events.', operations=[{'path': '/v3/OS-REVOKE/events', 'method': 'GET'}]) ] def list_rules(): return revoke_event_policies Add scope_types for revoke event policies This commit associates `system` to revoke event policies, since these policies were developed to assist the system in offline token validation. From now on, a warning will be logged when a project-scoped token is used to get revocation events. Operators can opt into requiring system-scoped tokens for these policies by enabling oslo.policy's `enforce_scope` configuration option, which will result in an HTTP Forbidden exception when mismatching scope is used. Change-Id: I1dddeb216b2523b8471e5f2d5370921bb7a45e7f# Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from oslo_policy import policy from keystone.common.policies import base revoke_event_policies = [ policy.DocumentedRuleDefault( name=base.IDENTITY % 'list_revoke_events', check_str=base.RULE_SERVICE_OR_ADMIN, # NOTE(lbragstad): This API was originally introduced so that services # could invalidate tokens based on revocation events. This is system # specific so it make sense to associate `system` as the scope type # required for this policy. scope_types=['system'], description='List revocation events.', operations=[{'path': '/v3/OS-REVOKE/events', 'method': 'GET'}]) ] def list_rules(): return revoke_event_policies
<commit_before># Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from oslo_policy import policy from keystone.common.policies import base revoke_event_policies = [ policy.DocumentedRuleDefault( name=base.IDENTITY % 'list_revoke_events', check_str=base.RULE_SERVICE_OR_ADMIN, description='List revocation events.', operations=[{'path': '/v3/OS-REVOKE/events', 'method': 'GET'}]) ] def list_rules(): return revoke_event_policies <commit_msg>Add scope_types for revoke event policies This commit associates `system` to revoke event policies, since these policies were developed to assist the system in offline token validation. From now on, a warning will be logged when a project-scoped token is used to get revocation events. Operators can opt into requiring system-scoped tokens for these policies by enabling oslo.policy's `enforce_scope` configuration option, which will result in an HTTP Forbidden exception when mismatching scope is used. Change-Id: I1dddeb216b2523b8471e5f2d5370921bb7a45e7f<commit_after># Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from oslo_policy import policy from keystone.common.policies import base revoke_event_policies = [ policy.DocumentedRuleDefault( name=base.IDENTITY % 'list_revoke_events', check_str=base.RULE_SERVICE_OR_ADMIN, # NOTE(lbragstad): This API was originally introduced so that services # could invalidate tokens based on revocation events. This is system # specific so it make sense to associate `system` as the scope type # required for this policy. scope_types=['system'], description='List revocation events.', operations=[{'path': '/v3/OS-REVOKE/events', 'method': 'GET'}]) ] def list_rules(): return revoke_event_policies
00ce59d43c4208846234652a0746f048836493f2
src/ggrc/services/signals.py
src/ggrc/services/signals.py
# Copyright (C) 2016 Google Inc., authors, and contributors <see AUTHORS file> # Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file> # Created By: urban@reciprocitylabs.com # Maintained By: urban@reciprocitylabs.com from blinker import Namespace class Signals(object): signals = Namespace() custom_attribute_changed = signals.signal( "Custom Attribute updated", """ Indicates that a custom attribute was successfully saved to database. :obj: The model instance :value: New custom attribute value :service: The instance of model handling the Custom Attribute update operation """, )
# Copyright (C) 2016 Google Inc., authors, and contributors <see AUTHORS file> # Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file> # Created By: urban@reciprocitylabs.com # Maintained By: urban@reciprocitylabs.com from blinker import Namespace class Signals(object): signals = Namespace() custom_attribute_changed = signals.signal( "Custom Attribute updated", """ Indicates that a custom attribute was successfully saved to database. :obj: The model instance :value: New custom attribute value :service: The instance of model handling the Custom Attribute update operation """, )
Fix new CA signal message
Fix new CA signal message
Python
apache-2.0
selahssea/ggrc-core,josthkko/ggrc-core,edofic/ggrc-core,andrei-karalionak/ggrc-core,VinnieJohns/ggrc-core,selahssea/ggrc-core,josthkko/ggrc-core,selahssea/ggrc-core,plamut/ggrc-core,selahssea/ggrc-core,NejcZupec/ggrc-core,edofic/ggrc-core,j0gurt/ggrc-core,j0gurt/ggrc-core,plamut/ggrc-core,plamut/ggrc-core,kr41/ggrc-core,VinnieJohns/ggrc-core,NejcZupec/ggrc-core,prasannav7/ggrc-core,j0gurt/ggrc-core,andrei-karalionak/ggrc-core,kr41/ggrc-core,prasannav7/ggrc-core,VinnieJohns/ggrc-core,josthkko/ggrc-core,kr41/ggrc-core,AleksNeStu/ggrc-core,NejcZupec/ggrc-core,AleksNeStu/ggrc-core,VinnieJohns/ggrc-core,AleksNeStu/ggrc-core,plamut/ggrc-core,j0gurt/ggrc-core,andrei-karalionak/ggrc-core,edofic/ggrc-core,prasannav7/ggrc-core,kr41/ggrc-core,josthkko/ggrc-core,AleksNeStu/ggrc-core,edofic/ggrc-core,andrei-karalionak/ggrc-core,prasannav7/ggrc-core,NejcZupec/ggrc-core
# Copyright (C) 2016 Google Inc., authors, and contributors <see AUTHORS file> # Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file> # Created By: urban@reciprocitylabs.com # Maintained By: urban@reciprocitylabs.com from blinker import Namespace class Signals(object): signals = Namespace() custom_attribute_changed = signals.signal( "Custom Attribute updated", """ Indicates that a custom attribute was successfully saved to database. :obj: The model instance :value: New custom attribute value :service: The instance of model handling the Custom Attribute update operation """, ) Fix new CA signal message
# Copyright (C) 2016 Google Inc., authors, and contributors <see AUTHORS file> # Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file> # Created By: urban@reciprocitylabs.com # Maintained By: urban@reciprocitylabs.com from blinker import Namespace class Signals(object): signals = Namespace() custom_attribute_changed = signals.signal( "Custom Attribute updated", """ Indicates that a custom attribute was successfully saved to database. :obj: The model instance :value: New custom attribute value :service: The instance of model handling the Custom Attribute update operation """, )
<commit_before># Copyright (C) 2016 Google Inc., authors, and contributors <see AUTHORS file> # Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file> # Created By: urban@reciprocitylabs.com # Maintained By: urban@reciprocitylabs.com from blinker import Namespace class Signals(object): signals = Namespace() custom_attribute_changed = signals.signal( "Custom Attribute updated", """ Indicates that a custom attribute was successfully saved to database. :obj: The model instance :value: New custom attribute value :service: The instance of model handling the Custom Attribute update operation """, ) <commit_msg>Fix new CA signal message<commit_after>
# Copyright (C) 2016 Google Inc., authors, and contributors <see AUTHORS file> # Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file> # Created By: urban@reciprocitylabs.com # Maintained By: urban@reciprocitylabs.com from blinker import Namespace class Signals(object): signals = Namespace() custom_attribute_changed = signals.signal( "Custom Attribute updated", """ Indicates that a custom attribute was successfully saved to database. :obj: The model instance :value: New custom attribute value :service: The instance of model handling the Custom Attribute update operation """, )
# Copyright (C) 2016 Google Inc., authors, and contributors <see AUTHORS file> # Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file> # Created By: urban@reciprocitylabs.com # Maintained By: urban@reciprocitylabs.com from blinker import Namespace class Signals(object): signals = Namespace() custom_attribute_changed = signals.signal( "Custom Attribute updated", """ Indicates that a custom attribute was successfully saved to database. :obj: The model instance :value: New custom attribute value :service: The instance of model handling the Custom Attribute update operation """, ) Fix new CA signal message# Copyright (C) 2016 Google Inc., authors, and contributors <see AUTHORS file> # Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file> # Created By: urban@reciprocitylabs.com # Maintained By: urban@reciprocitylabs.com from blinker import Namespace class Signals(object): signals = Namespace() custom_attribute_changed = signals.signal( "Custom Attribute updated", """ Indicates that a custom attribute was successfully saved to database. :obj: The model instance :value: New custom attribute value :service: The instance of model handling the Custom Attribute update operation """, )
<commit_before># Copyright (C) 2016 Google Inc., authors, and contributors <see AUTHORS file> # Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file> # Created By: urban@reciprocitylabs.com # Maintained By: urban@reciprocitylabs.com from blinker import Namespace class Signals(object): signals = Namespace() custom_attribute_changed = signals.signal( "Custom Attribute updated", """ Indicates that a custom attribute was successfully saved to database. :obj: The model instance :value: New custom attribute value :service: The instance of model handling the Custom Attribute update operation """, ) <commit_msg>Fix new CA signal message<commit_after># Copyright (C) 2016 Google Inc., authors, and contributors <see AUTHORS file> # Licensed under http://www.apache.org/licenses/LICENSE-2.0 <see LICENSE file> # Created By: urban@reciprocitylabs.com # Maintained By: urban@reciprocitylabs.com from blinker import Namespace class Signals(object): signals = Namespace() custom_attribute_changed = signals.signal( "Custom Attribute updated", """ Indicates that a custom attribute was successfully saved to database. :obj: The model instance :value: New custom attribute value :service: The instance of model handling the Custom Attribute update operation """, )
6c6934e8a36429e2a988835d8bd4d66fe95e306b
tensorflow_datasets/image/cifar_test.py
tensorflow_datasets/image/cifar_test.py
# coding=utf-8 # Copyright 2018 The TensorFlow Datasets Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Tests for cifar dataset module.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function from tensorflow_datasets.image import cifar from tensorflow_datasets.testing import dataset_builder_testing class Cifar10Test(dataset_builder_testing.TestCase): DATASET_CLASS = cifar.Cifar10 SPLITS = { "train": 10, # Number of examples. "test": 2, # See testing/generate_cifar10_like_example.py } if __name__ == "__main__": dataset_builder_testing.main()
# coding=utf-8 # Copyright 2018 The TensorFlow Datasets Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Tests for cifar dataset module.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function from tensorflow_datasets.image import cifar from tensorflow_datasets.testing import dataset_builder_testing class Cifar10Test(dataset_builder_testing.TestCase): DATASET_CLASS = cifar.Cifar10 SPLITS = { "train": 10, # Number of examples. "test": 2, # See testing/cifar10.py } if __name__ == "__main__": dataset_builder_testing.main()
Move references of deleted generate_cifar10_like_example.py to the new name cifar.py
Move references of deleted generate_cifar10_like_example.py to the new name cifar.py PiperOrigin-RevId: 225386826
Python
apache-2.0
tensorflow/datasets,tensorflow/datasets,tensorflow/datasets,tensorflow/datasets,tensorflow/datasets
# coding=utf-8 # Copyright 2018 The TensorFlow Datasets Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Tests for cifar dataset module.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function from tensorflow_datasets.image import cifar from tensorflow_datasets.testing import dataset_builder_testing class Cifar10Test(dataset_builder_testing.TestCase): DATASET_CLASS = cifar.Cifar10 SPLITS = { "train": 10, # Number of examples. "test": 2, # See testing/generate_cifar10_like_example.py } if __name__ == "__main__": dataset_builder_testing.main() Move references of deleted generate_cifar10_like_example.py to the new name cifar.py PiperOrigin-RevId: 225386826
# coding=utf-8 # Copyright 2018 The TensorFlow Datasets Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Tests for cifar dataset module.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function from tensorflow_datasets.image import cifar from tensorflow_datasets.testing import dataset_builder_testing class Cifar10Test(dataset_builder_testing.TestCase): DATASET_CLASS = cifar.Cifar10 SPLITS = { "train": 10, # Number of examples. "test": 2, # See testing/cifar10.py } if __name__ == "__main__": dataset_builder_testing.main()
<commit_before># coding=utf-8 # Copyright 2018 The TensorFlow Datasets Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Tests for cifar dataset module.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function from tensorflow_datasets.image import cifar from tensorflow_datasets.testing import dataset_builder_testing class Cifar10Test(dataset_builder_testing.TestCase): DATASET_CLASS = cifar.Cifar10 SPLITS = { "train": 10, # Number of examples. "test": 2, # See testing/generate_cifar10_like_example.py } if __name__ == "__main__": dataset_builder_testing.main() <commit_msg>Move references of deleted generate_cifar10_like_example.py to the new name cifar.py PiperOrigin-RevId: 225386826<commit_after>
# coding=utf-8 # Copyright 2018 The TensorFlow Datasets Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Tests for cifar dataset module.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function from tensorflow_datasets.image import cifar from tensorflow_datasets.testing import dataset_builder_testing class Cifar10Test(dataset_builder_testing.TestCase): DATASET_CLASS = cifar.Cifar10 SPLITS = { "train": 10, # Number of examples. "test": 2, # See testing/cifar10.py } if __name__ == "__main__": dataset_builder_testing.main()
# coding=utf-8 # Copyright 2018 The TensorFlow Datasets Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Tests for cifar dataset module.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function from tensorflow_datasets.image import cifar from tensorflow_datasets.testing import dataset_builder_testing class Cifar10Test(dataset_builder_testing.TestCase): DATASET_CLASS = cifar.Cifar10 SPLITS = { "train": 10, # Number of examples. "test": 2, # See testing/generate_cifar10_like_example.py } if __name__ == "__main__": dataset_builder_testing.main() Move references of deleted generate_cifar10_like_example.py to the new name cifar.py PiperOrigin-RevId: 225386826# coding=utf-8 # Copyright 2018 The TensorFlow Datasets Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Tests for cifar dataset module.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function from tensorflow_datasets.image import cifar from tensorflow_datasets.testing import dataset_builder_testing class Cifar10Test(dataset_builder_testing.TestCase): DATASET_CLASS = cifar.Cifar10 SPLITS = { "train": 10, # Number of examples. "test": 2, # See testing/cifar10.py } if __name__ == "__main__": dataset_builder_testing.main()
<commit_before># coding=utf-8 # Copyright 2018 The TensorFlow Datasets Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Tests for cifar dataset module.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function from tensorflow_datasets.image import cifar from tensorflow_datasets.testing import dataset_builder_testing class Cifar10Test(dataset_builder_testing.TestCase): DATASET_CLASS = cifar.Cifar10 SPLITS = { "train": 10, # Number of examples. "test": 2, # See testing/generate_cifar10_like_example.py } if __name__ == "__main__": dataset_builder_testing.main() <commit_msg>Move references of deleted generate_cifar10_like_example.py to the new name cifar.py PiperOrigin-RevId: 225386826<commit_after># coding=utf-8 # Copyright 2018 The TensorFlow Datasets Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Tests for cifar dataset module.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function from tensorflow_datasets.image import cifar from tensorflow_datasets.testing import dataset_builder_testing class Cifar10Test(dataset_builder_testing.TestCase): DATASET_CLASS = cifar.Cifar10 SPLITS = { "train": 10, # Number of examples. "test": 2, # See testing/cifar10.py } if __name__ == "__main__": dataset_builder_testing.main()
9feb84c39c6988ff31f3d7cb38852a457870111b
bin/readability_to_dr.py
bin/readability_to_dr.py
#!/usr/bin/env python3 import hashlib import json import os import sys print(sys.path) from gigacluster import Doc, dr, Tokenizer tok = Tokenizer() for root, dirs, files in os.walk(sys.argv[1]): dirs.sort() files = set(files) cluster_dr = os.path.join(root, 'cluster.dr') with open(cluster_dr, 'wb') as f_dr: writer = dr.Writer(f_dr, Doc) for filename in sorted(f for f in files if f.endswith('.readability')): dr_filename = filename.replace('.readability', '.dr') if not dr_filename in files: path = os.path.join(root, filename) with open(path) as f: try: data = json.load(f) except Exception: pass else: if not ('url' in data and 'title' in data and 'content' in data): continue id = hashlib.md5(data['url'].encode('utf8')).hexdigest() title = data.get('title', '') text = data.get('content', '') doc = Doc(id=id) title = title.encode('utf8') tok.tokenize(title, doc, 0, is_headline=True) tok.tokenize(text.encode('utf8'), doc, len(title) + 1, is_headline=False) print(doc) writer.write(doc)
#!/usr/bin/env python3 from collections import defaultdict import hashlib import json import os import sys from gigacluster import Doc, dr, Tokenizer tok = Tokenizer() stats = defaultdict(int) for root, dirs, files in os.walk(sys.argv[1]): dirs.sort() files = set(files) cluster_dr = os.path.join(root, 'cluster.dr') with open(cluster_dr, 'wb') as f_dr: writer = dr.Writer(f_dr, Doc) written = 0 for filename in sorted(f for f in files if f.endswith('.readability')): path = os.path.join(root, filename) with open(path) as f: try: data = json.load(f) except Exception: stats['bad json'] += 1 pass else: if not ('url' in data and 'title' in data and 'content' in data): stats['no url/title/content'] += 1 continue id = hashlib.md5(data['url'].encode('utf8')).hexdigest() title = data.get('title', '') text = data.get('content', '') doc = Doc(id=id) title = title.encode('utf8') tok.tokenize(title, doc, 0, is_headline=True) tok.tokenize(text.encode('utf8'), doc, len(title) + 1, is_headline=False) writer.write(doc) stats['ok'] += 1 written += 1 if not written: print('No docs written for {}'.format(cluster_dr)) os.remove(cluster_dr) print('Stats\t{}'.format(dict(stats)))
Check for empty cluster dr files.
Check for empty cluster dr files.
Python
mit
schwa-lab/gigacluster,schwa-lab/gigacluster
#!/usr/bin/env python3 import hashlib import json import os import sys print(sys.path) from gigacluster import Doc, dr, Tokenizer tok = Tokenizer() for root, dirs, files in os.walk(sys.argv[1]): dirs.sort() files = set(files) cluster_dr = os.path.join(root, 'cluster.dr') with open(cluster_dr, 'wb') as f_dr: writer = dr.Writer(f_dr, Doc) for filename in sorted(f for f in files if f.endswith('.readability')): dr_filename = filename.replace('.readability', '.dr') if not dr_filename in files: path = os.path.join(root, filename) with open(path) as f: try: data = json.load(f) except Exception: pass else: if not ('url' in data and 'title' in data and 'content' in data): continue id = hashlib.md5(data['url'].encode('utf8')).hexdigest() title = data.get('title', '') text = data.get('content', '') doc = Doc(id=id) title = title.encode('utf8') tok.tokenize(title, doc, 0, is_headline=True) tok.tokenize(text.encode('utf8'), doc, len(title) + 1, is_headline=False) print(doc) writer.write(doc) Check for empty cluster dr files.
#!/usr/bin/env python3 from collections import defaultdict import hashlib import json import os import sys from gigacluster import Doc, dr, Tokenizer tok = Tokenizer() stats = defaultdict(int) for root, dirs, files in os.walk(sys.argv[1]): dirs.sort() files = set(files) cluster_dr = os.path.join(root, 'cluster.dr') with open(cluster_dr, 'wb') as f_dr: writer = dr.Writer(f_dr, Doc) written = 0 for filename in sorted(f for f in files if f.endswith('.readability')): path = os.path.join(root, filename) with open(path) as f: try: data = json.load(f) except Exception: stats['bad json'] += 1 pass else: if not ('url' in data and 'title' in data and 'content' in data): stats['no url/title/content'] += 1 continue id = hashlib.md5(data['url'].encode('utf8')).hexdigest() title = data.get('title', '') text = data.get('content', '') doc = Doc(id=id) title = title.encode('utf8') tok.tokenize(title, doc, 0, is_headline=True) tok.tokenize(text.encode('utf8'), doc, len(title) + 1, is_headline=False) writer.write(doc) stats['ok'] += 1 written += 1 if not written: print('No docs written for {}'.format(cluster_dr)) os.remove(cluster_dr) print('Stats\t{}'.format(dict(stats)))
<commit_before>#!/usr/bin/env python3 import hashlib import json import os import sys print(sys.path) from gigacluster import Doc, dr, Tokenizer tok = Tokenizer() for root, dirs, files in os.walk(sys.argv[1]): dirs.sort() files = set(files) cluster_dr = os.path.join(root, 'cluster.dr') with open(cluster_dr, 'wb') as f_dr: writer = dr.Writer(f_dr, Doc) for filename in sorted(f for f in files if f.endswith('.readability')): dr_filename = filename.replace('.readability', '.dr') if not dr_filename in files: path = os.path.join(root, filename) with open(path) as f: try: data = json.load(f) except Exception: pass else: if not ('url' in data and 'title' in data and 'content' in data): continue id = hashlib.md5(data['url'].encode('utf8')).hexdigest() title = data.get('title', '') text = data.get('content', '') doc = Doc(id=id) title = title.encode('utf8') tok.tokenize(title, doc, 0, is_headline=True) tok.tokenize(text.encode('utf8'), doc, len(title) + 1, is_headline=False) print(doc) writer.write(doc) <commit_msg>Check for empty cluster dr files.<commit_after>
#!/usr/bin/env python3 from collections import defaultdict import hashlib import json import os import sys from gigacluster import Doc, dr, Tokenizer tok = Tokenizer() stats = defaultdict(int) for root, dirs, files in os.walk(sys.argv[1]): dirs.sort() files = set(files) cluster_dr = os.path.join(root, 'cluster.dr') with open(cluster_dr, 'wb') as f_dr: writer = dr.Writer(f_dr, Doc) written = 0 for filename in sorted(f for f in files if f.endswith('.readability')): path = os.path.join(root, filename) with open(path) as f: try: data = json.load(f) except Exception: stats['bad json'] += 1 pass else: if not ('url' in data and 'title' in data and 'content' in data): stats['no url/title/content'] += 1 continue id = hashlib.md5(data['url'].encode('utf8')).hexdigest() title = data.get('title', '') text = data.get('content', '') doc = Doc(id=id) title = title.encode('utf8') tok.tokenize(title, doc, 0, is_headline=True) tok.tokenize(text.encode('utf8'), doc, len(title) + 1, is_headline=False) writer.write(doc) stats['ok'] += 1 written += 1 if not written: print('No docs written for {}'.format(cluster_dr)) os.remove(cluster_dr) print('Stats\t{}'.format(dict(stats)))
#!/usr/bin/env python3 import hashlib import json import os import sys print(sys.path) from gigacluster import Doc, dr, Tokenizer tok = Tokenizer() for root, dirs, files in os.walk(sys.argv[1]): dirs.sort() files = set(files) cluster_dr = os.path.join(root, 'cluster.dr') with open(cluster_dr, 'wb') as f_dr: writer = dr.Writer(f_dr, Doc) for filename in sorted(f for f in files if f.endswith('.readability')): dr_filename = filename.replace('.readability', '.dr') if not dr_filename in files: path = os.path.join(root, filename) with open(path) as f: try: data = json.load(f) except Exception: pass else: if not ('url' in data and 'title' in data and 'content' in data): continue id = hashlib.md5(data['url'].encode('utf8')).hexdigest() title = data.get('title', '') text = data.get('content', '') doc = Doc(id=id) title = title.encode('utf8') tok.tokenize(title, doc, 0, is_headline=True) tok.tokenize(text.encode('utf8'), doc, len(title) + 1, is_headline=False) print(doc) writer.write(doc) Check for empty cluster dr files.#!/usr/bin/env python3 from collections import defaultdict import hashlib import json import os import sys from gigacluster import Doc, dr, Tokenizer tok = Tokenizer() stats = defaultdict(int) for root, dirs, files in os.walk(sys.argv[1]): dirs.sort() files = set(files) cluster_dr = os.path.join(root, 'cluster.dr') with open(cluster_dr, 'wb') as f_dr: writer = dr.Writer(f_dr, Doc) written = 0 for filename in sorted(f for f in files if f.endswith('.readability')): path = os.path.join(root, filename) with open(path) as f: try: data = json.load(f) except Exception: stats['bad json'] += 1 pass else: if not ('url' in data and 'title' in data and 'content' in data): stats['no url/title/content'] += 1 continue id = hashlib.md5(data['url'].encode('utf8')).hexdigest() title = data.get('title', '') text = data.get('content', '') doc = Doc(id=id) title = title.encode('utf8') tok.tokenize(title, doc, 0, is_headline=True) tok.tokenize(text.encode('utf8'), doc, len(title) + 1, is_headline=False) writer.write(doc) stats['ok'] += 1 written += 1 if not written: print('No docs written for {}'.format(cluster_dr)) os.remove(cluster_dr) print('Stats\t{}'.format(dict(stats)))
<commit_before>#!/usr/bin/env python3 import hashlib import json import os import sys print(sys.path) from gigacluster import Doc, dr, Tokenizer tok = Tokenizer() for root, dirs, files in os.walk(sys.argv[1]): dirs.sort() files = set(files) cluster_dr = os.path.join(root, 'cluster.dr') with open(cluster_dr, 'wb') as f_dr: writer = dr.Writer(f_dr, Doc) for filename in sorted(f for f in files if f.endswith('.readability')): dr_filename = filename.replace('.readability', '.dr') if not dr_filename in files: path = os.path.join(root, filename) with open(path) as f: try: data = json.load(f) except Exception: pass else: if not ('url' in data and 'title' in data and 'content' in data): continue id = hashlib.md5(data['url'].encode('utf8')).hexdigest() title = data.get('title', '') text = data.get('content', '') doc = Doc(id=id) title = title.encode('utf8') tok.tokenize(title, doc, 0, is_headline=True) tok.tokenize(text.encode('utf8'), doc, len(title) + 1, is_headline=False) print(doc) writer.write(doc) <commit_msg>Check for empty cluster dr files.<commit_after>#!/usr/bin/env python3 from collections import defaultdict import hashlib import json import os import sys from gigacluster import Doc, dr, Tokenizer tok = Tokenizer() stats = defaultdict(int) for root, dirs, files in os.walk(sys.argv[1]): dirs.sort() files = set(files) cluster_dr = os.path.join(root, 'cluster.dr') with open(cluster_dr, 'wb') as f_dr: writer = dr.Writer(f_dr, Doc) written = 0 for filename in sorted(f for f in files if f.endswith('.readability')): path = os.path.join(root, filename) with open(path) as f: try: data = json.load(f) except Exception: stats['bad json'] += 1 pass else: if not ('url' in data and 'title' in data and 'content' in data): stats['no url/title/content'] += 1 continue id = hashlib.md5(data['url'].encode('utf8')).hexdigest() title = data.get('title', '') text = data.get('content', '') doc = Doc(id=id) title = title.encode('utf8') tok.tokenize(title, doc, 0, is_headline=True) tok.tokenize(text.encode('utf8'), doc, len(title) + 1, is_headline=False) writer.write(doc) stats['ok'] += 1 written += 1 if not written: print('No docs written for {}'.format(cluster_dr)) os.remove(cluster_dr) print('Stats\t{}'.format(dict(stats)))
0d98a1c9b2682dde45196c8a3c8e89738aa3ab2a
litmus/cmds/__init__.py
litmus/cmds/__init__.py
#!/usr/bin/env python3 # Copyright 2015-2016 Samsung Electronics Co., Ltd. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from configparser import RawConfigParser from litmus.core.util import call def load_project_list(projects): """docstring for load_project_list""" configparser = RawConfigParser() configparser.read(projects) project_list = [] for section in configparser.sections(): item = dict(configparser.items(section)) item['name'] = section project_list.append(item) return project_list def sdb_does_exist(): help_url = 'https://github.com/dhs-shine/litmus#prerequisite' try: call('sdb version', shell=True, timeout=10) except FileNotFoundError: raise Exception('Please install sdb. Refer to {}'.format(help_url)) return
#!/usr/bin/env python3 # Copyright 2015-2016 Samsung Electronics Co., Ltd. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from configparser import RawConfigParser from litmus.core.util import call def load_project_list(projects): """docstring for load_project_list""" configparser = RawConfigParser() configparser.read(projects) project_list = [] for section in configparser.sections(): item = dict(configparser.items(section)) item['name'] = section project_list.append(item) return project_list def sdb_does_exist(): help_url = 'https://github.com/dhs-shine/litmus#prerequisite' try: call(['sdb', 'version'], timeout=10) except FileNotFoundError: raise Exception('Please install sdb. Refer to {}'.format(help_url)) return
Raise exception if sdb does not exist
Raise exception if sdb does not exist
Python
apache-2.0
dhs-shine/litmus
#!/usr/bin/env python3 # Copyright 2015-2016 Samsung Electronics Co., Ltd. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from configparser import RawConfigParser from litmus.core.util import call def load_project_list(projects): """docstring for load_project_list""" configparser = RawConfigParser() configparser.read(projects) project_list = [] for section in configparser.sections(): item = dict(configparser.items(section)) item['name'] = section project_list.append(item) return project_list def sdb_does_exist(): help_url = 'https://github.com/dhs-shine/litmus#prerequisite' try: call('sdb version', shell=True, timeout=10) except FileNotFoundError: raise Exception('Please install sdb. Refer to {}'.format(help_url)) return Raise exception if sdb does not exist
#!/usr/bin/env python3 # Copyright 2015-2016 Samsung Electronics Co., Ltd. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from configparser import RawConfigParser from litmus.core.util import call def load_project_list(projects): """docstring for load_project_list""" configparser = RawConfigParser() configparser.read(projects) project_list = [] for section in configparser.sections(): item = dict(configparser.items(section)) item['name'] = section project_list.append(item) return project_list def sdb_does_exist(): help_url = 'https://github.com/dhs-shine/litmus#prerequisite' try: call(['sdb', 'version'], timeout=10) except FileNotFoundError: raise Exception('Please install sdb. Refer to {}'.format(help_url)) return
<commit_before>#!/usr/bin/env python3 # Copyright 2015-2016 Samsung Electronics Co., Ltd. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from configparser import RawConfigParser from litmus.core.util import call def load_project_list(projects): """docstring for load_project_list""" configparser = RawConfigParser() configparser.read(projects) project_list = [] for section in configparser.sections(): item = dict(configparser.items(section)) item['name'] = section project_list.append(item) return project_list def sdb_does_exist(): help_url = 'https://github.com/dhs-shine/litmus#prerequisite' try: call('sdb version', shell=True, timeout=10) except FileNotFoundError: raise Exception('Please install sdb. Refer to {}'.format(help_url)) return <commit_msg>Raise exception if sdb does not exist<commit_after>
#!/usr/bin/env python3 # Copyright 2015-2016 Samsung Electronics Co., Ltd. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from configparser import RawConfigParser from litmus.core.util import call def load_project_list(projects): """docstring for load_project_list""" configparser = RawConfigParser() configparser.read(projects) project_list = [] for section in configparser.sections(): item = dict(configparser.items(section)) item['name'] = section project_list.append(item) return project_list def sdb_does_exist(): help_url = 'https://github.com/dhs-shine/litmus#prerequisite' try: call(['sdb', 'version'], timeout=10) except FileNotFoundError: raise Exception('Please install sdb. Refer to {}'.format(help_url)) return
#!/usr/bin/env python3 # Copyright 2015-2016 Samsung Electronics Co., Ltd. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from configparser import RawConfigParser from litmus.core.util import call def load_project_list(projects): """docstring for load_project_list""" configparser = RawConfigParser() configparser.read(projects) project_list = [] for section in configparser.sections(): item = dict(configparser.items(section)) item['name'] = section project_list.append(item) return project_list def sdb_does_exist(): help_url = 'https://github.com/dhs-shine/litmus#prerequisite' try: call('sdb version', shell=True, timeout=10) except FileNotFoundError: raise Exception('Please install sdb. Refer to {}'.format(help_url)) return Raise exception if sdb does not exist#!/usr/bin/env python3 # Copyright 2015-2016 Samsung Electronics Co., Ltd. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from configparser import RawConfigParser from litmus.core.util import call def load_project_list(projects): """docstring for load_project_list""" configparser = RawConfigParser() configparser.read(projects) project_list = [] for section in configparser.sections(): item = dict(configparser.items(section)) item['name'] = section project_list.append(item) return project_list def sdb_does_exist(): help_url = 'https://github.com/dhs-shine/litmus#prerequisite' try: call(['sdb', 'version'], timeout=10) except FileNotFoundError: raise Exception('Please install sdb. Refer to {}'.format(help_url)) return
<commit_before>#!/usr/bin/env python3 # Copyright 2015-2016 Samsung Electronics Co., Ltd. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from configparser import RawConfigParser from litmus.core.util import call def load_project_list(projects): """docstring for load_project_list""" configparser = RawConfigParser() configparser.read(projects) project_list = [] for section in configparser.sections(): item = dict(configparser.items(section)) item['name'] = section project_list.append(item) return project_list def sdb_does_exist(): help_url = 'https://github.com/dhs-shine/litmus#prerequisite' try: call('sdb version', shell=True, timeout=10) except FileNotFoundError: raise Exception('Please install sdb. Refer to {}'.format(help_url)) return <commit_msg>Raise exception if sdb does not exist<commit_after>#!/usr/bin/env python3 # Copyright 2015-2016 Samsung Electronics Co., Ltd. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from configparser import RawConfigParser from litmus.core.util import call def load_project_list(projects): """docstring for load_project_list""" configparser = RawConfigParser() configparser.read(projects) project_list = [] for section in configparser.sections(): item = dict(configparser.items(section)) item['name'] = section project_list.append(item) return project_list def sdb_does_exist(): help_url = 'https://github.com/dhs-shine/litmus#prerequisite' try: call(['sdb', 'version'], timeout=10) except FileNotFoundError: raise Exception('Please install sdb. Refer to {}'.format(help_url)) return
c74b777cc861963595593486cf803a1d7431ad65
matches/admin.py
matches/admin.py
from django.contrib import admin from .models import Match from .models import Tip def delete_tips(modeladmin, request, queryset): for match in queryset: tips = Tip.objects.filter(match = match) for tip in tips: tip.score = 0 tip.scoring_field = "" tip.is_score_calculated = False delete_tips.delete_tips = "Delete calculated scores for tips for these matches" class MatchAdmin(admin.ModelAdmin): actions = [delete_tips] admin.site.register(Match, MatchAdmin) admin.site.register(Tip)
from django.contrib import admin from .models import Match from .models import Tip def delete_tips(modeladmin, request, queryset): for match in queryset: tips = Tip.objects.filter(match = match) for tip in tips: tip.score = 0 tip.scoring_field = "" tip.is_score_calculated = False tip.save() delete_tips.delete_tips = "Delete calculated scores for tips for these matches" class MatchAdmin(admin.ModelAdmin): actions = [delete_tips] admin.site.register(Match, MatchAdmin) admin.site.register(Tip)
Add action to zero out tips for given match
Add action to zero out tips for given match
Python
mit
leventebakos/football-ech,leventebakos/football-ech
from django.contrib import admin from .models import Match from .models import Tip def delete_tips(modeladmin, request, queryset): for match in queryset: tips = Tip.objects.filter(match = match) for tip in tips: tip.score = 0 tip.scoring_field = "" tip.is_score_calculated = False delete_tips.delete_tips = "Delete calculated scores for tips for these matches" class MatchAdmin(admin.ModelAdmin): actions = [delete_tips] admin.site.register(Match, MatchAdmin) admin.site.register(Tip) Add action to zero out tips for given match
from django.contrib import admin from .models import Match from .models import Tip def delete_tips(modeladmin, request, queryset): for match in queryset: tips = Tip.objects.filter(match = match) for tip in tips: tip.score = 0 tip.scoring_field = "" tip.is_score_calculated = False tip.save() delete_tips.delete_tips = "Delete calculated scores for tips for these matches" class MatchAdmin(admin.ModelAdmin): actions = [delete_tips] admin.site.register(Match, MatchAdmin) admin.site.register(Tip)
<commit_before>from django.contrib import admin from .models import Match from .models import Tip def delete_tips(modeladmin, request, queryset): for match in queryset: tips = Tip.objects.filter(match = match) for tip in tips: tip.score = 0 tip.scoring_field = "" tip.is_score_calculated = False delete_tips.delete_tips = "Delete calculated scores for tips for these matches" class MatchAdmin(admin.ModelAdmin): actions = [delete_tips] admin.site.register(Match, MatchAdmin) admin.site.register(Tip) <commit_msg>Add action to zero out tips for given match<commit_after>
from django.contrib import admin from .models import Match from .models import Tip def delete_tips(modeladmin, request, queryset): for match in queryset: tips = Tip.objects.filter(match = match) for tip in tips: tip.score = 0 tip.scoring_field = "" tip.is_score_calculated = False tip.save() delete_tips.delete_tips = "Delete calculated scores for tips for these matches" class MatchAdmin(admin.ModelAdmin): actions = [delete_tips] admin.site.register(Match, MatchAdmin) admin.site.register(Tip)
from django.contrib import admin from .models import Match from .models import Tip def delete_tips(modeladmin, request, queryset): for match in queryset: tips = Tip.objects.filter(match = match) for tip in tips: tip.score = 0 tip.scoring_field = "" tip.is_score_calculated = False delete_tips.delete_tips = "Delete calculated scores for tips for these matches" class MatchAdmin(admin.ModelAdmin): actions = [delete_tips] admin.site.register(Match, MatchAdmin) admin.site.register(Tip) Add action to zero out tips for given matchfrom django.contrib import admin from .models import Match from .models import Tip def delete_tips(modeladmin, request, queryset): for match in queryset: tips = Tip.objects.filter(match = match) for tip in tips: tip.score = 0 tip.scoring_field = "" tip.is_score_calculated = False tip.save() delete_tips.delete_tips = "Delete calculated scores for tips for these matches" class MatchAdmin(admin.ModelAdmin): actions = [delete_tips] admin.site.register(Match, MatchAdmin) admin.site.register(Tip)
<commit_before>from django.contrib import admin from .models import Match from .models import Tip def delete_tips(modeladmin, request, queryset): for match in queryset: tips = Tip.objects.filter(match = match) for tip in tips: tip.score = 0 tip.scoring_field = "" tip.is_score_calculated = False delete_tips.delete_tips = "Delete calculated scores for tips for these matches" class MatchAdmin(admin.ModelAdmin): actions = [delete_tips] admin.site.register(Match, MatchAdmin) admin.site.register(Tip) <commit_msg>Add action to zero out tips for given match<commit_after>from django.contrib import admin from .models import Match from .models import Tip def delete_tips(modeladmin, request, queryset): for match in queryset: tips = Tip.objects.filter(match = match) for tip in tips: tip.score = 0 tip.scoring_field = "" tip.is_score_calculated = False tip.save() delete_tips.delete_tips = "Delete calculated scores for tips for these matches" class MatchAdmin(admin.ModelAdmin): actions = [delete_tips] admin.site.register(Match, MatchAdmin) admin.site.register(Tip)
d073de94f1896239bd6120893b6a94c43061a279
byceps/util/image/models.py
byceps/util/image/models.py
""" byceps.util.image.models ~~~~~~~~~~~~~~~~~~~~~~~~ :Copyright: 2006-2017 Jochen Kupperschmidt :License: Modified BSD, see LICENSE for details. """ from collections import namedtuple from enum import Enum class Dimensions(namedtuple('Dimensions', ['width', 'height'])): """A 2D image's width and height.""" __slots__ = () @property def is_square(self): return self.width == self.height ImageType = Enum('ImageType', ['gif', 'jpeg', 'png']) # type: ignore
""" byceps.util.image.models ~~~~~~~~~~~~~~~~~~~~~~~~ :Copyright: 2006-2017 Jochen Kupperschmidt :License: Modified BSD, see LICENSE for details. """ from collections import namedtuple from enum import Enum class Dimensions(namedtuple('Dimensions', ['width', 'height'])): """A 2D image's width and height.""" __slots__ = () @property def is_square(self): return self.width == self.height ImageType = Enum('ImageType', ['gif', 'jpeg', 'png'])
Remove type ignore comment from functional enum API call
Remove type ignore comment from functional enum API call This is supported as of Mypy 0.510 and no longer raises an error.
Python
bsd-3-clause
homeworkprod/byceps,homeworkprod/byceps,m-ober/byceps,m-ober/byceps,m-ober/byceps,homeworkprod/byceps
""" byceps.util.image.models ~~~~~~~~~~~~~~~~~~~~~~~~ :Copyright: 2006-2017 Jochen Kupperschmidt :License: Modified BSD, see LICENSE for details. """ from collections import namedtuple from enum import Enum class Dimensions(namedtuple('Dimensions', ['width', 'height'])): """A 2D image's width and height.""" __slots__ = () @property def is_square(self): return self.width == self.height ImageType = Enum('ImageType', ['gif', 'jpeg', 'png']) # type: ignore Remove type ignore comment from functional enum API call This is supported as of Mypy 0.510 and no longer raises an error.
""" byceps.util.image.models ~~~~~~~~~~~~~~~~~~~~~~~~ :Copyright: 2006-2017 Jochen Kupperschmidt :License: Modified BSD, see LICENSE for details. """ from collections import namedtuple from enum import Enum class Dimensions(namedtuple('Dimensions', ['width', 'height'])): """A 2D image's width and height.""" __slots__ = () @property def is_square(self): return self.width == self.height ImageType = Enum('ImageType', ['gif', 'jpeg', 'png'])
<commit_before>""" byceps.util.image.models ~~~~~~~~~~~~~~~~~~~~~~~~ :Copyright: 2006-2017 Jochen Kupperschmidt :License: Modified BSD, see LICENSE for details. """ from collections import namedtuple from enum import Enum class Dimensions(namedtuple('Dimensions', ['width', 'height'])): """A 2D image's width and height.""" __slots__ = () @property def is_square(self): return self.width == self.height ImageType = Enum('ImageType', ['gif', 'jpeg', 'png']) # type: ignore <commit_msg>Remove type ignore comment from functional enum API call This is supported as of Mypy 0.510 and no longer raises an error.<commit_after>
""" byceps.util.image.models ~~~~~~~~~~~~~~~~~~~~~~~~ :Copyright: 2006-2017 Jochen Kupperschmidt :License: Modified BSD, see LICENSE for details. """ from collections import namedtuple from enum import Enum class Dimensions(namedtuple('Dimensions', ['width', 'height'])): """A 2D image's width and height.""" __slots__ = () @property def is_square(self): return self.width == self.height ImageType = Enum('ImageType', ['gif', 'jpeg', 'png'])
""" byceps.util.image.models ~~~~~~~~~~~~~~~~~~~~~~~~ :Copyright: 2006-2017 Jochen Kupperschmidt :License: Modified BSD, see LICENSE for details. """ from collections import namedtuple from enum import Enum class Dimensions(namedtuple('Dimensions', ['width', 'height'])): """A 2D image's width and height.""" __slots__ = () @property def is_square(self): return self.width == self.height ImageType = Enum('ImageType', ['gif', 'jpeg', 'png']) # type: ignore Remove type ignore comment from functional enum API call This is supported as of Mypy 0.510 and no longer raises an error.""" byceps.util.image.models ~~~~~~~~~~~~~~~~~~~~~~~~ :Copyright: 2006-2017 Jochen Kupperschmidt :License: Modified BSD, see LICENSE for details. """ from collections import namedtuple from enum import Enum class Dimensions(namedtuple('Dimensions', ['width', 'height'])): """A 2D image's width and height.""" __slots__ = () @property def is_square(self): return self.width == self.height ImageType = Enum('ImageType', ['gif', 'jpeg', 'png'])
<commit_before>""" byceps.util.image.models ~~~~~~~~~~~~~~~~~~~~~~~~ :Copyright: 2006-2017 Jochen Kupperschmidt :License: Modified BSD, see LICENSE for details. """ from collections import namedtuple from enum import Enum class Dimensions(namedtuple('Dimensions', ['width', 'height'])): """A 2D image's width and height.""" __slots__ = () @property def is_square(self): return self.width == self.height ImageType = Enum('ImageType', ['gif', 'jpeg', 'png']) # type: ignore <commit_msg>Remove type ignore comment from functional enum API call This is supported as of Mypy 0.510 and no longer raises an error.<commit_after>""" byceps.util.image.models ~~~~~~~~~~~~~~~~~~~~~~~~ :Copyright: 2006-2017 Jochen Kupperschmidt :License: Modified BSD, see LICENSE for details. """ from collections import namedtuple from enum import Enum class Dimensions(namedtuple('Dimensions', ['width', 'height'])): """A 2D image's width and height.""" __slots__ = () @property def is_square(self): return self.width == self.height ImageType = Enum('ImageType', ['gif', 'jpeg', 'png'])