commit
stringlengths 40
40
| old_file
stringlengths 4
118
| new_file
stringlengths 4
118
| old_contents
stringlengths 0
2.94k
| new_contents
stringlengths 1
4.43k
| subject
stringlengths 15
444
| message
stringlengths 16
3.45k
| lang
stringclasses 1
value | license
stringclasses 13
values | repos
stringlengths 5
43.2k
| prompt
stringlengths 17
4.58k
| response
stringlengths 1
4.43k
| prompt_tagged
stringlengths 58
4.62k
| response_tagged
stringlengths 1
4.43k
| text
stringlengths 132
7.29k
| text_tagged
stringlengths 173
7.33k
|
|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
56186c985b87fbbf0a7ea0f04c8b089a13b29fe3
|
execute_all_tests.py
|
execute_all_tests.py
|
#! /bin/python3
"""
This program is free software: you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation, either version 3 of the License, or
(at your option) any later version.
This program is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with this program. If not, see <http://www.gnu.org/licenses/>.
"""
import os
import sys
from coalib.tests.TestHelper import TestHelper
def show_help():
print("Usage: {name} [OPTIONS]".format(name=sys.argv[0]))
print()
print("--help : Show this help text")
print("--cover : Use coverage to get statement and branch coverage of tests")
if __name__ == '__main__':
use_coverage = False
for arg in sys.argv[1:]:
arg = str(arg).strip().lower()
if arg == "--cover" and not use_coverage:
use_coverage = True
else:
show_help()
exit()
test_dir = os.path.abspath("coalib/tests")
files = TestHelper.get_test_files(test_dir)
exit(TestHelper.execute_python3_files(files, use_coverage))
|
#! /bin/python3
"""
This program is free software: you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation, either version 3 of the License, or
(at your option) any later version.
This program is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with this program. If not, see <http://www.gnu.org/licenses/>.
"""
import os
import sys
from coalib.tests.TestHelper import TestHelper
def show_help():
print("Usage: {name} [OPTIONS]".format(name=sys.argv[0]))
print()
print("--help : Show this help text")
print("--cover : Use coverage to get statement and branch coverage of tests")
if __name__ == '__main__':
use_coverage = False
for arg in sys.argv[1:]:
arg = str(arg).strip().lower()
if arg == "--cover" and not use_coverage:
use_coverage = True
else:
show_help()
exit()
files = TestHelper.get_test_files(os.path.abspath("coalib/tests"))
exit(TestHelper.execute_python3_files(files, use_coverage))
|
Test execution: Remove unneeded variable
|
Test execution: Remove unneeded variable
|
Python
|
agpl-3.0
|
Tanmay28/coala,Tanmay28/coala,meetmangukiya/coala,arush0311/coala,NalinG/coala,Tanmay28/coala,incorrectusername/coala,yashtrivedi96/coala,shreyans800755/coala,sagark123/coala,jayvdb/coala,MariosPanag/coala,Nosferatul/coala,vinc456/coala,damngamerz/coala,MattAllmendinger/coala,ManjiriBirajdar/coala,andreimacavei/coala,rresol/coala,AdeshAtole/coala,kartikeys98/coala,NalinG/coala,karansingh1559/coala,coala/coala,SambitAcharya/coala,Uran198/coala,Tanmay28/coala,nemaniarjun/coala,coala/coala,jayvdb/coala,netman92/coala,lonewolf07/coala,andreimacavei/coala,SanketDG/coala,d6e/coala,NiklasMM/coala,tltuan/coala,SanketDG/coala,refeed/coala,rimacone/testing2,SambitAcharya/coala,abhiroyg/coala,arush0311/coala,svsn2117/coala,shreyans800755/coala,sils1297/coala,arjunsinghy96/coala,rresol/coala,AbdealiJK/coala,tushar-rishav/coala,AdeshAtole/coala,NiklasMM/coala,arafsheikh/coala,scriptnull/coala,aptrishu/coala,scriptnull/coala,MattAllmendinger/coala,yland/coala,AbdealiJK/coala,SambitAcharya/coala,ayushin78/coala,arjunsinghy96/coala,yashtrivedi96/coala,FeodorFitsner/coala,SanketDG/coala,andreimacavei/coala,sils1297/coala,JohnS-01/coala,nemaniarjun/coala,mr-karan/coala,rimacone/testing2,karansingh1559/coala,incorrectusername/coala,yashLadha/coala,coala-analyzer/coala,NalinG/coala,NiklasMM/coala,saurabhiiit/coala,saurabhiiit/coala,MattAllmendinger/coala,NalinG/coala,karansingh1559/coala,Nosferatul/coala,Nosferatul/coala,ayushin78/coala,CruiseDevice/coala,d6e/coala,djkonro/coala,ManjiriBirajdar/coala,rimacone/testing2,sagark123/coala,SambitAcharya/coala,RJ722/coala,stevemontana1980/coala,kartikeys98/coala,NalinG/coala,Asnelchristian/coala,Tanmay28/coala,Uran198/coala,Shade5/coala,stevemontana1980/coala,incorrectusername/coala,Balaji2198/coala,RJ722/coala,swatilodha/coala,ManjiriBirajdar/coala,Tanmay28/coala,MariosPanag/coala,aptrishu/coala,sophiavanvalkenburg/coala,sils1297/coala,coala-analyzer/coala,RJ722/coala,arush0311/coala,saurabhiiit/coala,svsn2117/coala,vinc456/coala,refeed/coala,sophiavanvalkenburg/coala,impmihai/coala,meetmangukiya/coala,yashLadha/coala,djkonro/coala,svsn2117/coala,Shade5/coala,netman92/coala,scriptnull/coala,abhiroyg/coala,JohnS-01/coala,damngamerz/coala,tushar-rishav/coala,arafsheikh/coala,abhiroyg/coala,CruiseDevice/coala,lonewolf07/coala,ayushin78/coala,SambitAcharya/coala,sagark123/coala,tushar-rishav/coala,yland/coala,jayvdb/coala,shreyans800755/coala,JohnS-01/coala,Balaji2198/coala,impmihai/coala,swatilodha/coala,sophiavanvalkenburg/coala,stevemontana1980/coala,dagdaggo/coala,CruiseDevice/coala,coala/coala,netman92/coala,scriptnull/coala,scottbelden/coala,tltuan/coala,Asalle/coala,sudheesh001/coala,Uran198/coala,MariosPanag/coala,djkonro/coala,mr-karan/coala,SambitAcharya/coala,coala-analyzer/coala,meetmangukiya/coala,scriptnull/coala,swatilodha/coala,dagdaggo/coala,Tanmay28/coala,scottbelden/coala,impmihai/coala,dagdaggo/coala,FeodorFitsner/coala,vinc456/coala,AbdealiJK/coala,rresol/coala,NalinG/coala,d6e/coala,Shade5/coala,Asalle/coala,Tanmay28/coala,nemaniarjun/coala,Balaji2198/coala,lonewolf07/coala,arjunsinghy96/coala,yashtrivedi96/coala,aptrishu/coala,Asnelchristian/coala,yashLadha/coala,NalinG/coala,scriptnull/coala,arafsheikh/coala,scottbelden/coala,scriptnull/coala,kartikeys98/coala,yland/coala,AdeshAtole/coala,tltuan/coala,refeed/coala,Asnelchristian/coala,damngamerz/coala,mr-karan/coala,FeodorFitsner/coala,sudheesh001/coala,Asalle/coala,sudheesh001/coala,SambitAcharya/coala
|
#! /bin/python3
"""
This program is free software: you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation, either version 3 of the License, or
(at your option) any later version.
This program is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with this program. If not, see <http://www.gnu.org/licenses/>.
"""
import os
import sys
from coalib.tests.TestHelper import TestHelper
def show_help():
print("Usage: {name} [OPTIONS]".format(name=sys.argv[0]))
print()
print("--help : Show this help text")
print("--cover : Use coverage to get statement and branch coverage of tests")
if __name__ == '__main__':
use_coverage = False
for arg in sys.argv[1:]:
arg = str(arg).strip().lower()
if arg == "--cover" and not use_coverage:
use_coverage = True
else:
show_help()
exit()
test_dir = os.path.abspath("coalib/tests")
files = TestHelper.get_test_files(test_dir)
exit(TestHelper.execute_python3_files(files, use_coverage))
Test execution: Remove unneeded variable
|
#! /bin/python3
"""
This program is free software: you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation, either version 3 of the License, or
(at your option) any later version.
This program is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with this program. If not, see <http://www.gnu.org/licenses/>.
"""
import os
import sys
from coalib.tests.TestHelper import TestHelper
def show_help():
print("Usage: {name} [OPTIONS]".format(name=sys.argv[0]))
print()
print("--help : Show this help text")
print("--cover : Use coverage to get statement and branch coverage of tests")
if __name__ == '__main__':
use_coverage = False
for arg in sys.argv[1:]:
arg = str(arg).strip().lower()
if arg == "--cover" and not use_coverage:
use_coverage = True
else:
show_help()
exit()
files = TestHelper.get_test_files(os.path.abspath("coalib/tests"))
exit(TestHelper.execute_python3_files(files, use_coverage))
|
<commit_before>#! /bin/python3
"""
This program is free software: you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation, either version 3 of the License, or
(at your option) any later version.
This program is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with this program. If not, see <http://www.gnu.org/licenses/>.
"""
import os
import sys
from coalib.tests.TestHelper import TestHelper
def show_help():
print("Usage: {name} [OPTIONS]".format(name=sys.argv[0]))
print()
print("--help : Show this help text")
print("--cover : Use coverage to get statement and branch coverage of tests")
if __name__ == '__main__':
use_coverage = False
for arg in sys.argv[1:]:
arg = str(arg).strip().lower()
if arg == "--cover" and not use_coverage:
use_coverage = True
else:
show_help()
exit()
test_dir = os.path.abspath("coalib/tests")
files = TestHelper.get_test_files(test_dir)
exit(TestHelper.execute_python3_files(files, use_coverage))
<commit_msg>Test execution: Remove unneeded variable<commit_after>
|
#! /bin/python3
"""
This program is free software: you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation, either version 3 of the License, or
(at your option) any later version.
This program is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with this program. If not, see <http://www.gnu.org/licenses/>.
"""
import os
import sys
from coalib.tests.TestHelper import TestHelper
def show_help():
print("Usage: {name} [OPTIONS]".format(name=sys.argv[0]))
print()
print("--help : Show this help text")
print("--cover : Use coverage to get statement and branch coverage of tests")
if __name__ == '__main__':
use_coverage = False
for arg in sys.argv[1:]:
arg = str(arg).strip().lower()
if arg == "--cover" and not use_coverage:
use_coverage = True
else:
show_help()
exit()
files = TestHelper.get_test_files(os.path.abspath("coalib/tests"))
exit(TestHelper.execute_python3_files(files, use_coverage))
|
#! /bin/python3
"""
This program is free software: you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation, either version 3 of the License, or
(at your option) any later version.
This program is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with this program. If not, see <http://www.gnu.org/licenses/>.
"""
import os
import sys
from coalib.tests.TestHelper import TestHelper
def show_help():
print("Usage: {name} [OPTIONS]".format(name=sys.argv[0]))
print()
print("--help : Show this help text")
print("--cover : Use coverage to get statement and branch coverage of tests")
if __name__ == '__main__':
use_coverage = False
for arg in sys.argv[1:]:
arg = str(arg).strip().lower()
if arg == "--cover" and not use_coverage:
use_coverage = True
else:
show_help()
exit()
test_dir = os.path.abspath("coalib/tests")
files = TestHelper.get_test_files(test_dir)
exit(TestHelper.execute_python3_files(files, use_coverage))
Test execution: Remove unneeded variable#! /bin/python3
"""
This program is free software: you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation, either version 3 of the License, or
(at your option) any later version.
This program is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with this program. If not, see <http://www.gnu.org/licenses/>.
"""
import os
import sys
from coalib.tests.TestHelper import TestHelper
def show_help():
print("Usage: {name} [OPTIONS]".format(name=sys.argv[0]))
print()
print("--help : Show this help text")
print("--cover : Use coverage to get statement and branch coverage of tests")
if __name__ == '__main__':
use_coverage = False
for arg in sys.argv[1:]:
arg = str(arg).strip().lower()
if arg == "--cover" and not use_coverage:
use_coverage = True
else:
show_help()
exit()
files = TestHelper.get_test_files(os.path.abspath("coalib/tests"))
exit(TestHelper.execute_python3_files(files, use_coverage))
|
<commit_before>#! /bin/python3
"""
This program is free software: you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation, either version 3 of the License, or
(at your option) any later version.
This program is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with this program. If not, see <http://www.gnu.org/licenses/>.
"""
import os
import sys
from coalib.tests.TestHelper import TestHelper
def show_help():
print("Usage: {name} [OPTIONS]".format(name=sys.argv[0]))
print()
print("--help : Show this help text")
print("--cover : Use coverage to get statement and branch coverage of tests")
if __name__ == '__main__':
use_coverage = False
for arg in sys.argv[1:]:
arg = str(arg).strip().lower()
if arg == "--cover" and not use_coverage:
use_coverage = True
else:
show_help()
exit()
test_dir = os.path.abspath("coalib/tests")
files = TestHelper.get_test_files(test_dir)
exit(TestHelper.execute_python3_files(files, use_coverage))
<commit_msg>Test execution: Remove unneeded variable<commit_after>#! /bin/python3
"""
This program is free software: you can redistribute it and/or modify
it under the terms of the GNU General Public License as published by
the Free Software Foundation, either version 3 of the License, or
(at your option) any later version.
This program is distributed in the hope that it will be useful,
but WITHOUT ANY WARRANTY; without even the implied warranty of
MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
GNU General Public License for more details.
You should have received a copy of the GNU General Public License
along with this program. If not, see <http://www.gnu.org/licenses/>.
"""
import os
import sys
from coalib.tests.TestHelper import TestHelper
def show_help():
print("Usage: {name} [OPTIONS]".format(name=sys.argv[0]))
print()
print("--help : Show this help text")
print("--cover : Use coverage to get statement and branch coverage of tests")
if __name__ == '__main__':
use_coverage = False
for arg in sys.argv[1:]:
arg = str(arg).strip().lower()
if arg == "--cover" and not use_coverage:
use_coverage = True
else:
show_help()
exit()
files = TestHelper.get_test_files(os.path.abspath("coalib/tests"))
exit(TestHelper.execute_python3_files(files, use_coverage))
|
39f327bb9e37d6d290eb3f3179f7e79d60b5ab6d
|
model.py
|
model.py
|
from sqlalchemy.ext.declarative import declarative_base
from sqlalchemy import create_engine
engine = create_engine('postgresql://wn:wn@localhost:5432/wndb')
Base = declarative_base()
from sqlalchemy import Column, Integer, Float, DateTime, Boolean, String
class Observation(Base):
__tablename__ = 'obs'
id = Column(Integer, primary_key=True)
station_name = Column(String)
x = Column(Float)
y = Column(Float)
z = Column(Float)
jam_indicator = Column(Boolean)
jam_intensity = Column(Float)
date_time = Column(DateTime)
def __repr__(self):
return "<Observation(station_name='%r', x='%f', y='%f', z='%f', jam_indicator='%r', jam_intensity='%f', date_time='%r')>" % (
self.station_name, self.x, self.y, self.z, self.jam_indicator, self.jam_intensity, self.date_time)
Base.metadata.create_all(engine)
|
from sqlalchemy import create_engine
engine = create_engine('postgresql://wn:wn@localhost:5432/wndb')
from sqlalchemy import Column, Integer, Float, DateTime, Boolean, String, MetaData
metadata = MetaData()
table = Table('obs', metadata, Column(Integer, primary_key=True),
Column('station_name',String),
Column('x',Float),
Column('y',Float),
Column('z',Float),
Column('jam_indicator',Boolean),
Column('jam_intensity',Float),
Column('date_time',DateTime)
)
|
Switch from ORM to Core
|
Switch from ORM to Core
|
Python
|
mit
|
labhack/whiskeynovember,labhack/whiskeynovember,labhack/whiskeynovember
|
from sqlalchemy.ext.declarative import declarative_base
from sqlalchemy import create_engine
engine = create_engine('postgresql://wn:wn@localhost:5432/wndb')
Base = declarative_base()
from sqlalchemy import Column, Integer, Float, DateTime, Boolean, String
class Observation(Base):
__tablename__ = 'obs'
id = Column(Integer, primary_key=True)
station_name = Column(String)
x = Column(Float)
y = Column(Float)
z = Column(Float)
jam_indicator = Column(Boolean)
jam_intensity = Column(Float)
date_time = Column(DateTime)
def __repr__(self):
return "<Observation(station_name='%r', x='%f', y='%f', z='%f', jam_indicator='%r', jam_intensity='%f', date_time='%r')>" % (
self.station_name, self.x, self.y, self.z, self.jam_indicator, self.jam_intensity, self.date_time)
Base.metadata.create_all(engine)
Switch from ORM to Core
|
from sqlalchemy import create_engine
engine = create_engine('postgresql://wn:wn@localhost:5432/wndb')
from sqlalchemy import Column, Integer, Float, DateTime, Boolean, String, MetaData
metadata = MetaData()
table = Table('obs', metadata, Column(Integer, primary_key=True),
Column('station_name',String),
Column('x',Float),
Column('y',Float),
Column('z',Float),
Column('jam_indicator',Boolean),
Column('jam_intensity',Float),
Column('date_time',DateTime)
)
|
<commit_before>from sqlalchemy.ext.declarative import declarative_base
from sqlalchemy import create_engine
engine = create_engine('postgresql://wn:wn@localhost:5432/wndb')
Base = declarative_base()
from sqlalchemy import Column, Integer, Float, DateTime, Boolean, String
class Observation(Base):
__tablename__ = 'obs'
id = Column(Integer, primary_key=True)
station_name = Column(String)
x = Column(Float)
y = Column(Float)
z = Column(Float)
jam_indicator = Column(Boolean)
jam_intensity = Column(Float)
date_time = Column(DateTime)
def __repr__(self):
return "<Observation(station_name='%r', x='%f', y='%f', z='%f', jam_indicator='%r', jam_intensity='%f', date_time='%r')>" % (
self.station_name, self.x, self.y, self.z, self.jam_indicator, self.jam_intensity, self.date_time)
Base.metadata.create_all(engine)
<commit_msg>Switch from ORM to Core<commit_after>
|
from sqlalchemy import create_engine
engine = create_engine('postgresql://wn:wn@localhost:5432/wndb')
from sqlalchemy import Column, Integer, Float, DateTime, Boolean, String, MetaData
metadata = MetaData()
table = Table('obs', metadata, Column(Integer, primary_key=True),
Column('station_name',String),
Column('x',Float),
Column('y',Float),
Column('z',Float),
Column('jam_indicator',Boolean),
Column('jam_intensity',Float),
Column('date_time',DateTime)
)
|
from sqlalchemy.ext.declarative import declarative_base
from sqlalchemy import create_engine
engine = create_engine('postgresql://wn:wn@localhost:5432/wndb')
Base = declarative_base()
from sqlalchemy import Column, Integer, Float, DateTime, Boolean, String
class Observation(Base):
__tablename__ = 'obs'
id = Column(Integer, primary_key=True)
station_name = Column(String)
x = Column(Float)
y = Column(Float)
z = Column(Float)
jam_indicator = Column(Boolean)
jam_intensity = Column(Float)
date_time = Column(DateTime)
def __repr__(self):
return "<Observation(station_name='%r', x='%f', y='%f', z='%f', jam_indicator='%r', jam_intensity='%f', date_time='%r')>" % (
self.station_name, self.x, self.y, self.z, self.jam_indicator, self.jam_intensity, self.date_time)
Base.metadata.create_all(engine)
Switch from ORM to Corefrom sqlalchemy import create_engine
engine = create_engine('postgresql://wn:wn@localhost:5432/wndb')
from sqlalchemy import Column, Integer, Float, DateTime, Boolean, String, MetaData
metadata = MetaData()
table = Table('obs', metadata, Column(Integer, primary_key=True),
Column('station_name',String),
Column('x',Float),
Column('y',Float),
Column('z',Float),
Column('jam_indicator',Boolean),
Column('jam_intensity',Float),
Column('date_time',DateTime)
)
|
<commit_before>from sqlalchemy.ext.declarative import declarative_base
from sqlalchemy import create_engine
engine = create_engine('postgresql://wn:wn@localhost:5432/wndb')
Base = declarative_base()
from sqlalchemy import Column, Integer, Float, DateTime, Boolean, String
class Observation(Base):
__tablename__ = 'obs'
id = Column(Integer, primary_key=True)
station_name = Column(String)
x = Column(Float)
y = Column(Float)
z = Column(Float)
jam_indicator = Column(Boolean)
jam_intensity = Column(Float)
date_time = Column(DateTime)
def __repr__(self):
return "<Observation(station_name='%r', x='%f', y='%f', z='%f', jam_indicator='%r', jam_intensity='%f', date_time='%r')>" % (
self.station_name, self.x, self.y, self.z, self.jam_indicator, self.jam_intensity, self.date_time)
Base.metadata.create_all(engine)
<commit_msg>Switch from ORM to Core<commit_after>from sqlalchemy import create_engine
engine = create_engine('postgresql://wn:wn@localhost:5432/wndb')
from sqlalchemy import Column, Integer, Float, DateTime, Boolean, String, MetaData
metadata = MetaData()
table = Table('obs', metadata, Column(Integer, primary_key=True),
Column('station_name',String),
Column('x',Float),
Column('y',Float),
Column('z',Float),
Column('jam_indicator',Boolean),
Column('jam_intensity',Float),
Column('date_time',DateTime)
)
|
a5b89ed7aa9e2fe4305f6431a3bdd675a7eda03f
|
web/__init__.py
|
web/__init__.py
|
# -*- coding: utf-8 -*-
from os import path
from flask import Flask
PACKAGE_DIR = path.dirname(path.realpath(__file__))
ROOT_DIR = path.realpath(path.join(PACKAGE_DIR, '..'))
ROOT_URL = 'http://pythoncz.herokuapp.com'
GITHUB_URL = (
'https://github.com/honzajavorek/python.cz/'
'blob/master/{template_folder}/{filename}'
)
app = Flask('web',
static_folder=path.join(ROOT_DIR, 'files'),
static_url_path='',
template_folder=path.join(ROOT_DIR, 'pages'))
app.config.from_object(__name__)
from . import views # NOQA
|
# -*- coding: utf-8 -*-
from os import path
from flask import Flask
PACKAGE_DIR = path.dirname(path.realpath(__file__))
ROOT_DIR = path.realpath(path.join(PACKAGE_DIR, '..'))
ROOT_URL = 'http://pythoncz.herokuapp.com'
GITHUB_URL = (
'https://github.com/honzajavorek/python.cz/'
'blob/master/{template_folder}/{filename}'
)
app = Flask('web',
static_folder=path.join(ROOT_DIR, 'files'),
static_url_path='',
template_folder=path.join(ROOT_DIR, 'pages'))
app.config.from_object(__name__)
from . import views # NOQA
|
Fix newline at the end of file.
|
Fix newline at the end of file.
|
Python
|
mit
|
honzajavorek/python.cz,honzajavorek/python.cz,honzajavorek/python.cz,honzajavorek/python.cz
|
# -*- coding: utf-8 -*-
from os import path
from flask import Flask
PACKAGE_DIR = path.dirname(path.realpath(__file__))
ROOT_DIR = path.realpath(path.join(PACKAGE_DIR, '..'))
ROOT_URL = 'http://pythoncz.herokuapp.com'
GITHUB_URL = (
'https://github.com/honzajavorek/python.cz/'
'blob/master/{template_folder}/{filename}'
)
app = Flask('web',
static_folder=path.join(ROOT_DIR, 'files'),
static_url_path='',
template_folder=path.join(ROOT_DIR, 'pages'))
app.config.from_object(__name__)
from . import views # NOQAFix newline at the end of file.
|
# -*- coding: utf-8 -*-
from os import path
from flask import Flask
PACKAGE_DIR = path.dirname(path.realpath(__file__))
ROOT_DIR = path.realpath(path.join(PACKAGE_DIR, '..'))
ROOT_URL = 'http://pythoncz.herokuapp.com'
GITHUB_URL = (
'https://github.com/honzajavorek/python.cz/'
'blob/master/{template_folder}/{filename}'
)
app = Flask('web',
static_folder=path.join(ROOT_DIR, 'files'),
static_url_path='',
template_folder=path.join(ROOT_DIR, 'pages'))
app.config.from_object(__name__)
from . import views # NOQA
|
<commit_before># -*- coding: utf-8 -*-
from os import path
from flask import Flask
PACKAGE_DIR = path.dirname(path.realpath(__file__))
ROOT_DIR = path.realpath(path.join(PACKAGE_DIR, '..'))
ROOT_URL = 'http://pythoncz.herokuapp.com'
GITHUB_URL = (
'https://github.com/honzajavorek/python.cz/'
'blob/master/{template_folder}/{filename}'
)
app = Flask('web',
static_folder=path.join(ROOT_DIR, 'files'),
static_url_path='',
template_folder=path.join(ROOT_DIR, 'pages'))
app.config.from_object(__name__)
from . import views # NOQA<commit_msg>Fix newline at the end of file.<commit_after>
|
# -*- coding: utf-8 -*-
from os import path
from flask import Flask
PACKAGE_DIR = path.dirname(path.realpath(__file__))
ROOT_DIR = path.realpath(path.join(PACKAGE_DIR, '..'))
ROOT_URL = 'http://pythoncz.herokuapp.com'
GITHUB_URL = (
'https://github.com/honzajavorek/python.cz/'
'blob/master/{template_folder}/{filename}'
)
app = Flask('web',
static_folder=path.join(ROOT_DIR, 'files'),
static_url_path='',
template_folder=path.join(ROOT_DIR, 'pages'))
app.config.from_object(__name__)
from . import views # NOQA
|
# -*- coding: utf-8 -*-
from os import path
from flask import Flask
PACKAGE_DIR = path.dirname(path.realpath(__file__))
ROOT_DIR = path.realpath(path.join(PACKAGE_DIR, '..'))
ROOT_URL = 'http://pythoncz.herokuapp.com'
GITHUB_URL = (
'https://github.com/honzajavorek/python.cz/'
'blob/master/{template_folder}/{filename}'
)
app = Flask('web',
static_folder=path.join(ROOT_DIR, 'files'),
static_url_path='',
template_folder=path.join(ROOT_DIR, 'pages'))
app.config.from_object(__name__)
from . import views # NOQAFix newline at the end of file.# -*- coding: utf-8 -*-
from os import path
from flask import Flask
PACKAGE_DIR = path.dirname(path.realpath(__file__))
ROOT_DIR = path.realpath(path.join(PACKAGE_DIR, '..'))
ROOT_URL = 'http://pythoncz.herokuapp.com'
GITHUB_URL = (
'https://github.com/honzajavorek/python.cz/'
'blob/master/{template_folder}/{filename}'
)
app = Flask('web',
static_folder=path.join(ROOT_DIR, 'files'),
static_url_path='',
template_folder=path.join(ROOT_DIR, 'pages'))
app.config.from_object(__name__)
from . import views # NOQA
|
<commit_before># -*- coding: utf-8 -*-
from os import path
from flask import Flask
PACKAGE_DIR = path.dirname(path.realpath(__file__))
ROOT_DIR = path.realpath(path.join(PACKAGE_DIR, '..'))
ROOT_URL = 'http://pythoncz.herokuapp.com'
GITHUB_URL = (
'https://github.com/honzajavorek/python.cz/'
'blob/master/{template_folder}/{filename}'
)
app = Flask('web',
static_folder=path.join(ROOT_DIR, 'files'),
static_url_path='',
template_folder=path.join(ROOT_DIR, 'pages'))
app.config.from_object(__name__)
from . import views # NOQA<commit_msg>Fix newline at the end of file.<commit_after># -*- coding: utf-8 -*-
from os import path
from flask import Flask
PACKAGE_DIR = path.dirname(path.realpath(__file__))
ROOT_DIR = path.realpath(path.join(PACKAGE_DIR, '..'))
ROOT_URL = 'http://pythoncz.herokuapp.com'
GITHUB_URL = (
'https://github.com/honzajavorek/python.cz/'
'blob/master/{template_folder}/{filename}'
)
app = Flask('web',
static_folder=path.join(ROOT_DIR, 'files'),
static_url_path='',
template_folder=path.join(ROOT_DIR, 'pages'))
app.config.from_object(__name__)
from . import views # NOQA
|
5f716da231aa3f338300295695b1513aa404ae7d
|
lino_xl/lib/appypod/__init__.py
|
lino_xl/lib/appypod/__init__.py
|
# Copyright 2014-2019 Rumma & Ko Ltd
# License: BSD (see file COPYING for details)
"""
Adds functionality for generating printable documents using
LibreOffice and the `appy.pod <http://appyframework.org/pod.html>`__
package.
See also :ref:`lino.admin.appypod` and :doc:`/specs/appypod`.
"""
import six
from lino.api import ad, _
class Plugin(ad.Plugin):
verbose_name = _("Appy POD")
def get_requirements(self, site):
try:
import appy
# leave unchanged if it is already installed
except ImportError:
if six.PY3:
# yield "-e svn+https://svn.forge.pallavi.be/appy-dev/dev1#egg=appy"
yield "svn+https://svn.forge.pallavi.be/appy-dev/dev1#egg=appy"
else:
yield "appy"
def get_used_libs(self, html=None):
try:
# ~ import appy
from appy import version
version = version.verbose
except ImportError:
version = self.site.not_found_msg
yield ("Appy", version, "http://appyframework.org/pod.html")
|
# Copyright 2014-2019 Rumma & Ko Ltd
# License: BSD (see file COPYING for details)
"""
Adds functionality for generating printable documents using
LibreOffice and the `appy.pod <http://appyframework.org/pod.html>`__
package.
See also :ref:`lino.admin.appypod` and :doc:`/specs/appypod`.
"""
import six
from lino.api import ad, _
class Plugin(ad.Plugin):
verbose_name = _("Appy POD")
def get_requirements(self, site):
try:
import appy
# leave unchanged if it is already installed
except ImportError:
if six.PY3:
# yield "-e svn+https://svn.forge.pallavi.be/appy-dev/dev1#egg=appy"
yield "svn+http://svn.forge.pallavi.be/appy-dev/dev1#egg=appy"
else:
yield "appy"
def get_used_libs(self, html=None):
try:
# ~ import appy
from appy import version
version = version.verbose
except ImportError:
version = self.site.not_found_msg
yield ("Appy", version, "http://appyframework.org/pod.html")
|
Use http instead of https
|
Use http instead of https
|
Python
|
bsd-2-clause
|
lino-framework/xl,lino-framework/xl,lino-framework/xl,lino-framework/xl
|
# Copyright 2014-2019 Rumma & Ko Ltd
# License: BSD (see file COPYING for details)
"""
Adds functionality for generating printable documents using
LibreOffice and the `appy.pod <http://appyframework.org/pod.html>`__
package.
See also :ref:`lino.admin.appypod` and :doc:`/specs/appypod`.
"""
import six
from lino.api import ad, _
class Plugin(ad.Plugin):
verbose_name = _("Appy POD")
def get_requirements(self, site):
try:
import appy
# leave unchanged if it is already installed
except ImportError:
if six.PY3:
# yield "-e svn+https://svn.forge.pallavi.be/appy-dev/dev1#egg=appy"
yield "svn+https://svn.forge.pallavi.be/appy-dev/dev1#egg=appy"
else:
yield "appy"
def get_used_libs(self, html=None):
try:
# ~ import appy
from appy import version
version = version.verbose
except ImportError:
version = self.site.not_found_msg
yield ("Appy", version, "http://appyframework.org/pod.html")
Use http instead of https
|
# Copyright 2014-2019 Rumma & Ko Ltd
# License: BSD (see file COPYING for details)
"""
Adds functionality for generating printable documents using
LibreOffice and the `appy.pod <http://appyframework.org/pod.html>`__
package.
See also :ref:`lino.admin.appypod` and :doc:`/specs/appypod`.
"""
import six
from lino.api import ad, _
class Plugin(ad.Plugin):
verbose_name = _("Appy POD")
def get_requirements(self, site):
try:
import appy
# leave unchanged if it is already installed
except ImportError:
if six.PY3:
# yield "-e svn+https://svn.forge.pallavi.be/appy-dev/dev1#egg=appy"
yield "svn+http://svn.forge.pallavi.be/appy-dev/dev1#egg=appy"
else:
yield "appy"
def get_used_libs(self, html=None):
try:
# ~ import appy
from appy import version
version = version.verbose
except ImportError:
version = self.site.not_found_msg
yield ("Appy", version, "http://appyframework.org/pod.html")
|
<commit_before># Copyright 2014-2019 Rumma & Ko Ltd
# License: BSD (see file COPYING for details)
"""
Adds functionality for generating printable documents using
LibreOffice and the `appy.pod <http://appyframework.org/pod.html>`__
package.
See also :ref:`lino.admin.appypod` and :doc:`/specs/appypod`.
"""
import six
from lino.api import ad, _
class Plugin(ad.Plugin):
verbose_name = _("Appy POD")
def get_requirements(self, site):
try:
import appy
# leave unchanged if it is already installed
except ImportError:
if six.PY3:
# yield "-e svn+https://svn.forge.pallavi.be/appy-dev/dev1#egg=appy"
yield "svn+https://svn.forge.pallavi.be/appy-dev/dev1#egg=appy"
else:
yield "appy"
def get_used_libs(self, html=None):
try:
# ~ import appy
from appy import version
version = version.verbose
except ImportError:
version = self.site.not_found_msg
yield ("Appy", version, "http://appyframework.org/pod.html")
<commit_msg>Use http instead of https<commit_after>
|
# Copyright 2014-2019 Rumma & Ko Ltd
# License: BSD (see file COPYING for details)
"""
Adds functionality for generating printable documents using
LibreOffice and the `appy.pod <http://appyframework.org/pod.html>`__
package.
See also :ref:`lino.admin.appypod` and :doc:`/specs/appypod`.
"""
import six
from lino.api import ad, _
class Plugin(ad.Plugin):
verbose_name = _("Appy POD")
def get_requirements(self, site):
try:
import appy
# leave unchanged if it is already installed
except ImportError:
if six.PY3:
# yield "-e svn+https://svn.forge.pallavi.be/appy-dev/dev1#egg=appy"
yield "svn+http://svn.forge.pallavi.be/appy-dev/dev1#egg=appy"
else:
yield "appy"
def get_used_libs(self, html=None):
try:
# ~ import appy
from appy import version
version = version.verbose
except ImportError:
version = self.site.not_found_msg
yield ("Appy", version, "http://appyframework.org/pod.html")
|
# Copyright 2014-2019 Rumma & Ko Ltd
# License: BSD (see file COPYING for details)
"""
Adds functionality for generating printable documents using
LibreOffice and the `appy.pod <http://appyframework.org/pod.html>`__
package.
See also :ref:`lino.admin.appypod` and :doc:`/specs/appypod`.
"""
import six
from lino.api import ad, _
class Plugin(ad.Plugin):
verbose_name = _("Appy POD")
def get_requirements(self, site):
try:
import appy
# leave unchanged if it is already installed
except ImportError:
if six.PY3:
# yield "-e svn+https://svn.forge.pallavi.be/appy-dev/dev1#egg=appy"
yield "svn+https://svn.forge.pallavi.be/appy-dev/dev1#egg=appy"
else:
yield "appy"
def get_used_libs(self, html=None):
try:
# ~ import appy
from appy import version
version = version.verbose
except ImportError:
version = self.site.not_found_msg
yield ("Appy", version, "http://appyframework.org/pod.html")
Use http instead of https# Copyright 2014-2019 Rumma & Ko Ltd
# License: BSD (see file COPYING for details)
"""
Adds functionality for generating printable documents using
LibreOffice and the `appy.pod <http://appyframework.org/pod.html>`__
package.
See also :ref:`lino.admin.appypod` and :doc:`/specs/appypod`.
"""
import six
from lino.api import ad, _
class Plugin(ad.Plugin):
verbose_name = _("Appy POD")
def get_requirements(self, site):
try:
import appy
# leave unchanged if it is already installed
except ImportError:
if six.PY3:
# yield "-e svn+https://svn.forge.pallavi.be/appy-dev/dev1#egg=appy"
yield "svn+http://svn.forge.pallavi.be/appy-dev/dev1#egg=appy"
else:
yield "appy"
def get_used_libs(self, html=None):
try:
# ~ import appy
from appy import version
version = version.verbose
except ImportError:
version = self.site.not_found_msg
yield ("Appy", version, "http://appyframework.org/pod.html")
|
<commit_before># Copyright 2014-2019 Rumma & Ko Ltd
# License: BSD (see file COPYING for details)
"""
Adds functionality for generating printable documents using
LibreOffice and the `appy.pod <http://appyframework.org/pod.html>`__
package.
See also :ref:`lino.admin.appypod` and :doc:`/specs/appypod`.
"""
import six
from lino.api import ad, _
class Plugin(ad.Plugin):
verbose_name = _("Appy POD")
def get_requirements(self, site):
try:
import appy
# leave unchanged if it is already installed
except ImportError:
if six.PY3:
# yield "-e svn+https://svn.forge.pallavi.be/appy-dev/dev1#egg=appy"
yield "svn+https://svn.forge.pallavi.be/appy-dev/dev1#egg=appy"
else:
yield "appy"
def get_used_libs(self, html=None):
try:
# ~ import appy
from appy import version
version = version.verbose
except ImportError:
version = self.site.not_found_msg
yield ("Appy", version, "http://appyframework.org/pod.html")
<commit_msg>Use http instead of https<commit_after># Copyright 2014-2019 Rumma & Ko Ltd
# License: BSD (see file COPYING for details)
"""
Adds functionality for generating printable documents using
LibreOffice and the `appy.pod <http://appyframework.org/pod.html>`__
package.
See also :ref:`lino.admin.appypod` and :doc:`/specs/appypod`.
"""
import six
from lino.api import ad, _
class Plugin(ad.Plugin):
verbose_name = _("Appy POD")
def get_requirements(self, site):
try:
import appy
# leave unchanged if it is already installed
except ImportError:
if six.PY3:
# yield "-e svn+https://svn.forge.pallavi.be/appy-dev/dev1#egg=appy"
yield "svn+http://svn.forge.pallavi.be/appy-dev/dev1#egg=appy"
else:
yield "appy"
def get_used_libs(self, html=None):
try:
# ~ import appy
from appy import version
version = version.verbose
except ImportError:
version = self.site.not_found_msg
yield ("Appy", version, "http://appyframework.org/pod.html")
|
b49f733d675d537779bed931d0a079888a83a735
|
mpfmonitor/_version.py
|
mpfmonitor/_version.py
|
# mpf-monitor
__version__ = '0.3.0-dev.1'
__short_version__ = '0.3'
__bcp_version__ = '1.1'
__config_version__ = '4'
__mpf_version_required__ = '0.33.0'
version = "MPF Monitor v{} (config_version={}, BCP v{}, Requires MPF v{})".format(
__version__, __config_version__, __bcp_version__, __mpf_version_required__)
|
# mpf-monitor
__version__ = '0.2.0-dev.3'
__short_version__ = '0.2'
__bcp_version__ = '1.1'
__config_version__ = '4'
__mpf_version_required__ = '0.33.0.dev15'
version = "MPF Monitor v{} (config_version={}, BCP v{}, Requires MPF v{})".format(
__version__, __config_version__, __bcp_version__, __mpf_version_required__)
|
Revert "Bump dev version to 0.3.0-dev.1"
|
Revert "Bump dev version to 0.3.0-dev.1"
This reverts commit facc1caaca87e680321be0654882d6c5570bc2ad.
|
Python
|
mit
|
missionpinball/mpf-monitor
|
# mpf-monitor
__version__ = '0.3.0-dev.1'
__short_version__ = '0.3'
__bcp_version__ = '1.1'
__config_version__ = '4'
__mpf_version_required__ = '0.33.0'
version = "MPF Monitor v{} (config_version={}, BCP v{}, Requires MPF v{})".format(
__version__, __config_version__, __bcp_version__, __mpf_version_required__)
Revert "Bump dev version to 0.3.0-dev.1"
This reverts commit facc1caaca87e680321be0654882d6c5570bc2ad.
|
# mpf-monitor
__version__ = '0.2.0-dev.3'
__short_version__ = '0.2'
__bcp_version__ = '1.1'
__config_version__ = '4'
__mpf_version_required__ = '0.33.0.dev15'
version = "MPF Monitor v{} (config_version={}, BCP v{}, Requires MPF v{})".format(
__version__, __config_version__, __bcp_version__, __mpf_version_required__)
|
<commit_before># mpf-monitor
__version__ = '0.3.0-dev.1'
__short_version__ = '0.3'
__bcp_version__ = '1.1'
__config_version__ = '4'
__mpf_version_required__ = '0.33.0'
version = "MPF Monitor v{} (config_version={}, BCP v{}, Requires MPF v{})".format(
__version__, __config_version__, __bcp_version__, __mpf_version_required__)
<commit_msg>Revert "Bump dev version to 0.3.0-dev.1"
This reverts commit facc1caaca87e680321be0654882d6c5570bc2ad.<commit_after>
|
# mpf-monitor
__version__ = '0.2.0-dev.3'
__short_version__ = '0.2'
__bcp_version__ = '1.1'
__config_version__ = '4'
__mpf_version_required__ = '0.33.0.dev15'
version = "MPF Monitor v{} (config_version={}, BCP v{}, Requires MPF v{})".format(
__version__, __config_version__, __bcp_version__, __mpf_version_required__)
|
# mpf-monitor
__version__ = '0.3.0-dev.1'
__short_version__ = '0.3'
__bcp_version__ = '1.1'
__config_version__ = '4'
__mpf_version_required__ = '0.33.0'
version = "MPF Monitor v{} (config_version={}, BCP v{}, Requires MPF v{})".format(
__version__, __config_version__, __bcp_version__, __mpf_version_required__)
Revert "Bump dev version to 0.3.0-dev.1"
This reverts commit facc1caaca87e680321be0654882d6c5570bc2ad.# mpf-monitor
__version__ = '0.2.0-dev.3'
__short_version__ = '0.2'
__bcp_version__ = '1.1'
__config_version__ = '4'
__mpf_version_required__ = '0.33.0.dev15'
version = "MPF Monitor v{} (config_version={}, BCP v{}, Requires MPF v{})".format(
__version__, __config_version__, __bcp_version__, __mpf_version_required__)
|
<commit_before># mpf-monitor
__version__ = '0.3.0-dev.1'
__short_version__ = '0.3'
__bcp_version__ = '1.1'
__config_version__ = '4'
__mpf_version_required__ = '0.33.0'
version = "MPF Monitor v{} (config_version={}, BCP v{}, Requires MPF v{})".format(
__version__, __config_version__, __bcp_version__, __mpf_version_required__)
<commit_msg>Revert "Bump dev version to 0.3.0-dev.1"
This reverts commit facc1caaca87e680321be0654882d6c5570bc2ad.<commit_after># mpf-monitor
__version__ = '0.2.0-dev.3'
__short_version__ = '0.2'
__bcp_version__ = '1.1'
__config_version__ = '4'
__mpf_version_required__ = '0.33.0.dev15'
version = "MPF Monitor v{} (config_version={}, BCP v{}, Requires MPF v{})".format(
__version__, __config_version__, __bcp_version__, __mpf_version_required__)
|
3af2a5b6eda4af972e3a208e727483384f313cb9
|
octotribble/csv2tex.py
|
octotribble/csv2tex.py
|
#!/usr/bin/env python
"""Convert a CSV table into a latex tabular using pandas."""
import argparse
import pandas as pd
def convert(filename, transpose=False):
"""convert csv to tex table."""
df = pd.read_csv(filename)
if transpose:
df = df.transpose()
tex_name = filename.replace(".csv", "_transpose.tex")
else:
tex_name = filename.replace(".csv", ".tex")
with open(tex_name, "w") as f:
f.write(r"\begin{table}")
f.write("\n")
f.write(r"\label{}")
f.write("\n")
f.write(r"\caption{}")
f.write("\n")
f.write(df.to_latex(na_rep="-"))
f.write(r"\end{table}")
f.write("\n")
if __name__ == "__main__":
parser = argparse.ArgumentParser(description="Convert csv to latex tabular")
parser.add_argument("filename", help="Name of csv file", type=str)
parser.add_argument(
"-t", "--transpose", help="Transpose table", action="store_true"
)
args = parser.parse_args()
convert(args.filename, args.transpose)
|
#!/usr/bin/env python
"""Convert a CSV table into a latex tabular using pandas."""
import argparse
import pandas as pd
def convert(filename, transpose=False):
"""convert csv to tex table."""
df = pd.read_csv(filename)
if transpose:
df = df.transpose()
tex_name = filename.replace(".csv", "_transpose.tex")
index = True
else:
tex_name = filename.replace(".csv", ".tex")
index=False
assert tex_name != filename, "This will overwrite the file, did you pass in a csv?"
latex = df.to_latex(na_rep="-", index=index)
with open(tex_name, "w") as f:
f.write(r"\begin{table}")
f.write("\n")
f.write(r"\label{}")
f.write("\n")
f.write(r"\caption{}")
f.write("\n")
f.write(latex)
f.write(r"\end{table}")
f.write("\n")
if __name__ == "__main__":
parser = argparse.ArgumentParser(description="Convert csv to latex tabular")
parser.add_argument("filename", help="Name of csv file", type=str)
parser.add_argument(
"-t", "--transpose", help="Transpose table", action="store_true"
)
args = parser.parse_args()
convert(args.filename, args.transpose)
|
Remove index and check filename
|
Remove index and check filename
|
Python
|
mit
|
jason-neal/equanimous-octo-tribble,jason-neal/equanimous-octo-tribble,jason-neal/equanimous-octo-tribble
|
#!/usr/bin/env python
"""Convert a CSV table into a latex tabular using pandas."""
import argparse
import pandas as pd
def convert(filename, transpose=False):
"""convert csv to tex table."""
df = pd.read_csv(filename)
if transpose:
df = df.transpose()
tex_name = filename.replace(".csv", "_transpose.tex")
else:
tex_name = filename.replace(".csv", ".tex")
with open(tex_name, "w") as f:
f.write(r"\begin{table}")
f.write("\n")
f.write(r"\label{}")
f.write("\n")
f.write(r"\caption{}")
f.write("\n")
f.write(df.to_latex(na_rep="-"))
f.write(r"\end{table}")
f.write("\n")
if __name__ == "__main__":
parser = argparse.ArgumentParser(description="Convert csv to latex tabular")
parser.add_argument("filename", help="Name of csv file", type=str)
parser.add_argument(
"-t", "--transpose", help="Transpose table", action="store_true"
)
args = parser.parse_args()
convert(args.filename, args.transpose)
Remove index and check filename
|
#!/usr/bin/env python
"""Convert a CSV table into a latex tabular using pandas."""
import argparse
import pandas as pd
def convert(filename, transpose=False):
"""convert csv to tex table."""
df = pd.read_csv(filename)
if transpose:
df = df.transpose()
tex_name = filename.replace(".csv", "_transpose.tex")
index = True
else:
tex_name = filename.replace(".csv", ".tex")
index=False
assert tex_name != filename, "This will overwrite the file, did you pass in a csv?"
latex = df.to_latex(na_rep="-", index=index)
with open(tex_name, "w") as f:
f.write(r"\begin{table}")
f.write("\n")
f.write(r"\label{}")
f.write("\n")
f.write(r"\caption{}")
f.write("\n")
f.write(latex)
f.write(r"\end{table}")
f.write("\n")
if __name__ == "__main__":
parser = argparse.ArgumentParser(description="Convert csv to latex tabular")
parser.add_argument("filename", help="Name of csv file", type=str)
parser.add_argument(
"-t", "--transpose", help="Transpose table", action="store_true"
)
args = parser.parse_args()
convert(args.filename, args.transpose)
|
<commit_before>#!/usr/bin/env python
"""Convert a CSV table into a latex tabular using pandas."""
import argparse
import pandas as pd
def convert(filename, transpose=False):
"""convert csv to tex table."""
df = pd.read_csv(filename)
if transpose:
df = df.transpose()
tex_name = filename.replace(".csv", "_transpose.tex")
else:
tex_name = filename.replace(".csv", ".tex")
with open(tex_name, "w") as f:
f.write(r"\begin{table}")
f.write("\n")
f.write(r"\label{}")
f.write("\n")
f.write(r"\caption{}")
f.write("\n")
f.write(df.to_latex(na_rep="-"))
f.write(r"\end{table}")
f.write("\n")
if __name__ == "__main__":
parser = argparse.ArgumentParser(description="Convert csv to latex tabular")
parser.add_argument("filename", help="Name of csv file", type=str)
parser.add_argument(
"-t", "--transpose", help="Transpose table", action="store_true"
)
args = parser.parse_args()
convert(args.filename, args.transpose)
<commit_msg>Remove index and check filename<commit_after>
|
#!/usr/bin/env python
"""Convert a CSV table into a latex tabular using pandas."""
import argparse
import pandas as pd
def convert(filename, transpose=False):
"""convert csv to tex table."""
df = pd.read_csv(filename)
if transpose:
df = df.transpose()
tex_name = filename.replace(".csv", "_transpose.tex")
index = True
else:
tex_name = filename.replace(".csv", ".tex")
index=False
assert tex_name != filename, "This will overwrite the file, did you pass in a csv?"
latex = df.to_latex(na_rep="-", index=index)
with open(tex_name, "w") as f:
f.write(r"\begin{table}")
f.write("\n")
f.write(r"\label{}")
f.write("\n")
f.write(r"\caption{}")
f.write("\n")
f.write(latex)
f.write(r"\end{table}")
f.write("\n")
if __name__ == "__main__":
parser = argparse.ArgumentParser(description="Convert csv to latex tabular")
parser.add_argument("filename", help="Name of csv file", type=str)
parser.add_argument(
"-t", "--transpose", help="Transpose table", action="store_true"
)
args = parser.parse_args()
convert(args.filename, args.transpose)
|
#!/usr/bin/env python
"""Convert a CSV table into a latex tabular using pandas."""
import argparse
import pandas as pd
def convert(filename, transpose=False):
"""convert csv to tex table."""
df = pd.read_csv(filename)
if transpose:
df = df.transpose()
tex_name = filename.replace(".csv", "_transpose.tex")
else:
tex_name = filename.replace(".csv", ".tex")
with open(tex_name, "w") as f:
f.write(r"\begin{table}")
f.write("\n")
f.write(r"\label{}")
f.write("\n")
f.write(r"\caption{}")
f.write("\n")
f.write(df.to_latex(na_rep="-"))
f.write(r"\end{table}")
f.write("\n")
if __name__ == "__main__":
parser = argparse.ArgumentParser(description="Convert csv to latex tabular")
parser.add_argument("filename", help="Name of csv file", type=str)
parser.add_argument(
"-t", "--transpose", help="Transpose table", action="store_true"
)
args = parser.parse_args()
convert(args.filename, args.transpose)
Remove index and check filename#!/usr/bin/env python
"""Convert a CSV table into a latex tabular using pandas."""
import argparse
import pandas as pd
def convert(filename, transpose=False):
"""convert csv to tex table."""
df = pd.read_csv(filename)
if transpose:
df = df.transpose()
tex_name = filename.replace(".csv", "_transpose.tex")
index = True
else:
tex_name = filename.replace(".csv", ".tex")
index=False
assert tex_name != filename, "This will overwrite the file, did you pass in a csv?"
latex = df.to_latex(na_rep="-", index=index)
with open(tex_name, "w") as f:
f.write(r"\begin{table}")
f.write("\n")
f.write(r"\label{}")
f.write("\n")
f.write(r"\caption{}")
f.write("\n")
f.write(latex)
f.write(r"\end{table}")
f.write("\n")
if __name__ == "__main__":
parser = argparse.ArgumentParser(description="Convert csv to latex tabular")
parser.add_argument("filename", help="Name of csv file", type=str)
parser.add_argument(
"-t", "--transpose", help="Transpose table", action="store_true"
)
args = parser.parse_args()
convert(args.filename, args.transpose)
|
<commit_before>#!/usr/bin/env python
"""Convert a CSV table into a latex tabular using pandas."""
import argparse
import pandas as pd
def convert(filename, transpose=False):
"""convert csv to tex table."""
df = pd.read_csv(filename)
if transpose:
df = df.transpose()
tex_name = filename.replace(".csv", "_transpose.tex")
else:
tex_name = filename.replace(".csv", ".tex")
with open(tex_name, "w") as f:
f.write(r"\begin{table}")
f.write("\n")
f.write(r"\label{}")
f.write("\n")
f.write(r"\caption{}")
f.write("\n")
f.write(df.to_latex(na_rep="-"))
f.write(r"\end{table}")
f.write("\n")
if __name__ == "__main__":
parser = argparse.ArgumentParser(description="Convert csv to latex tabular")
parser.add_argument("filename", help="Name of csv file", type=str)
parser.add_argument(
"-t", "--transpose", help="Transpose table", action="store_true"
)
args = parser.parse_args()
convert(args.filename, args.transpose)
<commit_msg>Remove index and check filename<commit_after>#!/usr/bin/env python
"""Convert a CSV table into a latex tabular using pandas."""
import argparse
import pandas as pd
def convert(filename, transpose=False):
"""convert csv to tex table."""
df = pd.read_csv(filename)
if transpose:
df = df.transpose()
tex_name = filename.replace(".csv", "_transpose.tex")
index = True
else:
tex_name = filename.replace(".csv", ".tex")
index=False
assert tex_name != filename, "This will overwrite the file, did you pass in a csv?"
latex = df.to_latex(na_rep="-", index=index)
with open(tex_name, "w") as f:
f.write(r"\begin{table}")
f.write("\n")
f.write(r"\label{}")
f.write("\n")
f.write(r"\caption{}")
f.write("\n")
f.write(latex)
f.write(r"\end{table}")
f.write("\n")
if __name__ == "__main__":
parser = argparse.ArgumentParser(description="Convert csv to latex tabular")
parser.add_argument("filename", help="Name of csv file", type=str)
parser.add_argument(
"-t", "--transpose", help="Transpose table", action="store_true"
)
args = parser.parse_args()
convert(args.filename, args.transpose)
|
75b50ffcb6575e38e6792356dd58612089ee4f55
|
django_mercadopago/views.py
|
django_mercadopago/views.py
|
import logging
from django.conf import settings
from django.http import HttpResponse
from django.views.decorators.csrf import csrf_exempt
from .models import Notification, Account
logger = logging.getLogger(__name__)
# Maybe use a form for this? :D
@csrf_exempt
def create_notification(request, slug):
topic = request.GET.get('topic', None)
resource_id = request.GET.get('id', None)
if topic is None:
return HttpResponse(
'<h1>400 Bad Request.</h1>'
'Missing parameter topic',
status=400
)
if resource_id is None:
return HttpResponse(
'<h1>400 Bad Request.</h1>'
'Missing parameter id',
status=400
)
if topic == 'merchant_order':
topic = Notification.TOPIC_ORDER
elif topic == 'payment':
topic = Notification.TOPIC_PAYMENT
else:
return HttpResponse('invalid topic', status=400)
owner = Account.objects.get(slug=slug)
notification, created = Notification.objects.get_or_create(
topic=topic,
resource_id=resource_id,
owner=owner,
)
if not created:
notification.processed = False
notification.save()
if not settings.MERCADOPAGO_ASYNC:
notification.process()
# TODO: Else add to some queue?
return HttpResponse("<h1>200 OK</h1>", status=201)
|
import logging
from django.conf import settings
from django.http import HttpResponse
from django.views.decorators.csrf import csrf_exempt
from .models import Notification, Account
logger = logging.getLogger(__name__)
# Maybe use a form for this? :D
@csrf_exempt
def create_notification(request, slug):
topic = request.GET.get('topic', None)
resource_id = request.GET.get('id', None)
if topic is None:
return HttpResponse(
'<h1>400 Bad Request.</h1>'
'Missing parameter topic',
status=400
)
if resource_id is None:
return HttpResponse(
'<h1>400 Bad Request.</h1>'
'Missing parameter id',
status=400
)
if topic == 'merchant_order':
topic = Notification.TOPIC_ORDER
elif topic == 'payment':
topic = Notification.TOPIC_PAYMENT
else:
return HttpResponse('invalid topic', status=400)
try:
owner = Account.objects.get(slug=slug)
except Account.DoesNotExist:
return HttpResponse('Unknown account/slug', status=400)
notification, created = Notification.objects.get_or_create(
topic=topic,
resource_id=resource_id,
owner=owner,
)
if not created:
notification.processed = False
notification.save()
if not settings.MERCADOPAGO_ASYNC:
notification.process()
# TODO: Else add to some queue?
return HttpResponse("<h1>200 OK</h1>", status=201)
|
Return 400 for inexistant accounts
|
Return 400 for inexistant accounts
|
Python
|
isc
|
asermax/django-mercadopago
|
import logging
from django.conf import settings
from django.http import HttpResponse
from django.views.decorators.csrf import csrf_exempt
from .models import Notification, Account
logger = logging.getLogger(__name__)
# Maybe use a form for this? :D
@csrf_exempt
def create_notification(request, slug):
topic = request.GET.get('topic', None)
resource_id = request.GET.get('id', None)
if topic is None:
return HttpResponse(
'<h1>400 Bad Request.</h1>'
'Missing parameter topic',
status=400
)
if resource_id is None:
return HttpResponse(
'<h1>400 Bad Request.</h1>'
'Missing parameter id',
status=400
)
if topic == 'merchant_order':
topic = Notification.TOPIC_ORDER
elif topic == 'payment':
topic = Notification.TOPIC_PAYMENT
else:
return HttpResponse('invalid topic', status=400)
owner = Account.objects.get(slug=slug)
notification, created = Notification.objects.get_or_create(
topic=topic,
resource_id=resource_id,
owner=owner,
)
if not created:
notification.processed = False
notification.save()
if not settings.MERCADOPAGO_ASYNC:
notification.process()
# TODO: Else add to some queue?
return HttpResponse("<h1>200 OK</h1>", status=201)
Return 400 for inexistant accounts
|
import logging
from django.conf import settings
from django.http import HttpResponse
from django.views.decorators.csrf import csrf_exempt
from .models import Notification, Account
logger = logging.getLogger(__name__)
# Maybe use a form for this? :D
@csrf_exempt
def create_notification(request, slug):
topic = request.GET.get('topic', None)
resource_id = request.GET.get('id', None)
if topic is None:
return HttpResponse(
'<h1>400 Bad Request.</h1>'
'Missing parameter topic',
status=400
)
if resource_id is None:
return HttpResponse(
'<h1>400 Bad Request.</h1>'
'Missing parameter id',
status=400
)
if topic == 'merchant_order':
topic = Notification.TOPIC_ORDER
elif topic == 'payment':
topic = Notification.TOPIC_PAYMENT
else:
return HttpResponse('invalid topic', status=400)
try:
owner = Account.objects.get(slug=slug)
except Account.DoesNotExist:
return HttpResponse('Unknown account/slug', status=400)
notification, created = Notification.objects.get_or_create(
topic=topic,
resource_id=resource_id,
owner=owner,
)
if not created:
notification.processed = False
notification.save()
if not settings.MERCADOPAGO_ASYNC:
notification.process()
# TODO: Else add to some queue?
return HttpResponse("<h1>200 OK</h1>", status=201)
|
<commit_before>import logging
from django.conf import settings
from django.http import HttpResponse
from django.views.decorators.csrf import csrf_exempt
from .models import Notification, Account
logger = logging.getLogger(__name__)
# Maybe use a form for this? :D
@csrf_exempt
def create_notification(request, slug):
topic = request.GET.get('topic', None)
resource_id = request.GET.get('id', None)
if topic is None:
return HttpResponse(
'<h1>400 Bad Request.</h1>'
'Missing parameter topic',
status=400
)
if resource_id is None:
return HttpResponse(
'<h1>400 Bad Request.</h1>'
'Missing parameter id',
status=400
)
if topic == 'merchant_order':
topic = Notification.TOPIC_ORDER
elif topic == 'payment':
topic = Notification.TOPIC_PAYMENT
else:
return HttpResponse('invalid topic', status=400)
owner = Account.objects.get(slug=slug)
notification, created = Notification.objects.get_or_create(
topic=topic,
resource_id=resource_id,
owner=owner,
)
if not created:
notification.processed = False
notification.save()
if not settings.MERCADOPAGO_ASYNC:
notification.process()
# TODO: Else add to some queue?
return HttpResponse("<h1>200 OK</h1>", status=201)
<commit_msg>Return 400 for inexistant accounts<commit_after>
|
import logging
from django.conf import settings
from django.http import HttpResponse
from django.views.decorators.csrf import csrf_exempt
from .models import Notification, Account
logger = logging.getLogger(__name__)
# Maybe use a form for this? :D
@csrf_exempt
def create_notification(request, slug):
topic = request.GET.get('topic', None)
resource_id = request.GET.get('id', None)
if topic is None:
return HttpResponse(
'<h1>400 Bad Request.</h1>'
'Missing parameter topic',
status=400
)
if resource_id is None:
return HttpResponse(
'<h1>400 Bad Request.</h1>'
'Missing parameter id',
status=400
)
if topic == 'merchant_order':
topic = Notification.TOPIC_ORDER
elif topic == 'payment':
topic = Notification.TOPIC_PAYMENT
else:
return HttpResponse('invalid topic', status=400)
try:
owner = Account.objects.get(slug=slug)
except Account.DoesNotExist:
return HttpResponse('Unknown account/slug', status=400)
notification, created = Notification.objects.get_or_create(
topic=topic,
resource_id=resource_id,
owner=owner,
)
if not created:
notification.processed = False
notification.save()
if not settings.MERCADOPAGO_ASYNC:
notification.process()
# TODO: Else add to some queue?
return HttpResponse("<h1>200 OK</h1>", status=201)
|
import logging
from django.conf import settings
from django.http import HttpResponse
from django.views.decorators.csrf import csrf_exempt
from .models import Notification, Account
logger = logging.getLogger(__name__)
# Maybe use a form for this? :D
@csrf_exempt
def create_notification(request, slug):
topic = request.GET.get('topic', None)
resource_id = request.GET.get('id', None)
if topic is None:
return HttpResponse(
'<h1>400 Bad Request.</h1>'
'Missing parameter topic',
status=400
)
if resource_id is None:
return HttpResponse(
'<h1>400 Bad Request.</h1>'
'Missing parameter id',
status=400
)
if topic == 'merchant_order':
topic = Notification.TOPIC_ORDER
elif topic == 'payment':
topic = Notification.TOPIC_PAYMENT
else:
return HttpResponse('invalid topic', status=400)
owner = Account.objects.get(slug=slug)
notification, created = Notification.objects.get_or_create(
topic=topic,
resource_id=resource_id,
owner=owner,
)
if not created:
notification.processed = False
notification.save()
if not settings.MERCADOPAGO_ASYNC:
notification.process()
# TODO: Else add to some queue?
return HttpResponse("<h1>200 OK</h1>", status=201)
Return 400 for inexistant accountsimport logging
from django.conf import settings
from django.http import HttpResponse
from django.views.decorators.csrf import csrf_exempt
from .models import Notification, Account
logger = logging.getLogger(__name__)
# Maybe use a form for this? :D
@csrf_exempt
def create_notification(request, slug):
topic = request.GET.get('topic', None)
resource_id = request.GET.get('id', None)
if topic is None:
return HttpResponse(
'<h1>400 Bad Request.</h1>'
'Missing parameter topic',
status=400
)
if resource_id is None:
return HttpResponse(
'<h1>400 Bad Request.</h1>'
'Missing parameter id',
status=400
)
if topic == 'merchant_order':
topic = Notification.TOPIC_ORDER
elif topic == 'payment':
topic = Notification.TOPIC_PAYMENT
else:
return HttpResponse('invalid topic', status=400)
try:
owner = Account.objects.get(slug=slug)
except Account.DoesNotExist:
return HttpResponse('Unknown account/slug', status=400)
notification, created = Notification.objects.get_or_create(
topic=topic,
resource_id=resource_id,
owner=owner,
)
if not created:
notification.processed = False
notification.save()
if not settings.MERCADOPAGO_ASYNC:
notification.process()
# TODO: Else add to some queue?
return HttpResponse("<h1>200 OK</h1>", status=201)
|
<commit_before>import logging
from django.conf import settings
from django.http import HttpResponse
from django.views.decorators.csrf import csrf_exempt
from .models import Notification, Account
logger = logging.getLogger(__name__)
# Maybe use a form for this? :D
@csrf_exempt
def create_notification(request, slug):
topic = request.GET.get('topic', None)
resource_id = request.GET.get('id', None)
if topic is None:
return HttpResponse(
'<h1>400 Bad Request.</h1>'
'Missing parameter topic',
status=400
)
if resource_id is None:
return HttpResponse(
'<h1>400 Bad Request.</h1>'
'Missing parameter id',
status=400
)
if topic == 'merchant_order':
topic = Notification.TOPIC_ORDER
elif topic == 'payment':
topic = Notification.TOPIC_PAYMENT
else:
return HttpResponse('invalid topic', status=400)
owner = Account.objects.get(slug=slug)
notification, created = Notification.objects.get_or_create(
topic=topic,
resource_id=resource_id,
owner=owner,
)
if not created:
notification.processed = False
notification.save()
if not settings.MERCADOPAGO_ASYNC:
notification.process()
# TODO: Else add to some queue?
return HttpResponse("<h1>200 OK</h1>", status=201)
<commit_msg>Return 400 for inexistant accounts<commit_after>import logging
from django.conf import settings
from django.http import HttpResponse
from django.views.decorators.csrf import csrf_exempt
from .models import Notification, Account
logger = logging.getLogger(__name__)
# Maybe use a form for this? :D
@csrf_exempt
def create_notification(request, slug):
topic = request.GET.get('topic', None)
resource_id = request.GET.get('id', None)
if topic is None:
return HttpResponse(
'<h1>400 Bad Request.</h1>'
'Missing parameter topic',
status=400
)
if resource_id is None:
return HttpResponse(
'<h1>400 Bad Request.</h1>'
'Missing parameter id',
status=400
)
if topic == 'merchant_order':
topic = Notification.TOPIC_ORDER
elif topic == 'payment':
topic = Notification.TOPIC_PAYMENT
else:
return HttpResponse('invalid topic', status=400)
try:
owner = Account.objects.get(slug=slug)
except Account.DoesNotExist:
return HttpResponse('Unknown account/slug', status=400)
notification, created = Notification.objects.get_or_create(
topic=topic,
resource_id=resource_id,
owner=owner,
)
if not created:
notification.processed = False
notification.save()
if not settings.MERCADOPAGO_ASYNC:
notification.process()
# TODO: Else add to some queue?
return HttpResponse("<h1>200 OK</h1>", status=201)
|
b639e094f0ac9feb008c0d13deb26c55bbb50793
|
git_gutter_change.py
|
git_gutter_change.py
|
import sublime
import sublime_plugin
try:
from GitGutter.view_collection import ViewCollection
except ImportError:
from view_collection import ViewCollection
class GitGutterBaseChangeCommand(sublime_plugin.WindowCommand):
def lines_to_blocks(self, lines):
blocks = []
last_line = -2
for line in lines:
if line > last_line+1:
blocks.append(line)
last_line = line
return blocks
def run(self):
view = self.window.active_view()
inserted, modified, deleted = ViewCollection.diff(view)
inserted = self.lines_to_blocks(inserted)
modified = self.lines_to_blocks(modified)
all_changes = sorted(inserted + modified + deleted)
row, col = view.rowcol(view.sel()[0].begin())
current_row = row + 1
line = self.jump(all_changes, current_row)
self.window.active_view().run_command("goto_line", {"line": line})
class GitGutterNextChangeCommand(GitGutterBaseChangeCommand):
def jump(self, all_changes, current_row):
return next((change for change in all_changes
if change > current_row), current_row)
class GitGutterPrevChangeCommand(GitGutterBaseChangeCommand):
def jump(self, all_changes, current_row):
return next((change for change in reversed(all_changes)
if change < current_row), current_row)
|
import sublime
import sublime_plugin
try:
from GitGutter.view_collection import ViewCollection
except ImportError:
from view_collection import ViewCollection
class GitGutterBaseChangeCommand(sublime_plugin.WindowCommand):
def lines_to_blocks(self, lines):
blocks = []
last_line = -2
for line in lines:
if line > last_line+1:
blocks.append(line)
last_line = line
return blocks
def run(self):
view = self.window.active_view()
inserted, modified, deleted = ViewCollection.diff(view)
inserted = self.lines_to_blocks(inserted)
modified = self.lines_to_blocks(modified)
all_changes = sorted(inserted + modified + deleted)
row, col = view.rowcol(view.sel()[0].begin())
current_row = row + 1
line = self.jump(all_changes, current_row)
self.window.active_view().run_command("goto_line", {"line": line})
class GitGutterNextChangeCommand(GitGutterBaseChangeCommand):
def jump(self, all_changes, current_row):
return next((change for change in all_changes
if change > current_row), all_changes[0])
class GitGutterPrevChangeCommand(GitGutterBaseChangeCommand):
def jump(self, all_changes, current_row):
return next((change for change in reversed(all_changes)
if change < current_row), all_changes[-1])
|
Make jumping between changes loop back around
|
Make jumping between changes loop back around
|
Python
|
mit
|
michaelhogg/GitGutter,biodamasceno/GitGutter,biodamasceno/GitGutter,robfrawley/sublime-git-gutter,robfrawley/sublime-git-gutter,robfrawley/sublime-git-gutter,robfrawley/sublime-git-gutter,biodamasceno/GitGutter,natecavanaugh/GitGutter,tushortz/GitGutter,michaelhogg/GitGutter,akpersad/GitGutter,akpersad/GitGutter,natecavanaugh/GitGutter,akpersad/GitGutter,tushortz/GitGutter,akpersad/GitGutter,michaelhogg/GitGutter,tushortz/GitGutter,jisaacks/GitGutter,biodamasceno/GitGutter,natecavanaugh/GitGutter,michaelhogg/GitGutter,tushortz/GitGutter,natecavanaugh/GitGutter
|
import sublime
import sublime_plugin
try:
from GitGutter.view_collection import ViewCollection
except ImportError:
from view_collection import ViewCollection
class GitGutterBaseChangeCommand(sublime_plugin.WindowCommand):
def lines_to_blocks(self, lines):
blocks = []
last_line = -2
for line in lines:
if line > last_line+1:
blocks.append(line)
last_line = line
return blocks
def run(self):
view = self.window.active_view()
inserted, modified, deleted = ViewCollection.diff(view)
inserted = self.lines_to_blocks(inserted)
modified = self.lines_to_blocks(modified)
all_changes = sorted(inserted + modified + deleted)
row, col = view.rowcol(view.sel()[0].begin())
current_row = row + 1
line = self.jump(all_changes, current_row)
self.window.active_view().run_command("goto_line", {"line": line})
class GitGutterNextChangeCommand(GitGutterBaseChangeCommand):
def jump(self, all_changes, current_row):
return next((change for change in all_changes
if change > current_row), current_row)
class GitGutterPrevChangeCommand(GitGutterBaseChangeCommand):
def jump(self, all_changes, current_row):
return next((change for change in reversed(all_changes)
if change < current_row), current_row)Make jumping between changes loop back around
|
import sublime
import sublime_plugin
try:
from GitGutter.view_collection import ViewCollection
except ImportError:
from view_collection import ViewCollection
class GitGutterBaseChangeCommand(sublime_plugin.WindowCommand):
def lines_to_blocks(self, lines):
blocks = []
last_line = -2
for line in lines:
if line > last_line+1:
blocks.append(line)
last_line = line
return blocks
def run(self):
view = self.window.active_view()
inserted, modified, deleted = ViewCollection.diff(view)
inserted = self.lines_to_blocks(inserted)
modified = self.lines_to_blocks(modified)
all_changes = sorted(inserted + modified + deleted)
row, col = view.rowcol(view.sel()[0].begin())
current_row = row + 1
line = self.jump(all_changes, current_row)
self.window.active_view().run_command("goto_line", {"line": line})
class GitGutterNextChangeCommand(GitGutterBaseChangeCommand):
def jump(self, all_changes, current_row):
return next((change for change in all_changes
if change > current_row), all_changes[0])
class GitGutterPrevChangeCommand(GitGutterBaseChangeCommand):
def jump(self, all_changes, current_row):
return next((change for change in reversed(all_changes)
if change < current_row), all_changes[-1])
|
<commit_before>import sublime
import sublime_plugin
try:
from GitGutter.view_collection import ViewCollection
except ImportError:
from view_collection import ViewCollection
class GitGutterBaseChangeCommand(sublime_plugin.WindowCommand):
def lines_to_blocks(self, lines):
blocks = []
last_line = -2
for line in lines:
if line > last_line+1:
blocks.append(line)
last_line = line
return blocks
def run(self):
view = self.window.active_view()
inserted, modified, deleted = ViewCollection.diff(view)
inserted = self.lines_to_blocks(inserted)
modified = self.lines_to_blocks(modified)
all_changes = sorted(inserted + modified + deleted)
row, col = view.rowcol(view.sel()[0].begin())
current_row = row + 1
line = self.jump(all_changes, current_row)
self.window.active_view().run_command("goto_line", {"line": line})
class GitGutterNextChangeCommand(GitGutterBaseChangeCommand):
def jump(self, all_changes, current_row):
return next((change for change in all_changes
if change > current_row), current_row)
class GitGutterPrevChangeCommand(GitGutterBaseChangeCommand):
def jump(self, all_changes, current_row):
return next((change for change in reversed(all_changes)
if change < current_row), current_row)<commit_msg>Make jumping between changes loop back around<commit_after>
|
import sublime
import sublime_plugin
try:
from GitGutter.view_collection import ViewCollection
except ImportError:
from view_collection import ViewCollection
class GitGutterBaseChangeCommand(sublime_plugin.WindowCommand):
def lines_to_blocks(self, lines):
blocks = []
last_line = -2
for line in lines:
if line > last_line+1:
blocks.append(line)
last_line = line
return blocks
def run(self):
view = self.window.active_view()
inserted, modified, deleted = ViewCollection.diff(view)
inserted = self.lines_to_blocks(inserted)
modified = self.lines_to_blocks(modified)
all_changes = sorted(inserted + modified + deleted)
row, col = view.rowcol(view.sel()[0].begin())
current_row = row + 1
line = self.jump(all_changes, current_row)
self.window.active_view().run_command("goto_line", {"line": line})
class GitGutterNextChangeCommand(GitGutterBaseChangeCommand):
def jump(self, all_changes, current_row):
return next((change for change in all_changes
if change > current_row), all_changes[0])
class GitGutterPrevChangeCommand(GitGutterBaseChangeCommand):
def jump(self, all_changes, current_row):
return next((change for change in reversed(all_changes)
if change < current_row), all_changes[-1])
|
import sublime
import sublime_plugin
try:
from GitGutter.view_collection import ViewCollection
except ImportError:
from view_collection import ViewCollection
class GitGutterBaseChangeCommand(sublime_plugin.WindowCommand):
def lines_to_blocks(self, lines):
blocks = []
last_line = -2
for line in lines:
if line > last_line+1:
blocks.append(line)
last_line = line
return blocks
def run(self):
view = self.window.active_view()
inserted, modified, deleted = ViewCollection.diff(view)
inserted = self.lines_to_blocks(inserted)
modified = self.lines_to_blocks(modified)
all_changes = sorted(inserted + modified + deleted)
row, col = view.rowcol(view.sel()[0].begin())
current_row = row + 1
line = self.jump(all_changes, current_row)
self.window.active_view().run_command("goto_line", {"line": line})
class GitGutterNextChangeCommand(GitGutterBaseChangeCommand):
def jump(self, all_changes, current_row):
return next((change for change in all_changes
if change > current_row), current_row)
class GitGutterPrevChangeCommand(GitGutterBaseChangeCommand):
def jump(self, all_changes, current_row):
return next((change for change in reversed(all_changes)
if change < current_row), current_row)Make jumping between changes loop back aroundimport sublime
import sublime_plugin
try:
from GitGutter.view_collection import ViewCollection
except ImportError:
from view_collection import ViewCollection
class GitGutterBaseChangeCommand(sublime_plugin.WindowCommand):
def lines_to_blocks(self, lines):
blocks = []
last_line = -2
for line in lines:
if line > last_line+1:
blocks.append(line)
last_line = line
return blocks
def run(self):
view = self.window.active_view()
inserted, modified, deleted = ViewCollection.diff(view)
inserted = self.lines_to_blocks(inserted)
modified = self.lines_to_blocks(modified)
all_changes = sorted(inserted + modified + deleted)
row, col = view.rowcol(view.sel()[0].begin())
current_row = row + 1
line = self.jump(all_changes, current_row)
self.window.active_view().run_command("goto_line", {"line": line})
class GitGutterNextChangeCommand(GitGutterBaseChangeCommand):
def jump(self, all_changes, current_row):
return next((change for change in all_changes
if change > current_row), all_changes[0])
class GitGutterPrevChangeCommand(GitGutterBaseChangeCommand):
def jump(self, all_changes, current_row):
return next((change for change in reversed(all_changes)
if change < current_row), all_changes[-1])
|
<commit_before>import sublime
import sublime_plugin
try:
from GitGutter.view_collection import ViewCollection
except ImportError:
from view_collection import ViewCollection
class GitGutterBaseChangeCommand(sublime_plugin.WindowCommand):
def lines_to_blocks(self, lines):
blocks = []
last_line = -2
for line in lines:
if line > last_line+1:
blocks.append(line)
last_line = line
return blocks
def run(self):
view = self.window.active_view()
inserted, modified, deleted = ViewCollection.diff(view)
inserted = self.lines_to_blocks(inserted)
modified = self.lines_to_blocks(modified)
all_changes = sorted(inserted + modified + deleted)
row, col = view.rowcol(view.sel()[0].begin())
current_row = row + 1
line = self.jump(all_changes, current_row)
self.window.active_view().run_command("goto_line", {"line": line})
class GitGutterNextChangeCommand(GitGutterBaseChangeCommand):
def jump(self, all_changes, current_row):
return next((change for change in all_changes
if change > current_row), current_row)
class GitGutterPrevChangeCommand(GitGutterBaseChangeCommand):
def jump(self, all_changes, current_row):
return next((change for change in reversed(all_changes)
if change < current_row), current_row)<commit_msg>Make jumping between changes loop back around<commit_after>import sublime
import sublime_plugin
try:
from GitGutter.view_collection import ViewCollection
except ImportError:
from view_collection import ViewCollection
class GitGutterBaseChangeCommand(sublime_plugin.WindowCommand):
def lines_to_blocks(self, lines):
blocks = []
last_line = -2
for line in lines:
if line > last_line+1:
blocks.append(line)
last_line = line
return blocks
def run(self):
view = self.window.active_view()
inserted, modified, deleted = ViewCollection.diff(view)
inserted = self.lines_to_blocks(inserted)
modified = self.lines_to_blocks(modified)
all_changes = sorted(inserted + modified + deleted)
row, col = view.rowcol(view.sel()[0].begin())
current_row = row + 1
line = self.jump(all_changes, current_row)
self.window.active_view().run_command("goto_line", {"line": line})
class GitGutterNextChangeCommand(GitGutterBaseChangeCommand):
def jump(self, all_changes, current_row):
return next((change for change in all_changes
if change > current_row), all_changes[0])
class GitGutterPrevChangeCommand(GitGutterBaseChangeCommand):
def jump(self, all_changes, current_row):
return next((change for change in reversed(all_changes)
if change < current_row), all_changes[-1])
|
6a7fb1ff05202f60c7036db369926e3056372123
|
tests/chainer_tests/functions_tests/math_tests/test_sqrt.py
|
tests/chainer_tests/functions_tests/math_tests/test_sqrt.py
|
import unittest
import numpy
import chainer.functions as F
from chainer import testing
def make_data(dtype, shape):
x = numpy.random.uniform(0.1, 5, shape).astype(dtype)
gy = numpy.random.uniform(-1, 1, shape).astype(dtype)
return x, gy
#
# sqrt
@testing.math_function_test(F.Sqrt(), make_data=make_data)
class TestSqrt(unittest.TestCase):
pass
#
# rsqrt
def rsqrt(x, dtype=numpy.float32):
return numpy.reciprocal(numpy.sqrt(x, dtype=dtype))
# TODO(takagi) Fix test of rsqrt not to use this decorator.
@testing.math_function_test(F.rsqrt, func_expected=rsqrt, make_data=make_data)
class TestRsqrt(unittest.TestCase):
pass
testing.run_module(__name__, __file__)
|
import unittest
import numpy
import chainer.functions as F
from chainer import testing
#
# sqrt
def make_data(dtype, shape):
x = numpy.random.uniform(0.1, 5, shape).astype(dtype)
gy = numpy.random.uniform(-1, 1, shape).astype(dtype)
return x, gy
@testing.math_function_test(F.Sqrt(), make_data=make_data)
class TestSqrt(unittest.TestCase):
pass
#
# rsqrt
def rsqrt(x):
return numpy.reciprocal(numpy.sqrt(x))
class TestRsqrt(unittest.TestCase):
def test_rsqrt(self):
x = numpy.random.uniform(0.1, 5, (3, 2)).astype(numpy.float32)
testing.assert_allclose(F.rsqrt(x).data, rsqrt(x))
testing.run_module(__name__, __file__)
|
Simplify test of rsqrt function.
|
Simplify test of rsqrt function.
|
Python
|
mit
|
chainer/chainer,niboshi/chainer,hvy/chainer,hvy/chainer,anaruse/chainer,ktnyt/chainer,jnishi/chainer,pfnet/chainer,keisuke-umezawa/chainer,cupy/cupy,ysekky/chainer,kashif/chainer,keisuke-umezawa/chainer,ronekko/chainer,niboshi/chainer,wkentaro/chainer,cupy/cupy,kiyukuta/chainer,okuta/chainer,delta2323/chainer,hvy/chainer,ktnyt/chainer,keisuke-umezawa/chainer,niboshi/chainer,hvy/chainer,ktnyt/chainer,aonotas/chainer,rezoo/chainer,okuta/chainer,jnishi/chainer,cupy/cupy,ktnyt/chainer,keisuke-umezawa/chainer,wkentaro/chainer,okuta/chainer,wkentaro/chainer,niboshi/chainer,tkerola/chainer,chainer/chainer,okuta/chainer,jnishi/chainer,jnishi/chainer,wkentaro/chainer,chainer/chainer,cupy/cupy,chainer/chainer
|
import unittest
import numpy
import chainer.functions as F
from chainer import testing
def make_data(dtype, shape):
x = numpy.random.uniform(0.1, 5, shape).astype(dtype)
gy = numpy.random.uniform(-1, 1, shape).astype(dtype)
return x, gy
#
# sqrt
@testing.math_function_test(F.Sqrt(), make_data=make_data)
class TestSqrt(unittest.TestCase):
pass
#
# rsqrt
def rsqrt(x, dtype=numpy.float32):
return numpy.reciprocal(numpy.sqrt(x, dtype=dtype))
# TODO(takagi) Fix test of rsqrt not to use this decorator.
@testing.math_function_test(F.rsqrt, func_expected=rsqrt, make_data=make_data)
class TestRsqrt(unittest.TestCase):
pass
testing.run_module(__name__, __file__)
Simplify test of rsqrt function.
|
import unittest
import numpy
import chainer.functions as F
from chainer import testing
#
# sqrt
def make_data(dtype, shape):
x = numpy.random.uniform(0.1, 5, shape).astype(dtype)
gy = numpy.random.uniform(-1, 1, shape).astype(dtype)
return x, gy
@testing.math_function_test(F.Sqrt(), make_data=make_data)
class TestSqrt(unittest.TestCase):
pass
#
# rsqrt
def rsqrt(x):
return numpy.reciprocal(numpy.sqrt(x))
class TestRsqrt(unittest.TestCase):
def test_rsqrt(self):
x = numpy.random.uniform(0.1, 5, (3, 2)).astype(numpy.float32)
testing.assert_allclose(F.rsqrt(x).data, rsqrt(x))
testing.run_module(__name__, __file__)
|
<commit_before>import unittest
import numpy
import chainer.functions as F
from chainer import testing
def make_data(dtype, shape):
x = numpy.random.uniform(0.1, 5, shape).astype(dtype)
gy = numpy.random.uniform(-1, 1, shape).astype(dtype)
return x, gy
#
# sqrt
@testing.math_function_test(F.Sqrt(), make_data=make_data)
class TestSqrt(unittest.TestCase):
pass
#
# rsqrt
def rsqrt(x, dtype=numpy.float32):
return numpy.reciprocal(numpy.sqrt(x, dtype=dtype))
# TODO(takagi) Fix test of rsqrt not to use this decorator.
@testing.math_function_test(F.rsqrt, func_expected=rsqrt, make_data=make_data)
class TestRsqrt(unittest.TestCase):
pass
testing.run_module(__name__, __file__)
<commit_msg>Simplify test of rsqrt function.<commit_after>
|
import unittest
import numpy
import chainer.functions as F
from chainer import testing
#
# sqrt
def make_data(dtype, shape):
x = numpy.random.uniform(0.1, 5, shape).astype(dtype)
gy = numpy.random.uniform(-1, 1, shape).astype(dtype)
return x, gy
@testing.math_function_test(F.Sqrt(), make_data=make_data)
class TestSqrt(unittest.TestCase):
pass
#
# rsqrt
def rsqrt(x):
return numpy.reciprocal(numpy.sqrt(x))
class TestRsqrt(unittest.TestCase):
def test_rsqrt(self):
x = numpy.random.uniform(0.1, 5, (3, 2)).astype(numpy.float32)
testing.assert_allclose(F.rsqrt(x).data, rsqrt(x))
testing.run_module(__name__, __file__)
|
import unittest
import numpy
import chainer.functions as F
from chainer import testing
def make_data(dtype, shape):
x = numpy.random.uniform(0.1, 5, shape).astype(dtype)
gy = numpy.random.uniform(-1, 1, shape).astype(dtype)
return x, gy
#
# sqrt
@testing.math_function_test(F.Sqrt(), make_data=make_data)
class TestSqrt(unittest.TestCase):
pass
#
# rsqrt
def rsqrt(x, dtype=numpy.float32):
return numpy.reciprocal(numpy.sqrt(x, dtype=dtype))
# TODO(takagi) Fix test of rsqrt not to use this decorator.
@testing.math_function_test(F.rsqrt, func_expected=rsqrt, make_data=make_data)
class TestRsqrt(unittest.TestCase):
pass
testing.run_module(__name__, __file__)
Simplify test of rsqrt function.import unittest
import numpy
import chainer.functions as F
from chainer import testing
#
# sqrt
def make_data(dtype, shape):
x = numpy.random.uniform(0.1, 5, shape).astype(dtype)
gy = numpy.random.uniform(-1, 1, shape).astype(dtype)
return x, gy
@testing.math_function_test(F.Sqrt(), make_data=make_data)
class TestSqrt(unittest.TestCase):
pass
#
# rsqrt
def rsqrt(x):
return numpy.reciprocal(numpy.sqrt(x))
class TestRsqrt(unittest.TestCase):
def test_rsqrt(self):
x = numpy.random.uniform(0.1, 5, (3, 2)).astype(numpy.float32)
testing.assert_allclose(F.rsqrt(x).data, rsqrt(x))
testing.run_module(__name__, __file__)
|
<commit_before>import unittest
import numpy
import chainer.functions as F
from chainer import testing
def make_data(dtype, shape):
x = numpy.random.uniform(0.1, 5, shape).astype(dtype)
gy = numpy.random.uniform(-1, 1, shape).astype(dtype)
return x, gy
#
# sqrt
@testing.math_function_test(F.Sqrt(), make_data=make_data)
class TestSqrt(unittest.TestCase):
pass
#
# rsqrt
def rsqrt(x, dtype=numpy.float32):
return numpy.reciprocal(numpy.sqrt(x, dtype=dtype))
# TODO(takagi) Fix test of rsqrt not to use this decorator.
@testing.math_function_test(F.rsqrt, func_expected=rsqrt, make_data=make_data)
class TestRsqrt(unittest.TestCase):
pass
testing.run_module(__name__, __file__)
<commit_msg>Simplify test of rsqrt function.<commit_after>import unittest
import numpy
import chainer.functions as F
from chainer import testing
#
# sqrt
def make_data(dtype, shape):
x = numpy.random.uniform(0.1, 5, shape).astype(dtype)
gy = numpy.random.uniform(-1, 1, shape).astype(dtype)
return x, gy
@testing.math_function_test(F.Sqrt(), make_data=make_data)
class TestSqrt(unittest.TestCase):
pass
#
# rsqrt
def rsqrt(x):
return numpy.reciprocal(numpy.sqrt(x))
class TestRsqrt(unittest.TestCase):
def test_rsqrt(self):
x = numpy.random.uniform(0.1, 5, (3, 2)).astype(numpy.float32)
testing.assert_allclose(F.rsqrt(x).data, rsqrt(x))
testing.run_module(__name__, __file__)
|
63bf9c267ff891f1a2bd1f472a5d77f8df1e0209
|
tests/iam/test_iam_valid_json.py
|
tests/iam/test_iam_valid_json.py
|
"""Test IAM Policy templates are valid JSON."""
import jinja2
from foremast.iam.construct_policy import render_policy_template
from foremast.utils.templates import LOCAL_TEMPLATES
def iam_templates():
"""Generate list of IAM templates."""
jinjaenv = jinja2.Environment(loader=jinja2.FileSystemLoader([LOCAL_TEMPLATES]))
iam_template_names = jinjaenv.list_templates(filter_func=lambda x: all([
x.startswith('infrastructure/iam/'),
'trust' not in x,
'wrapper' not in x, ]))
for iam_template_name in iam_template_names:
yield iam_template_name
items = ['resource1', 'resource2']
if service == 'rds-db':
items = {
'resource1': 'user1',
'resource2': 'user2',
}
rendered = render_policy_template(
account_number='',
app='coreforrest',
env='dev',
group='forrest',
items=items,
pipeline_settings={
'lambda': {
'vpc_enabled': False,
},
},
region='us-east-1',
service=service)
assert isinstance(rendered, list)
|
"""Test IAM Policy templates are valid JSON."""
import json
import jinja2
import pytest
from foremast.iam.construct_policy import render_policy_template
from foremast.utils.templates import LOCAL_TEMPLATES
def iam_templates():
"""Generate list of IAM templates."""
jinjaenv = jinja2.Environment(loader=jinja2.FileSystemLoader([LOCAL_TEMPLATES]))
iam_template_names = jinjaenv.list_templates(filter_func=lambda x: all([
x.startswith('infrastructure/iam/'),
'trust' not in x,
'wrapper' not in x, ]))
for iam_template_name in iam_template_names:
yield iam_template_name
@pytest.mark.parametrize(argnames='template_name', argvalues=iam_templates())
def test_all_iam_templates(template_name):
"""Verify all IAM templates render as proper JSON."""
*_, service_json = template_name.split('/')
service, *_ = service_json.split('.')
items = ['resource1', 'resource2']
if service == 'rds-db':
items = {
'resource1': 'user1',
'resource2': 'user2',
}
try:
rendered = render_policy_template(
account_number='',
app='coreforrest',
env='dev',
group='forrest',
items=items,
pipeline_settings={
'lambda': {
'vpc_enabled': False,
},
},
region='us-east-1',
service=service)
except json.decoder.JSONDecodeError:
pytest.fail('Bad template: {0}'.format(template_name), pytrace=False)
assert isinstance(rendered, list)
|
Split IAM template tests with paramtrize
|
test: Split IAM template tests with paramtrize
See also: #208
|
Python
|
apache-2.0
|
gogoair/foremast,gogoair/foremast
|
"""Test IAM Policy templates are valid JSON."""
import jinja2
from foremast.iam.construct_policy import render_policy_template
from foremast.utils.templates import LOCAL_TEMPLATES
def iam_templates():
"""Generate list of IAM templates."""
jinjaenv = jinja2.Environment(loader=jinja2.FileSystemLoader([LOCAL_TEMPLATES]))
iam_template_names = jinjaenv.list_templates(filter_func=lambda x: all([
x.startswith('infrastructure/iam/'),
'trust' not in x,
'wrapper' not in x, ]))
for iam_template_name in iam_template_names:
yield iam_template_name
items = ['resource1', 'resource2']
if service == 'rds-db':
items = {
'resource1': 'user1',
'resource2': 'user2',
}
rendered = render_policy_template(
account_number='',
app='coreforrest',
env='dev',
group='forrest',
items=items,
pipeline_settings={
'lambda': {
'vpc_enabled': False,
},
},
region='us-east-1',
service=service)
assert isinstance(rendered, list)
test: Split IAM template tests with paramtrize
See also: #208
|
"""Test IAM Policy templates are valid JSON."""
import json
import jinja2
import pytest
from foremast.iam.construct_policy import render_policy_template
from foremast.utils.templates import LOCAL_TEMPLATES
def iam_templates():
"""Generate list of IAM templates."""
jinjaenv = jinja2.Environment(loader=jinja2.FileSystemLoader([LOCAL_TEMPLATES]))
iam_template_names = jinjaenv.list_templates(filter_func=lambda x: all([
x.startswith('infrastructure/iam/'),
'trust' not in x,
'wrapper' not in x, ]))
for iam_template_name in iam_template_names:
yield iam_template_name
@pytest.mark.parametrize(argnames='template_name', argvalues=iam_templates())
def test_all_iam_templates(template_name):
"""Verify all IAM templates render as proper JSON."""
*_, service_json = template_name.split('/')
service, *_ = service_json.split('.')
items = ['resource1', 'resource2']
if service == 'rds-db':
items = {
'resource1': 'user1',
'resource2': 'user2',
}
try:
rendered = render_policy_template(
account_number='',
app='coreforrest',
env='dev',
group='forrest',
items=items,
pipeline_settings={
'lambda': {
'vpc_enabled': False,
},
},
region='us-east-1',
service=service)
except json.decoder.JSONDecodeError:
pytest.fail('Bad template: {0}'.format(template_name), pytrace=False)
assert isinstance(rendered, list)
|
<commit_before>"""Test IAM Policy templates are valid JSON."""
import jinja2
from foremast.iam.construct_policy import render_policy_template
from foremast.utils.templates import LOCAL_TEMPLATES
def iam_templates():
"""Generate list of IAM templates."""
jinjaenv = jinja2.Environment(loader=jinja2.FileSystemLoader([LOCAL_TEMPLATES]))
iam_template_names = jinjaenv.list_templates(filter_func=lambda x: all([
x.startswith('infrastructure/iam/'),
'trust' not in x,
'wrapper' not in x, ]))
for iam_template_name in iam_template_names:
yield iam_template_name
items = ['resource1', 'resource2']
if service == 'rds-db':
items = {
'resource1': 'user1',
'resource2': 'user2',
}
rendered = render_policy_template(
account_number='',
app='coreforrest',
env='dev',
group='forrest',
items=items,
pipeline_settings={
'lambda': {
'vpc_enabled': False,
},
},
region='us-east-1',
service=service)
assert isinstance(rendered, list)
<commit_msg>test: Split IAM template tests with paramtrize
See also: #208<commit_after>
|
"""Test IAM Policy templates are valid JSON."""
import json
import jinja2
import pytest
from foremast.iam.construct_policy import render_policy_template
from foremast.utils.templates import LOCAL_TEMPLATES
def iam_templates():
"""Generate list of IAM templates."""
jinjaenv = jinja2.Environment(loader=jinja2.FileSystemLoader([LOCAL_TEMPLATES]))
iam_template_names = jinjaenv.list_templates(filter_func=lambda x: all([
x.startswith('infrastructure/iam/'),
'trust' not in x,
'wrapper' not in x, ]))
for iam_template_name in iam_template_names:
yield iam_template_name
@pytest.mark.parametrize(argnames='template_name', argvalues=iam_templates())
def test_all_iam_templates(template_name):
"""Verify all IAM templates render as proper JSON."""
*_, service_json = template_name.split('/')
service, *_ = service_json.split('.')
items = ['resource1', 'resource2']
if service == 'rds-db':
items = {
'resource1': 'user1',
'resource2': 'user2',
}
try:
rendered = render_policy_template(
account_number='',
app='coreforrest',
env='dev',
group='forrest',
items=items,
pipeline_settings={
'lambda': {
'vpc_enabled': False,
},
},
region='us-east-1',
service=service)
except json.decoder.JSONDecodeError:
pytest.fail('Bad template: {0}'.format(template_name), pytrace=False)
assert isinstance(rendered, list)
|
"""Test IAM Policy templates are valid JSON."""
import jinja2
from foremast.iam.construct_policy import render_policy_template
from foremast.utils.templates import LOCAL_TEMPLATES
def iam_templates():
"""Generate list of IAM templates."""
jinjaenv = jinja2.Environment(loader=jinja2.FileSystemLoader([LOCAL_TEMPLATES]))
iam_template_names = jinjaenv.list_templates(filter_func=lambda x: all([
x.startswith('infrastructure/iam/'),
'trust' not in x,
'wrapper' not in x, ]))
for iam_template_name in iam_template_names:
yield iam_template_name
items = ['resource1', 'resource2']
if service == 'rds-db':
items = {
'resource1': 'user1',
'resource2': 'user2',
}
rendered = render_policy_template(
account_number='',
app='coreforrest',
env='dev',
group='forrest',
items=items,
pipeline_settings={
'lambda': {
'vpc_enabled': False,
},
},
region='us-east-1',
service=service)
assert isinstance(rendered, list)
test: Split IAM template tests with paramtrize
See also: #208"""Test IAM Policy templates are valid JSON."""
import json
import jinja2
import pytest
from foremast.iam.construct_policy import render_policy_template
from foremast.utils.templates import LOCAL_TEMPLATES
def iam_templates():
"""Generate list of IAM templates."""
jinjaenv = jinja2.Environment(loader=jinja2.FileSystemLoader([LOCAL_TEMPLATES]))
iam_template_names = jinjaenv.list_templates(filter_func=lambda x: all([
x.startswith('infrastructure/iam/'),
'trust' not in x,
'wrapper' not in x, ]))
for iam_template_name in iam_template_names:
yield iam_template_name
@pytest.mark.parametrize(argnames='template_name', argvalues=iam_templates())
def test_all_iam_templates(template_name):
"""Verify all IAM templates render as proper JSON."""
*_, service_json = template_name.split('/')
service, *_ = service_json.split('.')
items = ['resource1', 'resource2']
if service == 'rds-db':
items = {
'resource1': 'user1',
'resource2': 'user2',
}
try:
rendered = render_policy_template(
account_number='',
app='coreforrest',
env='dev',
group='forrest',
items=items,
pipeline_settings={
'lambda': {
'vpc_enabled': False,
},
},
region='us-east-1',
service=service)
except json.decoder.JSONDecodeError:
pytest.fail('Bad template: {0}'.format(template_name), pytrace=False)
assert isinstance(rendered, list)
|
<commit_before>"""Test IAM Policy templates are valid JSON."""
import jinja2
from foremast.iam.construct_policy import render_policy_template
from foremast.utils.templates import LOCAL_TEMPLATES
def iam_templates():
"""Generate list of IAM templates."""
jinjaenv = jinja2.Environment(loader=jinja2.FileSystemLoader([LOCAL_TEMPLATES]))
iam_template_names = jinjaenv.list_templates(filter_func=lambda x: all([
x.startswith('infrastructure/iam/'),
'trust' not in x,
'wrapper' not in x, ]))
for iam_template_name in iam_template_names:
yield iam_template_name
items = ['resource1', 'resource2']
if service == 'rds-db':
items = {
'resource1': 'user1',
'resource2': 'user2',
}
rendered = render_policy_template(
account_number='',
app='coreforrest',
env='dev',
group='forrest',
items=items,
pipeline_settings={
'lambda': {
'vpc_enabled': False,
},
},
region='us-east-1',
service=service)
assert isinstance(rendered, list)
<commit_msg>test: Split IAM template tests with paramtrize
See also: #208<commit_after>"""Test IAM Policy templates are valid JSON."""
import json
import jinja2
import pytest
from foremast.iam.construct_policy import render_policy_template
from foremast.utils.templates import LOCAL_TEMPLATES
def iam_templates():
"""Generate list of IAM templates."""
jinjaenv = jinja2.Environment(loader=jinja2.FileSystemLoader([LOCAL_TEMPLATES]))
iam_template_names = jinjaenv.list_templates(filter_func=lambda x: all([
x.startswith('infrastructure/iam/'),
'trust' not in x,
'wrapper' not in x, ]))
for iam_template_name in iam_template_names:
yield iam_template_name
@pytest.mark.parametrize(argnames='template_name', argvalues=iam_templates())
def test_all_iam_templates(template_name):
"""Verify all IAM templates render as proper JSON."""
*_, service_json = template_name.split('/')
service, *_ = service_json.split('.')
items = ['resource1', 'resource2']
if service == 'rds-db':
items = {
'resource1': 'user1',
'resource2': 'user2',
}
try:
rendered = render_policy_template(
account_number='',
app='coreforrest',
env='dev',
group='forrest',
items=items,
pipeline_settings={
'lambda': {
'vpc_enabled': False,
},
},
region='us-east-1',
service=service)
except json.decoder.JSONDecodeError:
pytest.fail('Bad template: {0}'.format(template_name), pytrace=False)
assert isinstance(rendered, list)
|
79158c269669fcbe506ae83e803ef58ba1b40913
|
examples/olfaction/data/gen_olf_input.py
|
examples/olfaction/data/gen_olf_input.py
|
#!/usr/bin/env python
"""
Generate sample olfactory model stimulus.
"""
import numpy as np
import h5py
osn_num = 1375
dt = 1e-4 # time step
Ot = 2000 # number of data point during reset period
Rt = 1000 # number of data point during odor delivery period
Nt = 4*Ot + 3*Rt # number of data points in time
t = np.arange(0, dt*Nt, dt)
I = -1.*0.0195 # amplitude of odorant concentration
u_on = I*np.ones(Ot, dtype=np.float64)
u_off = np.zeros(Ot, dtype=np.float64)
u_reset = np.zeros(Rt, dtype=np.float64)
u = np.concatenate((u_off, u_reset, u_on, u_reset, u_off, u_reset, u_on))
u_all = np.transpose(np.kron(np.ones((osn_num, 1)), u))
with h5py.File('olfactory_input.h5', 'w') as f:
f.create_dataset('real', (Nt, osn_num),
dtype=np.float64,
data=u_all)
|
#!/usr/bin/env python
"""
Generate sample olfactory model stimulus.
"""
import numpy as np
import h5py
osn_num = 1375
dt = 1e-4 # time step
Ot = 2000 # number of data point during reset period
Rt = 1000 # number of data point during odor delivery period
Nt = 4*Ot + 3*Rt # number of data points in time
t = np.arange(0, dt*Nt, dt)
I = 10.0 # amplitude of odorant concentration
u_on = I*np.ones(Ot, dtype=np.float64)
u_off = np.zeros(Ot, dtype=np.float64)
u_reset = np.zeros(Rt, dtype=np.float64)
u = np.concatenate((u_off, u_reset, u_on, u_reset, u_off, u_reset, u_on))
u_all = np.transpose(np.kron(np.ones((osn_num, 1)), u))
with h5py.File('olfactory_input.h5', 'w') as f:
f.create_dataset('real', (Nt, osn_num),
dtype=np.float64,
data=u_all)
|
Tweak olfactory input stimulus to produce more interesting output.
|
Tweak olfactory input stimulus to produce more interesting output.
|
Python
|
bsd-3-clause
|
cerrno/neurokernel
|
#!/usr/bin/env python
"""
Generate sample olfactory model stimulus.
"""
import numpy as np
import h5py
osn_num = 1375
dt = 1e-4 # time step
Ot = 2000 # number of data point during reset period
Rt = 1000 # number of data point during odor delivery period
Nt = 4*Ot + 3*Rt # number of data points in time
t = np.arange(0, dt*Nt, dt)
I = -1.*0.0195 # amplitude of odorant concentration
u_on = I*np.ones(Ot, dtype=np.float64)
u_off = np.zeros(Ot, dtype=np.float64)
u_reset = np.zeros(Rt, dtype=np.float64)
u = np.concatenate((u_off, u_reset, u_on, u_reset, u_off, u_reset, u_on))
u_all = np.transpose(np.kron(np.ones((osn_num, 1)), u))
with h5py.File('olfactory_input.h5', 'w') as f:
f.create_dataset('real', (Nt, osn_num),
dtype=np.float64,
data=u_all)
Tweak olfactory input stimulus to produce more interesting output.
|
#!/usr/bin/env python
"""
Generate sample olfactory model stimulus.
"""
import numpy as np
import h5py
osn_num = 1375
dt = 1e-4 # time step
Ot = 2000 # number of data point during reset period
Rt = 1000 # number of data point during odor delivery period
Nt = 4*Ot + 3*Rt # number of data points in time
t = np.arange(0, dt*Nt, dt)
I = 10.0 # amplitude of odorant concentration
u_on = I*np.ones(Ot, dtype=np.float64)
u_off = np.zeros(Ot, dtype=np.float64)
u_reset = np.zeros(Rt, dtype=np.float64)
u = np.concatenate((u_off, u_reset, u_on, u_reset, u_off, u_reset, u_on))
u_all = np.transpose(np.kron(np.ones((osn_num, 1)), u))
with h5py.File('olfactory_input.h5', 'w') as f:
f.create_dataset('real', (Nt, osn_num),
dtype=np.float64,
data=u_all)
|
<commit_before>#!/usr/bin/env python
"""
Generate sample olfactory model stimulus.
"""
import numpy as np
import h5py
osn_num = 1375
dt = 1e-4 # time step
Ot = 2000 # number of data point during reset period
Rt = 1000 # number of data point during odor delivery period
Nt = 4*Ot + 3*Rt # number of data points in time
t = np.arange(0, dt*Nt, dt)
I = -1.*0.0195 # amplitude of odorant concentration
u_on = I*np.ones(Ot, dtype=np.float64)
u_off = np.zeros(Ot, dtype=np.float64)
u_reset = np.zeros(Rt, dtype=np.float64)
u = np.concatenate((u_off, u_reset, u_on, u_reset, u_off, u_reset, u_on))
u_all = np.transpose(np.kron(np.ones((osn_num, 1)), u))
with h5py.File('olfactory_input.h5', 'w') as f:
f.create_dataset('real', (Nt, osn_num),
dtype=np.float64,
data=u_all)
<commit_msg>Tweak olfactory input stimulus to produce more interesting output.<commit_after>
|
#!/usr/bin/env python
"""
Generate sample olfactory model stimulus.
"""
import numpy as np
import h5py
osn_num = 1375
dt = 1e-4 # time step
Ot = 2000 # number of data point during reset period
Rt = 1000 # number of data point during odor delivery period
Nt = 4*Ot + 3*Rt # number of data points in time
t = np.arange(0, dt*Nt, dt)
I = 10.0 # amplitude of odorant concentration
u_on = I*np.ones(Ot, dtype=np.float64)
u_off = np.zeros(Ot, dtype=np.float64)
u_reset = np.zeros(Rt, dtype=np.float64)
u = np.concatenate((u_off, u_reset, u_on, u_reset, u_off, u_reset, u_on))
u_all = np.transpose(np.kron(np.ones((osn_num, 1)), u))
with h5py.File('olfactory_input.h5', 'w') as f:
f.create_dataset('real', (Nt, osn_num),
dtype=np.float64,
data=u_all)
|
#!/usr/bin/env python
"""
Generate sample olfactory model stimulus.
"""
import numpy as np
import h5py
osn_num = 1375
dt = 1e-4 # time step
Ot = 2000 # number of data point during reset period
Rt = 1000 # number of data point during odor delivery period
Nt = 4*Ot + 3*Rt # number of data points in time
t = np.arange(0, dt*Nt, dt)
I = -1.*0.0195 # amplitude of odorant concentration
u_on = I*np.ones(Ot, dtype=np.float64)
u_off = np.zeros(Ot, dtype=np.float64)
u_reset = np.zeros(Rt, dtype=np.float64)
u = np.concatenate((u_off, u_reset, u_on, u_reset, u_off, u_reset, u_on))
u_all = np.transpose(np.kron(np.ones((osn_num, 1)), u))
with h5py.File('olfactory_input.h5', 'w') as f:
f.create_dataset('real', (Nt, osn_num),
dtype=np.float64,
data=u_all)
Tweak olfactory input stimulus to produce more interesting output.#!/usr/bin/env python
"""
Generate sample olfactory model stimulus.
"""
import numpy as np
import h5py
osn_num = 1375
dt = 1e-4 # time step
Ot = 2000 # number of data point during reset period
Rt = 1000 # number of data point during odor delivery period
Nt = 4*Ot + 3*Rt # number of data points in time
t = np.arange(0, dt*Nt, dt)
I = 10.0 # amplitude of odorant concentration
u_on = I*np.ones(Ot, dtype=np.float64)
u_off = np.zeros(Ot, dtype=np.float64)
u_reset = np.zeros(Rt, dtype=np.float64)
u = np.concatenate((u_off, u_reset, u_on, u_reset, u_off, u_reset, u_on))
u_all = np.transpose(np.kron(np.ones((osn_num, 1)), u))
with h5py.File('olfactory_input.h5', 'w') as f:
f.create_dataset('real', (Nt, osn_num),
dtype=np.float64,
data=u_all)
|
<commit_before>#!/usr/bin/env python
"""
Generate sample olfactory model stimulus.
"""
import numpy as np
import h5py
osn_num = 1375
dt = 1e-4 # time step
Ot = 2000 # number of data point during reset period
Rt = 1000 # number of data point during odor delivery period
Nt = 4*Ot + 3*Rt # number of data points in time
t = np.arange(0, dt*Nt, dt)
I = -1.*0.0195 # amplitude of odorant concentration
u_on = I*np.ones(Ot, dtype=np.float64)
u_off = np.zeros(Ot, dtype=np.float64)
u_reset = np.zeros(Rt, dtype=np.float64)
u = np.concatenate((u_off, u_reset, u_on, u_reset, u_off, u_reset, u_on))
u_all = np.transpose(np.kron(np.ones((osn_num, 1)), u))
with h5py.File('olfactory_input.h5', 'w') as f:
f.create_dataset('real', (Nt, osn_num),
dtype=np.float64,
data=u_all)
<commit_msg>Tweak olfactory input stimulus to produce more interesting output.<commit_after>#!/usr/bin/env python
"""
Generate sample olfactory model stimulus.
"""
import numpy as np
import h5py
osn_num = 1375
dt = 1e-4 # time step
Ot = 2000 # number of data point during reset period
Rt = 1000 # number of data point during odor delivery period
Nt = 4*Ot + 3*Rt # number of data points in time
t = np.arange(0, dt*Nt, dt)
I = 10.0 # amplitude of odorant concentration
u_on = I*np.ones(Ot, dtype=np.float64)
u_off = np.zeros(Ot, dtype=np.float64)
u_reset = np.zeros(Rt, dtype=np.float64)
u = np.concatenate((u_off, u_reset, u_on, u_reset, u_off, u_reset, u_on))
u_all = np.transpose(np.kron(np.ones((osn_num, 1)), u))
with h5py.File('olfactory_input.h5', 'w') as f:
f.create_dataset('real', (Nt, osn_num),
dtype=np.float64,
data=u_all)
|
0ff797d60c2ddc93579e7c486e8ebb77593014d8
|
apiclient/__init__.py
|
apiclient/__init__.py
|
"""Retain apiclient as an alias for googleapiclient."""
import googleapiclient
from googleapiclient import channel
from googleapiclient import discovery
from googleapiclient import errors
from googleapiclient import http
from googleapiclient import mimeparse
from googleapiclient import model
from googleapiclient import sample_tools
from googleapiclient import schema
__version__ = googleapiclient.__version__
|
"""Retain apiclient as an alias for googleapiclient."""
import googleapiclient
try:
import oauth2client
except ImportError:
raise RuntimeError(
'Previous version of google-api-python-client detected; due to a '
'packaging issue, we cannot perform an in-place upgrade. To repair, '
'remove and reinstall this package, along with oauth2client and '
'uritemplate. One can do this with pip via\n'
' pip install -I google-api-python-client'
)
from googleapiclient import channel
from googleapiclient import discovery
from googleapiclient import errors
from googleapiclient import http
from googleapiclient import mimeparse
from googleapiclient import model
from googleapiclient import sample_tools
from googleapiclient import schema
__version__ = googleapiclient.__version__
|
Add another check for a failed googleapiclient upgrade.
|
Add another check for a failed googleapiclient upgrade.
This adds one more check for a failed 1.2 -> 1.3 upgrade, specifically in the
`apiclient` import (which was the only import available in 1.2). Even combined
with the check in setup.py, this won't catch everything, but it now covers all
the most common cases.
|
Python
|
apache-2.0
|
googleapis/google-api-python-client,googleapis/google-api-python-client
|
"""Retain apiclient as an alias for googleapiclient."""
import googleapiclient
from googleapiclient import channel
from googleapiclient import discovery
from googleapiclient import errors
from googleapiclient import http
from googleapiclient import mimeparse
from googleapiclient import model
from googleapiclient import sample_tools
from googleapiclient import schema
__version__ = googleapiclient.__version__
Add another check for a failed googleapiclient upgrade.
This adds one more check for a failed 1.2 -> 1.3 upgrade, specifically in the
`apiclient` import (which was the only import available in 1.2). Even combined
with the check in setup.py, this won't catch everything, but it now covers all
the most common cases.
|
"""Retain apiclient as an alias for googleapiclient."""
import googleapiclient
try:
import oauth2client
except ImportError:
raise RuntimeError(
'Previous version of google-api-python-client detected; due to a '
'packaging issue, we cannot perform an in-place upgrade. To repair, '
'remove and reinstall this package, along with oauth2client and '
'uritemplate. One can do this with pip via\n'
' pip install -I google-api-python-client'
)
from googleapiclient import channel
from googleapiclient import discovery
from googleapiclient import errors
from googleapiclient import http
from googleapiclient import mimeparse
from googleapiclient import model
from googleapiclient import sample_tools
from googleapiclient import schema
__version__ = googleapiclient.__version__
|
<commit_before>"""Retain apiclient as an alias for googleapiclient."""
import googleapiclient
from googleapiclient import channel
from googleapiclient import discovery
from googleapiclient import errors
from googleapiclient import http
from googleapiclient import mimeparse
from googleapiclient import model
from googleapiclient import sample_tools
from googleapiclient import schema
__version__ = googleapiclient.__version__
<commit_msg>Add another check for a failed googleapiclient upgrade.
This adds one more check for a failed 1.2 -> 1.3 upgrade, specifically in the
`apiclient` import (which was the only import available in 1.2). Even combined
with the check in setup.py, this won't catch everything, but it now covers all
the most common cases.<commit_after>
|
"""Retain apiclient as an alias for googleapiclient."""
import googleapiclient
try:
import oauth2client
except ImportError:
raise RuntimeError(
'Previous version of google-api-python-client detected; due to a '
'packaging issue, we cannot perform an in-place upgrade. To repair, '
'remove and reinstall this package, along with oauth2client and '
'uritemplate. One can do this with pip via\n'
' pip install -I google-api-python-client'
)
from googleapiclient import channel
from googleapiclient import discovery
from googleapiclient import errors
from googleapiclient import http
from googleapiclient import mimeparse
from googleapiclient import model
from googleapiclient import sample_tools
from googleapiclient import schema
__version__ = googleapiclient.__version__
|
"""Retain apiclient as an alias for googleapiclient."""
import googleapiclient
from googleapiclient import channel
from googleapiclient import discovery
from googleapiclient import errors
from googleapiclient import http
from googleapiclient import mimeparse
from googleapiclient import model
from googleapiclient import sample_tools
from googleapiclient import schema
__version__ = googleapiclient.__version__
Add another check for a failed googleapiclient upgrade.
This adds one more check for a failed 1.2 -> 1.3 upgrade, specifically in the
`apiclient` import (which was the only import available in 1.2). Even combined
with the check in setup.py, this won't catch everything, but it now covers all
the most common cases."""Retain apiclient as an alias for googleapiclient."""
import googleapiclient
try:
import oauth2client
except ImportError:
raise RuntimeError(
'Previous version of google-api-python-client detected; due to a '
'packaging issue, we cannot perform an in-place upgrade. To repair, '
'remove and reinstall this package, along with oauth2client and '
'uritemplate. One can do this with pip via\n'
' pip install -I google-api-python-client'
)
from googleapiclient import channel
from googleapiclient import discovery
from googleapiclient import errors
from googleapiclient import http
from googleapiclient import mimeparse
from googleapiclient import model
from googleapiclient import sample_tools
from googleapiclient import schema
__version__ = googleapiclient.__version__
|
<commit_before>"""Retain apiclient as an alias for googleapiclient."""
import googleapiclient
from googleapiclient import channel
from googleapiclient import discovery
from googleapiclient import errors
from googleapiclient import http
from googleapiclient import mimeparse
from googleapiclient import model
from googleapiclient import sample_tools
from googleapiclient import schema
__version__ = googleapiclient.__version__
<commit_msg>Add another check for a failed googleapiclient upgrade.
This adds one more check for a failed 1.2 -> 1.3 upgrade, specifically in the
`apiclient` import (which was the only import available in 1.2). Even combined
with the check in setup.py, this won't catch everything, but it now covers all
the most common cases.<commit_after>"""Retain apiclient as an alias for googleapiclient."""
import googleapiclient
try:
import oauth2client
except ImportError:
raise RuntimeError(
'Previous version of google-api-python-client detected; due to a '
'packaging issue, we cannot perform an in-place upgrade. To repair, '
'remove and reinstall this package, along with oauth2client and '
'uritemplate. One can do this with pip via\n'
' pip install -I google-api-python-client'
)
from googleapiclient import channel
from googleapiclient import discovery
from googleapiclient import errors
from googleapiclient import http
from googleapiclient import mimeparse
from googleapiclient import model
from googleapiclient import sample_tools
from googleapiclient import schema
__version__ = googleapiclient.__version__
|
2995f15c1bcb1bc85d83c7407be199b27882a215
|
examples/translations/japanese_test_1.py
|
examples/translations/japanese_test_1.py
|
# Japanese Language Test - Python 3 Only!
from seleniumbase.translate.japanese import γ»γ¬γ³γγΉγγ±γΌγΉ # noqa
class γγΉγγ―γ©γΉ(γ»γ¬γ³γγΉγγ±γΌγΉ): # noqa
def test_δΎ1(self):
self.URLγιγ("https://ja.wikipedia.org/wiki/")
self.γγγΉγγη’Ίθͺγγ("γ¦γ£γγγγ£γ’")
self.θ¦η΄ γη’Ίθͺγγ('[title="γ‘γ€γ³γγΌγΈγ«η§»εγγ"]')
self.γγγΉγγζ΄ζ°("#searchInput", "γ’γγ‘")
self.γ―γͺγγ―γγ¦("#searchButton")
self.γγγΉγγη’Ίθͺγγ("γ’γγ‘", "#firstHeading")
self.γγγΉγγζ΄ζ°("#searchInput", "ε―ΏεΈ")
self.γ―γͺγγ―γγ¦("#searchButton")
self.γγγΉγγη’Ίθͺγγ("ε―ΏεΈ", "#firstHeading")
self.θ¦η΄ γη’Ίθͺγγ('img[alt="ζ‘γε―ΏεΈ"]')
|
# Japanese Language Test - Python 3 Only!
from seleniumbase.translate.japanese import γ»γ¬γγ¦γ γγΉγγ±γΌγΉ # noqa
class γγΉγγ―γ©γΉ(γ»γ¬γγ¦γ γγΉγγ±γΌγΉ): # noqa
def test_δΎ1(self):
self.URLγιγ("https://ja.wikipedia.org/wiki/")
self.γγγΉγγη’Ίθͺγγ("γ¦γ£γγγγ£γ’")
self.θ¦η΄ γη’Ίθͺγγ('[title="γ‘γ€γ³γγΌγΈγ«η§»εγγ"]')
self.γγγΉγγζ΄ζ°("#searchInput", "γ’γγ‘")
self.γ―γͺγγ―γγ¦("#searchButton")
self.γγγΉγγη’Ίθͺγγ("γ’γγ‘", "#firstHeading")
self.γγγΉγγζ΄ζ°("#searchInput", "ε―ΏεΈ")
self.γ―γͺγγ―γγ¦("#searchButton")
self.γγγΉγγη’Ίθͺγγ("ε―ΏεΈ", "#firstHeading")
self.θ¦η΄ γη’Ίθͺγγ('img[alt="ζ‘γε―ΏεΈ"]')
|
Update for fixing odd Japanese
|
Update for fixing odd Japanese
Selenium is "γ»γ¬γγ¦γ " in Japanese. (Most people don't write Selenium in Japanese, by the way)
|
Python
|
mit
|
mdmintz/SeleniumBase,seleniumbase/SeleniumBase,seleniumbase/SeleniumBase,mdmintz/SeleniumBase,mdmintz/SeleniumBase,seleniumbase/SeleniumBase,seleniumbase/SeleniumBase,mdmintz/SeleniumBase
|
# Japanese Language Test - Python 3 Only!
from seleniumbase.translate.japanese import γ»γ¬γ³γγΉγγ±γΌγΉ # noqa
class γγΉγγ―γ©γΉ(γ»γ¬γ³γγΉγγ±γΌγΉ): # noqa
def test_δΎ1(self):
self.URLγιγ("https://ja.wikipedia.org/wiki/")
self.γγγΉγγη’Ίθͺγγ("γ¦γ£γγγγ£γ’")
self.θ¦η΄ γη’Ίθͺγγ('[title="γ‘γ€γ³γγΌγΈγ«η§»εγγ"]')
self.γγγΉγγζ΄ζ°("#searchInput", "γ’γγ‘")
self.γ―γͺγγ―γγ¦("#searchButton")
self.γγγΉγγη’Ίθͺγγ("γ’γγ‘", "#firstHeading")
self.γγγΉγγζ΄ζ°("#searchInput", "ε―ΏεΈ")
self.γ―γͺγγ―γγ¦("#searchButton")
self.γγγΉγγη’Ίθͺγγ("ε―ΏεΈ", "#firstHeading")
self.θ¦η΄ γη’Ίθͺγγ('img[alt="ζ‘γε―ΏεΈ"]')
Update for fixing odd Japanese
Selenium is "γ»γ¬γγ¦γ " in Japanese. (Most people don't write Selenium in Japanese, by the way)
|
# Japanese Language Test - Python 3 Only!
from seleniumbase.translate.japanese import γ»γ¬γγ¦γ γγΉγγ±γΌγΉ # noqa
class γγΉγγ―γ©γΉ(γ»γ¬γγ¦γ γγΉγγ±γΌγΉ): # noqa
def test_δΎ1(self):
self.URLγιγ("https://ja.wikipedia.org/wiki/")
self.γγγΉγγη’Ίθͺγγ("γ¦γ£γγγγ£γ’")
self.θ¦η΄ γη’Ίθͺγγ('[title="γ‘γ€γ³γγΌγΈγ«η§»εγγ"]')
self.γγγΉγγζ΄ζ°("#searchInput", "γ’γγ‘")
self.γ―γͺγγ―γγ¦("#searchButton")
self.γγγΉγγη’Ίθͺγγ("γ’γγ‘", "#firstHeading")
self.γγγΉγγζ΄ζ°("#searchInput", "ε―ΏεΈ")
self.γ―γͺγγ―γγ¦("#searchButton")
self.γγγΉγγη’Ίθͺγγ("ε―ΏεΈ", "#firstHeading")
self.θ¦η΄ γη’Ίθͺγγ('img[alt="ζ‘γε―ΏεΈ"]')
|
<commit_before># Japanese Language Test - Python 3 Only!
from seleniumbase.translate.japanese import γ»γ¬γ³γγΉγγ±γΌγΉ # noqa
class γγΉγγ―γ©γΉ(γ»γ¬γ³γγΉγγ±γΌγΉ): # noqa
def test_δΎ1(self):
self.URLγιγ("https://ja.wikipedia.org/wiki/")
self.γγγΉγγη’Ίθͺγγ("γ¦γ£γγγγ£γ’")
self.θ¦η΄ γη’Ίθͺγγ('[title="γ‘γ€γ³γγΌγΈγ«η§»εγγ"]')
self.γγγΉγγζ΄ζ°("#searchInput", "γ’γγ‘")
self.γ―γͺγγ―γγ¦("#searchButton")
self.γγγΉγγη’Ίθͺγγ("γ’γγ‘", "#firstHeading")
self.γγγΉγγζ΄ζ°("#searchInput", "ε―ΏεΈ")
self.γ―γͺγγ―γγ¦("#searchButton")
self.γγγΉγγη’Ίθͺγγ("ε―ΏεΈ", "#firstHeading")
self.θ¦η΄ γη’Ίθͺγγ('img[alt="ζ‘γε―ΏεΈ"]')
<commit_msg>Update for fixing odd Japanese
Selenium is "γ»γ¬γγ¦γ " in Japanese. (Most people don't write Selenium in Japanese, by the way)<commit_after>
|
# Japanese Language Test - Python 3 Only!
from seleniumbase.translate.japanese import γ»γ¬γγ¦γ γγΉγγ±γΌγΉ # noqa
class γγΉγγ―γ©γΉ(γ»γ¬γγ¦γ γγΉγγ±γΌγΉ): # noqa
def test_δΎ1(self):
self.URLγιγ("https://ja.wikipedia.org/wiki/")
self.γγγΉγγη’Ίθͺγγ("γ¦γ£γγγγ£γ’")
self.θ¦η΄ γη’Ίθͺγγ('[title="γ‘γ€γ³γγΌγΈγ«η§»εγγ"]')
self.γγγΉγγζ΄ζ°("#searchInput", "γ’γγ‘")
self.γ―γͺγγ―γγ¦("#searchButton")
self.γγγΉγγη’Ίθͺγγ("γ’γγ‘", "#firstHeading")
self.γγγΉγγζ΄ζ°("#searchInput", "ε―ΏεΈ")
self.γ―γͺγγ―γγ¦("#searchButton")
self.γγγΉγγη’Ίθͺγγ("ε―ΏεΈ", "#firstHeading")
self.θ¦η΄ γη’Ίθͺγγ('img[alt="ζ‘γε―ΏεΈ"]')
|
# Japanese Language Test - Python 3 Only!
from seleniumbase.translate.japanese import γ»γ¬γ³γγΉγγ±γΌγΉ # noqa
class γγΉγγ―γ©γΉ(γ»γ¬γ³γγΉγγ±γΌγΉ): # noqa
def test_δΎ1(self):
self.URLγιγ("https://ja.wikipedia.org/wiki/")
self.γγγΉγγη’Ίθͺγγ("γ¦γ£γγγγ£γ’")
self.θ¦η΄ γη’Ίθͺγγ('[title="γ‘γ€γ³γγΌγΈγ«η§»εγγ"]')
self.γγγΉγγζ΄ζ°("#searchInput", "γ’γγ‘")
self.γ―γͺγγ―γγ¦("#searchButton")
self.γγγΉγγη’Ίθͺγγ("γ’γγ‘", "#firstHeading")
self.γγγΉγγζ΄ζ°("#searchInput", "ε―ΏεΈ")
self.γ―γͺγγ―γγ¦("#searchButton")
self.γγγΉγγη’Ίθͺγγ("ε―ΏεΈ", "#firstHeading")
self.θ¦η΄ γη’Ίθͺγγ('img[alt="ζ‘γε―ΏεΈ"]')
Update for fixing odd Japanese
Selenium is "γ»γ¬γγ¦γ " in Japanese. (Most people don't write Selenium in Japanese, by the way)# Japanese Language Test - Python 3 Only!
from seleniumbase.translate.japanese import γ»γ¬γγ¦γ γγΉγγ±γΌγΉ # noqa
class γγΉγγ―γ©γΉ(γ»γ¬γγ¦γ γγΉγγ±γΌγΉ): # noqa
def test_δΎ1(self):
self.URLγιγ("https://ja.wikipedia.org/wiki/")
self.γγγΉγγη’Ίθͺγγ("γ¦γ£γγγγ£γ’")
self.θ¦η΄ γη’Ίθͺγγ('[title="γ‘γ€γ³γγΌγΈγ«η§»εγγ"]')
self.γγγΉγγζ΄ζ°("#searchInput", "γ’γγ‘")
self.γ―γͺγγ―γγ¦("#searchButton")
self.γγγΉγγη’Ίθͺγγ("γ’γγ‘", "#firstHeading")
self.γγγΉγγζ΄ζ°("#searchInput", "ε―ΏεΈ")
self.γ―γͺγγ―γγ¦("#searchButton")
self.γγγΉγγη’Ίθͺγγ("ε―ΏεΈ", "#firstHeading")
self.θ¦η΄ γη’Ίθͺγγ('img[alt="ζ‘γε―ΏεΈ"]')
|
<commit_before># Japanese Language Test - Python 3 Only!
from seleniumbase.translate.japanese import γ»γ¬γ³γγΉγγ±γΌγΉ # noqa
class γγΉγγ―γ©γΉ(γ»γ¬γ³γγΉγγ±γΌγΉ): # noqa
def test_δΎ1(self):
self.URLγιγ("https://ja.wikipedia.org/wiki/")
self.γγγΉγγη’Ίθͺγγ("γ¦γ£γγγγ£γ’")
self.θ¦η΄ γη’Ίθͺγγ('[title="γ‘γ€γ³γγΌγΈγ«η§»εγγ"]')
self.γγγΉγγζ΄ζ°("#searchInput", "γ’γγ‘")
self.γ―γͺγγ―γγ¦("#searchButton")
self.γγγΉγγη’Ίθͺγγ("γ’γγ‘", "#firstHeading")
self.γγγΉγγζ΄ζ°("#searchInput", "ε―ΏεΈ")
self.γ―γͺγγ―γγ¦("#searchButton")
self.γγγΉγγη’Ίθͺγγ("ε―ΏεΈ", "#firstHeading")
self.θ¦η΄ γη’Ίθͺγγ('img[alt="ζ‘γε―ΏεΈ"]')
<commit_msg>Update for fixing odd Japanese
Selenium is "γ»γ¬γγ¦γ " in Japanese. (Most people don't write Selenium in Japanese, by the way)<commit_after># Japanese Language Test - Python 3 Only!
from seleniumbase.translate.japanese import γ»γ¬γγ¦γ γγΉγγ±γΌγΉ # noqa
class γγΉγγ―γ©γΉ(γ»γ¬γγ¦γ γγΉγγ±γΌγΉ): # noqa
def test_δΎ1(self):
self.URLγιγ("https://ja.wikipedia.org/wiki/")
self.γγγΉγγη’Ίθͺγγ("γ¦γ£γγγγ£γ’")
self.θ¦η΄ γη’Ίθͺγγ('[title="γ‘γ€γ³γγΌγΈγ«η§»εγγ"]')
self.γγγΉγγζ΄ζ°("#searchInput", "γ’γγ‘")
self.γ―γͺγγ―γγ¦("#searchButton")
self.γγγΉγγη’Ίθͺγγ("γ’γγ‘", "#firstHeading")
self.γγγΉγγζ΄ζ°("#searchInput", "ε―ΏεΈ")
self.γ―γͺγγ―γγ¦("#searchButton")
self.γγγΉγγη’Ίθͺγγ("ε―ΏεΈ", "#firstHeading")
self.θ¦η΄ γη’Ίθͺγγ('img[alt="ζ‘γε―ΏεΈ"]')
|
decb9212a5e0adae31d8e7562fa8258c222aae23
|
dbmigrator/__init__.py
|
dbmigrator/__init__.py
|
# -*- coding: utf-8 -*-
import logging
import sys
logger = logging.getLogger('dbmigrator')
logger.setLevel(logging.INFO)
handler = logging.StreamHandler(sys.stdout)
handler.setFormatter(logging.Formatter('[%(levelname)s] %(name)s (%(filename)s) - %(message)s'))
logger.addHandler(handler)
|
Add a logger for dbmigrator that writes to stdout
|
Add a logger for dbmigrator that writes to stdout
For example in a migration file
20160128111115_mimetype_removal_from_module_files.py:
```
from dbmigrator import logger
logger.info('message from migration')
```
You will see this when you run the migration:
```
[INFO] dbmigrator (20160128111115_mimetype_removal_from_module_files.py) - message from migration
```
Close #3
|
Python
|
agpl-3.0
|
karenc/db-migrator
|
Add a logger for dbmigrator that writes to stdout
For example in a migration file
20160128111115_mimetype_removal_from_module_files.py:
```
from dbmigrator import logger
logger.info('message from migration')
```
You will see this when you run the migration:
```
[INFO] dbmigrator (20160128111115_mimetype_removal_from_module_files.py) - message from migration
```
Close #3
|
# -*- coding: utf-8 -*-
import logging
import sys
logger = logging.getLogger('dbmigrator')
logger.setLevel(logging.INFO)
handler = logging.StreamHandler(sys.stdout)
handler.setFormatter(logging.Formatter('[%(levelname)s] %(name)s (%(filename)s) - %(message)s'))
logger.addHandler(handler)
|
<commit_before><commit_msg>Add a logger for dbmigrator that writes to stdout
For example in a migration file
20160128111115_mimetype_removal_from_module_files.py:
```
from dbmigrator import logger
logger.info('message from migration')
```
You will see this when you run the migration:
```
[INFO] dbmigrator (20160128111115_mimetype_removal_from_module_files.py) - message from migration
```
Close #3<commit_after>
|
# -*- coding: utf-8 -*-
import logging
import sys
logger = logging.getLogger('dbmigrator')
logger.setLevel(logging.INFO)
handler = logging.StreamHandler(sys.stdout)
handler.setFormatter(logging.Formatter('[%(levelname)s] %(name)s (%(filename)s) - %(message)s'))
logger.addHandler(handler)
|
Add a logger for dbmigrator that writes to stdout
For example in a migration file
20160128111115_mimetype_removal_from_module_files.py:
```
from dbmigrator import logger
logger.info('message from migration')
```
You will see this when you run the migration:
```
[INFO] dbmigrator (20160128111115_mimetype_removal_from_module_files.py) - message from migration
```
Close #3# -*- coding: utf-8 -*-
import logging
import sys
logger = logging.getLogger('dbmigrator')
logger.setLevel(logging.INFO)
handler = logging.StreamHandler(sys.stdout)
handler.setFormatter(logging.Formatter('[%(levelname)s] %(name)s (%(filename)s) - %(message)s'))
logger.addHandler(handler)
|
<commit_before><commit_msg>Add a logger for dbmigrator that writes to stdout
For example in a migration file
20160128111115_mimetype_removal_from_module_files.py:
```
from dbmigrator import logger
logger.info('message from migration')
```
You will see this when you run the migration:
```
[INFO] dbmigrator (20160128111115_mimetype_removal_from_module_files.py) - message from migration
```
Close #3<commit_after># -*- coding: utf-8 -*-
import logging
import sys
logger = logging.getLogger('dbmigrator')
logger.setLevel(logging.INFO)
handler = logging.StreamHandler(sys.stdout)
handler.setFormatter(logging.Formatter('[%(levelname)s] %(name)s (%(filename)s) - %(message)s'))
logger.addHandler(handler)
|
|
e15fb53c0fd63942cafd3a6f11418447df6b6800
|
siphon/cdmr/tests/test_coveragedataset.py
|
siphon/cdmr/tests/test_coveragedataset.py
|
# Copyright (c) 2016 Unidata.
# Distributed under the terms of the MIT License.
# SPDX-License-Identifier: MIT
import warnings
from siphon.testing import get_recorder
from siphon.cdmr.coveragedataset import CoverageDataset
recorder = get_recorder(__file__)
# Ignore warnings about CoverageDataset
warnings.simplefilter('ignore')
@recorder.use_cassette('hrrr_cdmremotefeature')
def test_simple_cdmremotefeature():
'Just a smoke test for CDMRemoteFeature'
cd = CoverageDataset('http://localhost:8080/thredds/cdmrfeature/grid/'
'test/HRRR_CONUS_2p5km_20160309_1600.grib2')
assert cd.grids
|
# Copyright (c) 2016 Unidata.
# Distributed under the terms of the MIT License.
# SPDX-License-Identifier: MIT
import warnings
from siphon.testing import get_recorder
from siphon.cdmr.coveragedataset import CoverageDataset
recorder = get_recorder(__file__)
# Ignore warnings about CoverageDataset
warnings.simplefilter('ignore')
@recorder.use_cassette('hrrr_cdmremotefeature')
def test_simple_cdmremotefeature():
'Just a smoke test for CDMRemoteFeature'
cd = CoverageDataset('http://localhost:8080/thredds/cdmrfeature/grid/'
'test/HRRR_CONUS_2p5km_20160309_1600.grib2')
assert cd.grids
@recorder.use_cassette('hrrr_cdmremotefeature')
def test_simple_cdmremotefeature_str():
'Just a smoke test for converting CoverageDataset to str'
cd = CoverageDataset('http://localhost:8080/thredds/cdmrfeature/grid/'
'test/HRRR_CONUS_2p5km_20160309_1600.grib2')
assert str(cd)
|
Add smoketest for convering CoverageDataset to str.
|
Add smoketest for convering CoverageDataset to str.
|
Python
|
bsd-3-clause
|
dopplershift/siphon,dopplershift/siphon,Unidata/siphon
|
# Copyright (c) 2016 Unidata.
# Distributed under the terms of the MIT License.
# SPDX-License-Identifier: MIT
import warnings
from siphon.testing import get_recorder
from siphon.cdmr.coveragedataset import CoverageDataset
recorder = get_recorder(__file__)
# Ignore warnings about CoverageDataset
warnings.simplefilter('ignore')
@recorder.use_cassette('hrrr_cdmremotefeature')
def test_simple_cdmremotefeature():
'Just a smoke test for CDMRemoteFeature'
cd = CoverageDataset('http://localhost:8080/thredds/cdmrfeature/grid/'
'test/HRRR_CONUS_2p5km_20160309_1600.grib2')
assert cd.grids
Add smoketest for convering CoverageDataset to str.
|
# Copyright (c) 2016 Unidata.
# Distributed under the terms of the MIT License.
# SPDX-License-Identifier: MIT
import warnings
from siphon.testing import get_recorder
from siphon.cdmr.coveragedataset import CoverageDataset
recorder = get_recorder(__file__)
# Ignore warnings about CoverageDataset
warnings.simplefilter('ignore')
@recorder.use_cassette('hrrr_cdmremotefeature')
def test_simple_cdmremotefeature():
'Just a smoke test for CDMRemoteFeature'
cd = CoverageDataset('http://localhost:8080/thredds/cdmrfeature/grid/'
'test/HRRR_CONUS_2p5km_20160309_1600.grib2')
assert cd.grids
@recorder.use_cassette('hrrr_cdmremotefeature')
def test_simple_cdmremotefeature_str():
'Just a smoke test for converting CoverageDataset to str'
cd = CoverageDataset('http://localhost:8080/thredds/cdmrfeature/grid/'
'test/HRRR_CONUS_2p5km_20160309_1600.grib2')
assert str(cd)
|
<commit_before># Copyright (c) 2016 Unidata.
# Distributed under the terms of the MIT License.
# SPDX-License-Identifier: MIT
import warnings
from siphon.testing import get_recorder
from siphon.cdmr.coveragedataset import CoverageDataset
recorder = get_recorder(__file__)
# Ignore warnings about CoverageDataset
warnings.simplefilter('ignore')
@recorder.use_cassette('hrrr_cdmremotefeature')
def test_simple_cdmremotefeature():
'Just a smoke test for CDMRemoteFeature'
cd = CoverageDataset('http://localhost:8080/thredds/cdmrfeature/grid/'
'test/HRRR_CONUS_2p5km_20160309_1600.grib2')
assert cd.grids
<commit_msg>Add smoketest for convering CoverageDataset to str.<commit_after>
|
# Copyright (c) 2016 Unidata.
# Distributed under the terms of the MIT License.
# SPDX-License-Identifier: MIT
import warnings
from siphon.testing import get_recorder
from siphon.cdmr.coveragedataset import CoverageDataset
recorder = get_recorder(__file__)
# Ignore warnings about CoverageDataset
warnings.simplefilter('ignore')
@recorder.use_cassette('hrrr_cdmremotefeature')
def test_simple_cdmremotefeature():
'Just a smoke test for CDMRemoteFeature'
cd = CoverageDataset('http://localhost:8080/thredds/cdmrfeature/grid/'
'test/HRRR_CONUS_2p5km_20160309_1600.grib2')
assert cd.grids
@recorder.use_cassette('hrrr_cdmremotefeature')
def test_simple_cdmremotefeature_str():
'Just a smoke test for converting CoverageDataset to str'
cd = CoverageDataset('http://localhost:8080/thredds/cdmrfeature/grid/'
'test/HRRR_CONUS_2p5km_20160309_1600.grib2')
assert str(cd)
|
# Copyright (c) 2016 Unidata.
# Distributed under the terms of the MIT License.
# SPDX-License-Identifier: MIT
import warnings
from siphon.testing import get_recorder
from siphon.cdmr.coveragedataset import CoverageDataset
recorder = get_recorder(__file__)
# Ignore warnings about CoverageDataset
warnings.simplefilter('ignore')
@recorder.use_cassette('hrrr_cdmremotefeature')
def test_simple_cdmremotefeature():
'Just a smoke test for CDMRemoteFeature'
cd = CoverageDataset('http://localhost:8080/thredds/cdmrfeature/grid/'
'test/HRRR_CONUS_2p5km_20160309_1600.grib2')
assert cd.grids
Add smoketest for convering CoverageDataset to str.# Copyright (c) 2016 Unidata.
# Distributed under the terms of the MIT License.
# SPDX-License-Identifier: MIT
import warnings
from siphon.testing import get_recorder
from siphon.cdmr.coveragedataset import CoverageDataset
recorder = get_recorder(__file__)
# Ignore warnings about CoverageDataset
warnings.simplefilter('ignore')
@recorder.use_cassette('hrrr_cdmremotefeature')
def test_simple_cdmremotefeature():
'Just a smoke test for CDMRemoteFeature'
cd = CoverageDataset('http://localhost:8080/thredds/cdmrfeature/grid/'
'test/HRRR_CONUS_2p5km_20160309_1600.grib2')
assert cd.grids
@recorder.use_cassette('hrrr_cdmremotefeature')
def test_simple_cdmremotefeature_str():
'Just a smoke test for converting CoverageDataset to str'
cd = CoverageDataset('http://localhost:8080/thredds/cdmrfeature/grid/'
'test/HRRR_CONUS_2p5km_20160309_1600.grib2')
assert str(cd)
|
<commit_before># Copyright (c) 2016 Unidata.
# Distributed under the terms of the MIT License.
# SPDX-License-Identifier: MIT
import warnings
from siphon.testing import get_recorder
from siphon.cdmr.coveragedataset import CoverageDataset
recorder = get_recorder(__file__)
# Ignore warnings about CoverageDataset
warnings.simplefilter('ignore')
@recorder.use_cassette('hrrr_cdmremotefeature')
def test_simple_cdmremotefeature():
'Just a smoke test for CDMRemoteFeature'
cd = CoverageDataset('http://localhost:8080/thredds/cdmrfeature/grid/'
'test/HRRR_CONUS_2p5km_20160309_1600.grib2')
assert cd.grids
<commit_msg>Add smoketest for convering CoverageDataset to str.<commit_after># Copyright (c) 2016 Unidata.
# Distributed under the terms of the MIT License.
# SPDX-License-Identifier: MIT
import warnings
from siphon.testing import get_recorder
from siphon.cdmr.coveragedataset import CoverageDataset
recorder = get_recorder(__file__)
# Ignore warnings about CoverageDataset
warnings.simplefilter('ignore')
@recorder.use_cassette('hrrr_cdmremotefeature')
def test_simple_cdmremotefeature():
'Just a smoke test for CDMRemoteFeature'
cd = CoverageDataset('http://localhost:8080/thredds/cdmrfeature/grid/'
'test/HRRR_CONUS_2p5km_20160309_1600.grib2')
assert cd.grids
@recorder.use_cassette('hrrr_cdmremotefeature')
def test_simple_cdmremotefeature_str():
'Just a smoke test for converting CoverageDataset to str'
cd = CoverageDataset('http://localhost:8080/thredds/cdmrfeature/grid/'
'test/HRRR_CONUS_2p5km_20160309_1600.grib2')
assert str(cd)
|
95add18b382898eb82c7ff3dd0aa0fd6db0f5cb9
|
setup.py
|
setup.py
|
"""
Flask-MongoKit
--------------
Flask-MongoKit simplifies to use MongoKit, a powerful MongoDB ORM in Flask
applications.
Links
`````
* `documentation <http://packages.python.org/Flask-MongoKit>`_
* `development version <http://github.com/jarus/flask-mongokit/zipball/master#egg=Flask-MongoKit-dev>`_
* `MongoKit <http://namlook.github.com/mongokit/>`_
* `Flask <http://flask.pocoo.org>`_
"""
from setuptools import setup
setup(
name='Flask-MongoKit',
version='0.6',
url='http://github.com/jarus/flask-mongokit',
license='BSD',
author='Christoph Heer',
author_email='Christoph.Heer@googlemail.com',
description='A Flask extension simplifies to use MongoKit',
long_description=__doc__,
py_modules=['flask_mongokit'],
zip_safe=False,
platforms='any',
install_requires=[
'Flask',
'MongoKit'
],
test_suite='tests.suite',
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
'Topic :: Software Development :: Libraries :: Python Modules'
]
)
|
"""
Flask-MongoKit
--------------
Flask-MongoKit simplifies to use MongoKit, a powerful MongoDB ORM in Flask
applications.
Links
`````
* `documentation <http://packages.python.org/Flask-MongoKit>`_
* `development version <http://github.com/jarus/flask-mongokit/zipball/master#egg=Flask-MongoKit-dev>`_
* `MongoKit <http://namlook.github.com/mongokit/>`_
* `Flask <http://flask.pocoo.org>`_
"""
import sys
from setuptools import setup
install_requires = [
"Flask",
"MongoKit"
]
if sys.version_info < (2, 6):
install_requires.append('simplejson')
setup(
name='Flask-MongoKit',
version='0.6',
url='http://github.com/jarus/flask-mongokit',
license='BSD',
author='Christoph Heer',
author_email='Christoph.Heer@googlemail.com',
description='A Flask extension simplifies to use MongoKit',
long_description=__doc__,
py_modules=['flask_mongokit'],
zip_safe=False,
platforms='any',
install_requires=install_requires,
test_suite='tests.suite',
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
'Topic :: Software Development :: Libraries :: Python Modules'
]
)
|
Add simplejson as requirement for python 2.5
|
Add simplejson as requirement for python 2.5
|
Python
|
bsd-3-clause
|
jarus/flask-mongokit,VishvajitP/flask-mongokit,jarus/flask-mongokit,VishvajitP/flask-mongokit
|
"""
Flask-MongoKit
--------------
Flask-MongoKit simplifies to use MongoKit, a powerful MongoDB ORM in Flask
applications.
Links
`````
* `documentation <http://packages.python.org/Flask-MongoKit>`_
* `development version <http://github.com/jarus/flask-mongokit/zipball/master#egg=Flask-MongoKit-dev>`_
* `MongoKit <http://namlook.github.com/mongokit/>`_
* `Flask <http://flask.pocoo.org>`_
"""
from setuptools import setup
setup(
name='Flask-MongoKit',
version='0.6',
url='http://github.com/jarus/flask-mongokit',
license='BSD',
author='Christoph Heer',
author_email='Christoph.Heer@googlemail.com',
description='A Flask extension simplifies to use MongoKit',
long_description=__doc__,
py_modules=['flask_mongokit'],
zip_safe=False,
platforms='any',
install_requires=[
'Flask',
'MongoKit'
],
test_suite='tests.suite',
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
'Topic :: Software Development :: Libraries :: Python Modules'
]
)
Add simplejson as requirement for python 2.5
|
"""
Flask-MongoKit
--------------
Flask-MongoKit simplifies to use MongoKit, a powerful MongoDB ORM in Flask
applications.
Links
`````
* `documentation <http://packages.python.org/Flask-MongoKit>`_
* `development version <http://github.com/jarus/flask-mongokit/zipball/master#egg=Flask-MongoKit-dev>`_
* `MongoKit <http://namlook.github.com/mongokit/>`_
* `Flask <http://flask.pocoo.org>`_
"""
import sys
from setuptools import setup
install_requires = [
"Flask",
"MongoKit"
]
if sys.version_info < (2, 6):
install_requires.append('simplejson')
setup(
name='Flask-MongoKit',
version='0.6',
url='http://github.com/jarus/flask-mongokit',
license='BSD',
author='Christoph Heer',
author_email='Christoph.Heer@googlemail.com',
description='A Flask extension simplifies to use MongoKit',
long_description=__doc__,
py_modules=['flask_mongokit'],
zip_safe=False,
platforms='any',
install_requires=install_requires,
test_suite='tests.suite',
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
'Topic :: Software Development :: Libraries :: Python Modules'
]
)
|
<commit_before>"""
Flask-MongoKit
--------------
Flask-MongoKit simplifies to use MongoKit, a powerful MongoDB ORM in Flask
applications.
Links
`````
* `documentation <http://packages.python.org/Flask-MongoKit>`_
* `development version <http://github.com/jarus/flask-mongokit/zipball/master#egg=Flask-MongoKit-dev>`_
* `MongoKit <http://namlook.github.com/mongokit/>`_
* `Flask <http://flask.pocoo.org>`_
"""
from setuptools import setup
setup(
name='Flask-MongoKit',
version='0.6',
url='http://github.com/jarus/flask-mongokit',
license='BSD',
author='Christoph Heer',
author_email='Christoph.Heer@googlemail.com',
description='A Flask extension simplifies to use MongoKit',
long_description=__doc__,
py_modules=['flask_mongokit'],
zip_safe=False,
platforms='any',
install_requires=[
'Flask',
'MongoKit'
],
test_suite='tests.suite',
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
'Topic :: Software Development :: Libraries :: Python Modules'
]
)
<commit_msg>Add simplejson as requirement for python 2.5<commit_after>
|
"""
Flask-MongoKit
--------------
Flask-MongoKit simplifies to use MongoKit, a powerful MongoDB ORM in Flask
applications.
Links
`````
* `documentation <http://packages.python.org/Flask-MongoKit>`_
* `development version <http://github.com/jarus/flask-mongokit/zipball/master#egg=Flask-MongoKit-dev>`_
* `MongoKit <http://namlook.github.com/mongokit/>`_
* `Flask <http://flask.pocoo.org>`_
"""
import sys
from setuptools import setup
install_requires = [
"Flask",
"MongoKit"
]
if sys.version_info < (2, 6):
install_requires.append('simplejson')
setup(
name='Flask-MongoKit',
version='0.6',
url='http://github.com/jarus/flask-mongokit',
license='BSD',
author='Christoph Heer',
author_email='Christoph.Heer@googlemail.com',
description='A Flask extension simplifies to use MongoKit',
long_description=__doc__,
py_modules=['flask_mongokit'],
zip_safe=False,
platforms='any',
install_requires=install_requires,
test_suite='tests.suite',
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
'Topic :: Software Development :: Libraries :: Python Modules'
]
)
|
"""
Flask-MongoKit
--------------
Flask-MongoKit simplifies to use MongoKit, a powerful MongoDB ORM in Flask
applications.
Links
`````
* `documentation <http://packages.python.org/Flask-MongoKit>`_
* `development version <http://github.com/jarus/flask-mongokit/zipball/master#egg=Flask-MongoKit-dev>`_
* `MongoKit <http://namlook.github.com/mongokit/>`_
* `Flask <http://flask.pocoo.org>`_
"""
from setuptools import setup
setup(
name='Flask-MongoKit',
version='0.6',
url='http://github.com/jarus/flask-mongokit',
license='BSD',
author='Christoph Heer',
author_email='Christoph.Heer@googlemail.com',
description='A Flask extension simplifies to use MongoKit',
long_description=__doc__,
py_modules=['flask_mongokit'],
zip_safe=False,
platforms='any',
install_requires=[
'Flask',
'MongoKit'
],
test_suite='tests.suite',
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
'Topic :: Software Development :: Libraries :: Python Modules'
]
)
Add simplejson as requirement for python 2.5"""
Flask-MongoKit
--------------
Flask-MongoKit simplifies to use MongoKit, a powerful MongoDB ORM in Flask
applications.
Links
`````
* `documentation <http://packages.python.org/Flask-MongoKit>`_
* `development version <http://github.com/jarus/flask-mongokit/zipball/master#egg=Flask-MongoKit-dev>`_
* `MongoKit <http://namlook.github.com/mongokit/>`_
* `Flask <http://flask.pocoo.org>`_
"""
import sys
from setuptools import setup
install_requires = [
"Flask",
"MongoKit"
]
if sys.version_info < (2, 6):
install_requires.append('simplejson')
setup(
name='Flask-MongoKit',
version='0.6',
url='http://github.com/jarus/flask-mongokit',
license='BSD',
author='Christoph Heer',
author_email='Christoph.Heer@googlemail.com',
description='A Flask extension simplifies to use MongoKit',
long_description=__doc__,
py_modules=['flask_mongokit'],
zip_safe=False,
platforms='any',
install_requires=install_requires,
test_suite='tests.suite',
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
'Topic :: Software Development :: Libraries :: Python Modules'
]
)
|
<commit_before>"""
Flask-MongoKit
--------------
Flask-MongoKit simplifies to use MongoKit, a powerful MongoDB ORM in Flask
applications.
Links
`````
* `documentation <http://packages.python.org/Flask-MongoKit>`_
* `development version <http://github.com/jarus/flask-mongokit/zipball/master#egg=Flask-MongoKit-dev>`_
* `MongoKit <http://namlook.github.com/mongokit/>`_
* `Flask <http://flask.pocoo.org>`_
"""
from setuptools import setup
setup(
name='Flask-MongoKit',
version='0.6',
url='http://github.com/jarus/flask-mongokit',
license='BSD',
author='Christoph Heer',
author_email='Christoph.Heer@googlemail.com',
description='A Flask extension simplifies to use MongoKit',
long_description=__doc__,
py_modules=['flask_mongokit'],
zip_safe=False,
platforms='any',
install_requires=[
'Flask',
'MongoKit'
],
test_suite='tests.suite',
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
'Topic :: Software Development :: Libraries :: Python Modules'
]
)
<commit_msg>Add simplejson as requirement for python 2.5<commit_after>"""
Flask-MongoKit
--------------
Flask-MongoKit simplifies to use MongoKit, a powerful MongoDB ORM in Flask
applications.
Links
`````
* `documentation <http://packages.python.org/Flask-MongoKit>`_
* `development version <http://github.com/jarus/flask-mongokit/zipball/master#egg=Flask-MongoKit-dev>`_
* `MongoKit <http://namlook.github.com/mongokit/>`_
* `Flask <http://flask.pocoo.org>`_
"""
import sys
from setuptools import setup
install_requires = [
"Flask",
"MongoKit"
]
if sys.version_info < (2, 6):
install_requires.append('simplejson')
setup(
name='Flask-MongoKit',
version='0.6',
url='http://github.com/jarus/flask-mongokit',
license='BSD',
author='Christoph Heer',
author_email='Christoph.Heer@googlemail.com',
description='A Flask extension simplifies to use MongoKit',
long_description=__doc__,
py_modules=['flask_mongokit'],
zip_safe=False,
platforms='any',
install_requires=install_requires,
test_suite='tests.suite',
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
'Topic :: Software Development :: Libraries :: Python Modules'
]
)
|
c575d69ce253f9eb4d9beb6ffcd3e8a57ed804f0
|
setup.py
|
setup.py
|
from setuptools import setup, find_packages
setup(name='diaspy',
version='0.1.0',
author='Moritz Kiefer',
author_email='moritz.kiefer@gmail.com',
url='https://github.com/Javafant/diaspora-api',
description='A python api to the social network diaspora',
packages=find_packages(),
install_requires=['requests']
)
|
from setuptools import setup, find_packages
setup(name='diaspy',
version='0.2.0',
author='Moritz Kiefer',
author_email='moritz.kiefer@gmail.com',
url='https://github.com/Javafant/diaspora-api',
description='A python api to the social network diaspora',
packages=find_packages(),
install_requires=['requests']
)
|
Update to version 0.2.0 (according with semantic versioning)
|
Update to version 0.2.0 (according with semantic versioning)
|
Python
|
mit
|
marekjm/diaspy
|
from setuptools import setup, find_packages
setup(name='diaspy',
version='0.1.0',
author='Moritz Kiefer',
author_email='moritz.kiefer@gmail.com',
url='https://github.com/Javafant/diaspora-api',
description='A python api to the social network diaspora',
packages=find_packages(),
install_requires=['requests']
)
Update to version 0.2.0 (according with semantic versioning)
|
from setuptools import setup, find_packages
setup(name='diaspy',
version='0.2.0',
author='Moritz Kiefer',
author_email='moritz.kiefer@gmail.com',
url='https://github.com/Javafant/diaspora-api',
description='A python api to the social network diaspora',
packages=find_packages(),
install_requires=['requests']
)
|
<commit_before>from setuptools import setup, find_packages
setup(name='diaspy',
version='0.1.0',
author='Moritz Kiefer',
author_email='moritz.kiefer@gmail.com',
url='https://github.com/Javafant/diaspora-api',
description='A python api to the social network diaspora',
packages=find_packages(),
install_requires=['requests']
)
<commit_msg>Update to version 0.2.0 (according with semantic versioning)<commit_after>
|
from setuptools import setup, find_packages
setup(name='diaspy',
version='0.2.0',
author='Moritz Kiefer',
author_email='moritz.kiefer@gmail.com',
url='https://github.com/Javafant/diaspora-api',
description='A python api to the social network diaspora',
packages=find_packages(),
install_requires=['requests']
)
|
from setuptools import setup, find_packages
setup(name='diaspy',
version='0.1.0',
author='Moritz Kiefer',
author_email='moritz.kiefer@gmail.com',
url='https://github.com/Javafant/diaspora-api',
description='A python api to the social network diaspora',
packages=find_packages(),
install_requires=['requests']
)
Update to version 0.2.0 (according with semantic versioning)from setuptools import setup, find_packages
setup(name='diaspy',
version='0.2.0',
author='Moritz Kiefer',
author_email='moritz.kiefer@gmail.com',
url='https://github.com/Javafant/diaspora-api',
description='A python api to the social network diaspora',
packages=find_packages(),
install_requires=['requests']
)
|
<commit_before>from setuptools import setup, find_packages
setup(name='diaspy',
version='0.1.0',
author='Moritz Kiefer',
author_email='moritz.kiefer@gmail.com',
url='https://github.com/Javafant/diaspora-api',
description='A python api to the social network diaspora',
packages=find_packages(),
install_requires=['requests']
)
<commit_msg>Update to version 0.2.0 (according with semantic versioning)<commit_after>from setuptools import setup, find_packages
setup(name='diaspy',
version='0.2.0',
author='Moritz Kiefer',
author_email='moritz.kiefer@gmail.com',
url='https://github.com/Javafant/diaspora-api',
description='A python api to the social network diaspora',
packages=find_packages(),
install_requires=['requests']
)
|
535ac4c6eae416461e11f33c1a1ef67e92c73914
|
tests/test_exception_wrapping.py
|
tests/test_exception_wrapping.py
|
import safe
def test_simple_exception():
class MockReponse(object):
def json(self):
return {'status': False,
'method': 'synchronize',
'module': 'cluster',
'error': {'message': 'Example error'}}
exception = safe.library.raise_from_json(MockReponse())
assert str(exception) == 'Example error'
|
import safe
class MockResponse(object):
def __init__(self, data):
self.data = data
def json(self):
return self.data
def test_basic_exception():
error_message = 'Example error'
response = MockResponse({
'status': False,
'method': 'synchronize',
'module': 'cluster',
'error': {'message': error_message}
})
exception = safe.library.raise_from_json(response)
assert str(exception) == error_message
def test_commit_failed_exception():
error_message = 'Default ipv4 gateway is not on eth0 subnet'
response = MockResponse({
'status': False,
'type': 'configuration',
'method': 'smartapply',
'module': 'nsc',
'error': {
'message': 'Apply configuration failed.',
'reason': [{
'url': '/SAFe/sng_network_config/modify/network',
'obj_type': 'configuration',
'type': 'ERROR',
'description': error_message,
'module': 'network'
}]
}
})
exception = safe.library.raise_from_json(response)
assert isinstance(exception, safe.CommitFailed)
assert str(exception) == 'Apply changes failed: ' + error_message
assert len(exception.reasons) == 1
reason = exception.reasons[0]
assert reason.obj == 'configuration'
assert reason.module == 'network'
assert reason.description == error_message
|
Add a commit failed test
|
Add a commit failed test
|
Python
|
mpl-2.0
|
sangoma/safepy2,leonardolang/safepy2
|
import safe
def test_simple_exception():
class MockReponse(object):
def json(self):
return {'status': False,
'method': 'synchronize',
'module': 'cluster',
'error': {'message': 'Example error'}}
exception = safe.library.raise_from_json(MockReponse())
assert str(exception) == 'Example error'
Add a commit failed test
|
import safe
class MockResponse(object):
def __init__(self, data):
self.data = data
def json(self):
return self.data
def test_basic_exception():
error_message = 'Example error'
response = MockResponse({
'status': False,
'method': 'synchronize',
'module': 'cluster',
'error': {'message': error_message}
})
exception = safe.library.raise_from_json(response)
assert str(exception) == error_message
def test_commit_failed_exception():
error_message = 'Default ipv4 gateway is not on eth0 subnet'
response = MockResponse({
'status': False,
'type': 'configuration',
'method': 'smartapply',
'module': 'nsc',
'error': {
'message': 'Apply configuration failed.',
'reason': [{
'url': '/SAFe/sng_network_config/modify/network',
'obj_type': 'configuration',
'type': 'ERROR',
'description': error_message,
'module': 'network'
}]
}
})
exception = safe.library.raise_from_json(response)
assert isinstance(exception, safe.CommitFailed)
assert str(exception) == 'Apply changes failed: ' + error_message
assert len(exception.reasons) == 1
reason = exception.reasons[0]
assert reason.obj == 'configuration'
assert reason.module == 'network'
assert reason.description == error_message
|
<commit_before>import safe
def test_simple_exception():
class MockReponse(object):
def json(self):
return {'status': False,
'method': 'synchronize',
'module': 'cluster',
'error': {'message': 'Example error'}}
exception = safe.library.raise_from_json(MockReponse())
assert str(exception) == 'Example error'
<commit_msg>Add a commit failed test<commit_after>
|
import safe
class MockResponse(object):
def __init__(self, data):
self.data = data
def json(self):
return self.data
def test_basic_exception():
error_message = 'Example error'
response = MockResponse({
'status': False,
'method': 'synchronize',
'module': 'cluster',
'error': {'message': error_message}
})
exception = safe.library.raise_from_json(response)
assert str(exception) == error_message
def test_commit_failed_exception():
error_message = 'Default ipv4 gateway is not on eth0 subnet'
response = MockResponse({
'status': False,
'type': 'configuration',
'method': 'smartapply',
'module': 'nsc',
'error': {
'message': 'Apply configuration failed.',
'reason': [{
'url': '/SAFe/sng_network_config/modify/network',
'obj_type': 'configuration',
'type': 'ERROR',
'description': error_message,
'module': 'network'
}]
}
})
exception = safe.library.raise_from_json(response)
assert isinstance(exception, safe.CommitFailed)
assert str(exception) == 'Apply changes failed: ' + error_message
assert len(exception.reasons) == 1
reason = exception.reasons[0]
assert reason.obj == 'configuration'
assert reason.module == 'network'
assert reason.description == error_message
|
import safe
def test_simple_exception():
class MockReponse(object):
def json(self):
return {'status': False,
'method': 'synchronize',
'module': 'cluster',
'error': {'message': 'Example error'}}
exception = safe.library.raise_from_json(MockReponse())
assert str(exception) == 'Example error'
Add a commit failed testimport safe
class MockResponse(object):
def __init__(self, data):
self.data = data
def json(self):
return self.data
def test_basic_exception():
error_message = 'Example error'
response = MockResponse({
'status': False,
'method': 'synchronize',
'module': 'cluster',
'error': {'message': error_message}
})
exception = safe.library.raise_from_json(response)
assert str(exception) == error_message
def test_commit_failed_exception():
error_message = 'Default ipv4 gateway is not on eth0 subnet'
response = MockResponse({
'status': False,
'type': 'configuration',
'method': 'smartapply',
'module': 'nsc',
'error': {
'message': 'Apply configuration failed.',
'reason': [{
'url': '/SAFe/sng_network_config/modify/network',
'obj_type': 'configuration',
'type': 'ERROR',
'description': error_message,
'module': 'network'
}]
}
})
exception = safe.library.raise_from_json(response)
assert isinstance(exception, safe.CommitFailed)
assert str(exception) == 'Apply changes failed: ' + error_message
assert len(exception.reasons) == 1
reason = exception.reasons[0]
assert reason.obj == 'configuration'
assert reason.module == 'network'
assert reason.description == error_message
|
<commit_before>import safe
def test_simple_exception():
class MockReponse(object):
def json(self):
return {'status': False,
'method': 'synchronize',
'module': 'cluster',
'error': {'message': 'Example error'}}
exception = safe.library.raise_from_json(MockReponse())
assert str(exception) == 'Example error'
<commit_msg>Add a commit failed test<commit_after>import safe
class MockResponse(object):
def __init__(self, data):
self.data = data
def json(self):
return self.data
def test_basic_exception():
error_message = 'Example error'
response = MockResponse({
'status': False,
'method': 'synchronize',
'module': 'cluster',
'error': {'message': error_message}
})
exception = safe.library.raise_from_json(response)
assert str(exception) == error_message
def test_commit_failed_exception():
error_message = 'Default ipv4 gateway is not on eth0 subnet'
response = MockResponse({
'status': False,
'type': 'configuration',
'method': 'smartapply',
'module': 'nsc',
'error': {
'message': 'Apply configuration failed.',
'reason': [{
'url': '/SAFe/sng_network_config/modify/network',
'obj_type': 'configuration',
'type': 'ERROR',
'description': error_message,
'module': 'network'
}]
}
})
exception = safe.library.raise_from_json(response)
assert isinstance(exception, safe.CommitFailed)
assert str(exception) == 'Apply changes failed: ' + error_message
assert len(exception.reasons) == 1
reason = exception.reasons[0]
assert reason.obj == 'configuration'
assert reason.module == 'network'
assert reason.description == error_message
|
73d69274a21818830b3a0b87ad574321c958c0f7
|
setup.py
|
setup.py
|
from setuptools import setup
setup(
name="pytest-cpp",
version='0.4',
packages=['pytest_cpp'],
entry_points={
'pytest11': ['cpp = pytest_cpp.plugin'],
},
install_requires=['pytest', 'colorama'],
# metadata for upload to PyPI
author="Bruno Oliveira",
author_email="nicoddemus@gmail.com",
description="Use pytest's runner to discover and execute C++ tests",
long_description=open('README.rst').read(),
license="MIT",
keywords="pytest test unittest",
url="http://github.com/pytest-dev/pytest-cpp",
classifiers=[
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Programming Language :: C++',
'Topic :: Software Development :: Quality Assurance',
'Topic :: Software Development :: Testing',
],
)
|
from setuptools import setup
setup(
name="pytest-cpp",
version='0.4',
packages=['pytest_cpp'],
entry_points={
'pytest11': ['cpp = pytest_cpp.plugin'],
},
install_requires=['pytest', 'colorama'],
# metadata for upload to PyPI
author="Bruno Oliveira",
author_email="nicoddemus@gmail.com",
description="Use pytest's runner to discover and execute C++ tests",
long_description=open('README.rst').read(),
license="MIT",
keywords="pytest test unittest",
url="http://github.com/pytest-dev/pytest-cpp",
classifiers=[
'Development Status :: 4 - Beta',
'Framework :: Pytest',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Programming Language :: C++',
'Topic :: Software Development :: Quality Assurance',
'Topic :: Software Development :: Testing',
],
)
|
Add Framework::Pytest to list of classifiers
|
Add Framework::Pytest to list of classifiers
|
Python
|
mit
|
pytest-dev/pytest-cpp,pytest-dev/pytest-cpp,pytest-dev/pytest-cpp
|
from setuptools import setup
setup(
name="pytest-cpp",
version='0.4',
packages=['pytest_cpp'],
entry_points={
'pytest11': ['cpp = pytest_cpp.plugin'],
},
install_requires=['pytest', 'colorama'],
# metadata for upload to PyPI
author="Bruno Oliveira",
author_email="nicoddemus@gmail.com",
description="Use pytest's runner to discover and execute C++ tests",
long_description=open('README.rst').read(),
license="MIT",
keywords="pytest test unittest",
url="http://github.com/pytest-dev/pytest-cpp",
classifiers=[
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Programming Language :: C++',
'Topic :: Software Development :: Quality Assurance',
'Topic :: Software Development :: Testing',
],
)
Add Framework::Pytest to list of classifiers
|
from setuptools import setup
setup(
name="pytest-cpp",
version='0.4',
packages=['pytest_cpp'],
entry_points={
'pytest11': ['cpp = pytest_cpp.plugin'],
},
install_requires=['pytest', 'colorama'],
# metadata for upload to PyPI
author="Bruno Oliveira",
author_email="nicoddemus@gmail.com",
description="Use pytest's runner to discover and execute C++ tests",
long_description=open('README.rst').read(),
license="MIT",
keywords="pytest test unittest",
url="http://github.com/pytest-dev/pytest-cpp",
classifiers=[
'Development Status :: 4 - Beta',
'Framework :: Pytest',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Programming Language :: C++',
'Topic :: Software Development :: Quality Assurance',
'Topic :: Software Development :: Testing',
],
)
|
<commit_before>from setuptools import setup
setup(
name="pytest-cpp",
version='0.4',
packages=['pytest_cpp'],
entry_points={
'pytest11': ['cpp = pytest_cpp.plugin'],
},
install_requires=['pytest', 'colorama'],
# metadata for upload to PyPI
author="Bruno Oliveira",
author_email="nicoddemus@gmail.com",
description="Use pytest's runner to discover and execute C++ tests",
long_description=open('README.rst').read(),
license="MIT",
keywords="pytest test unittest",
url="http://github.com/pytest-dev/pytest-cpp",
classifiers=[
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Programming Language :: C++',
'Topic :: Software Development :: Quality Assurance',
'Topic :: Software Development :: Testing',
],
)
<commit_msg>Add Framework::Pytest to list of classifiers<commit_after>
|
from setuptools import setup
setup(
name="pytest-cpp",
version='0.4',
packages=['pytest_cpp'],
entry_points={
'pytest11': ['cpp = pytest_cpp.plugin'],
},
install_requires=['pytest', 'colorama'],
# metadata for upload to PyPI
author="Bruno Oliveira",
author_email="nicoddemus@gmail.com",
description="Use pytest's runner to discover and execute C++ tests",
long_description=open('README.rst').read(),
license="MIT",
keywords="pytest test unittest",
url="http://github.com/pytest-dev/pytest-cpp",
classifiers=[
'Development Status :: 4 - Beta',
'Framework :: Pytest',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Programming Language :: C++',
'Topic :: Software Development :: Quality Assurance',
'Topic :: Software Development :: Testing',
],
)
|
from setuptools import setup
setup(
name="pytest-cpp",
version='0.4',
packages=['pytest_cpp'],
entry_points={
'pytest11': ['cpp = pytest_cpp.plugin'],
},
install_requires=['pytest', 'colorama'],
# metadata for upload to PyPI
author="Bruno Oliveira",
author_email="nicoddemus@gmail.com",
description="Use pytest's runner to discover and execute C++ tests",
long_description=open('README.rst').read(),
license="MIT",
keywords="pytest test unittest",
url="http://github.com/pytest-dev/pytest-cpp",
classifiers=[
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Programming Language :: C++',
'Topic :: Software Development :: Quality Assurance',
'Topic :: Software Development :: Testing',
],
)
Add Framework::Pytest to list of classifiersfrom setuptools import setup
setup(
name="pytest-cpp",
version='0.4',
packages=['pytest_cpp'],
entry_points={
'pytest11': ['cpp = pytest_cpp.plugin'],
},
install_requires=['pytest', 'colorama'],
# metadata for upload to PyPI
author="Bruno Oliveira",
author_email="nicoddemus@gmail.com",
description="Use pytest's runner to discover and execute C++ tests",
long_description=open('README.rst').read(),
license="MIT",
keywords="pytest test unittest",
url="http://github.com/pytest-dev/pytest-cpp",
classifiers=[
'Development Status :: 4 - Beta',
'Framework :: Pytest',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Programming Language :: C++',
'Topic :: Software Development :: Quality Assurance',
'Topic :: Software Development :: Testing',
],
)
|
<commit_before>from setuptools import setup
setup(
name="pytest-cpp",
version='0.4',
packages=['pytest_cpp'],
entry_points={
'pytest11': ['cpp = pytest_cpp.plugin'],
},
install_requires=['pytest', 'colorama'],
# metadata for upload to PyPI
author="Bruno Oliveira",
author_email="nicoddemus@gmail.com",
description="Use pytest's runner to discover and execute C++ tests",
long_description=open('README.rst').read(),
license="MIT",
keywords="pytest test unittest",
url="http://github.com/pytest-dev/pytest-cpp",
classifiers=[
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Programming Language :: C++',
'Topic :: Software Development :: Quality Assurance',
'Topic :: Software Development :: Testing',
],
)
<commit_msg>Add Framework::Pytest to list of classifiers<commit_after>from setuptools import setup
setup(
name="pytest-cpp",
version='0.4',
packages=['pytest_cpp'],
entry_points={
'pytest11': ['cpp = pytest_cpp.plugin'],
},
install_requires=['pytest', 'colorama'],
# metadata for upload to PyPI
author="Bruno Oliveira",
author_email="nicoddemus@gmail.com",
description="Use pytest's runner to discover and execute C++ tests",
long_description=open('README.rst').read(),
license="MIT",
keywords="pytest test unittest",
url="http://github.com/pytest-dev/pytest-cpp",
classifiers=[
'Development Status :: 4 - Beta',
'Framework :: Pytest',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.4',
'Programming Language :: Python :: 3.5',
'Programming Language :: C++',
'Topic :: Software Development :: Quality Assurance',
'Topic :: Software Development :: Testing',
],
)
|
6e57b110750e3e871156a7716e95ffed3adf2cd1
|
setup.py
|
setup.py
|
import os, io
from setuptools import setup, find_packages
long_description = (
io.open('README.rst', encoding='utf-8').read()
+ '\n' +
io.open('CHANGES.txt', encoding='utf-8').read())
setup(name='more.chameleon',
version='0.3.dev0',
description="Chameleon template integration for Morepath",
long_description=long_description,
author="Martijn Faassen",
author_email="faassen@startifact.com",
keywords='morepath chameleon',
license="BSD",
url="http://pypi.python.org/pypi/more.chameleon",
namespace_packages=['more'],
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=[
'setuptools',
'morepath >= 0.10',
'chameleon >= 2.20'
],
extras_require = dict(
test=['pytest >= 2.6.0',
'pytest-cov',
'WebTest'],
),
)
|
import io
from setuptools import setup, find_packages
long_description = '\n'.join((
io.open('README.rst', encoding='utf-8').read(),
io.open('CHANGES.txt', encoding='utf-8').read()
))
setup(name='more.chameleon',
version='0.3.dev0',
description="Chameleon template integration for Morepath",
long_description=long_description,
author="Martijn Faassen",
author_email="faassen@startifact.com",
keywords='morepath chameleon',
license="BSD",
url="http://pypi.python.org/pypi/more.chameleon",
namespace_packages=['more'],
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=[
'setuptools',
'morepath >= 0.10',
'chameleon >= 2.20'
],
extras_require=dict(
test=['pytest >= 2.6.0',
'pytest-cov',
'WebTest'],
),
)
|
Use io.open with encoding='utf-8' and flake8 compliance
|
Use io.open with encoding='utf-8' and flake8 compliance
|
Python
|
bsd-3-clause
|
morepath/more.chameleon
|
import os, io
from setuptools import setup, find_packages
long_description = (
io.open('README.rst', encoding='utf-8').read()
+ '\n' +
io.open('CHANGES.txt', encoding='utf-8').read())
setup(name='more.chameleon',
version='0.3.dev0',
description="Chameleon template integration for Morepath",
long_description=long_description,
author="Martijn Faassen",
author_email="faassen@startifact.com",
keywords='morepath chameleon',
license="BSD",
url="http://pypi.python.org/pypi/more.chameleon",
namespace_packages=['more'],
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=[
'setuptools',
'morepath >= 0.10',
'chameleon >= 2.20'
],
extras_require = dict(
test=['pytest >= 2.6.0',
'pytest-cov',
'WebTest'],
),
)
Use io.open with encoding='utf-8' and flake8 compliance
|
import io
from setuptools import setup, find_packages
long_description = '\n'.join((
io.open('README.rst', encoding='utf-8').read(),
io.open('CHANGES.txt', encoding='utf-8').read()
))
setup(name='more.chameleon',
version='0.3.dev0',
description="Chameleon template integration for Morepath",
long_description=long_description,
author="Martijn Faassen",
author_email="faassen@startifact.com",
keywords='morepath chameleon',
license="BSD",
url="http://pypi.python.org/pypi/more.chameleon",
namespace_packages=['more'],
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=[
'setuptools',
'morepath >= 0.10',
'chameleon >= 2.20'
],
extras_require=dict(
test=['pytest >= 2.6.0',
'pytest-cov',
'WebTest'],
),
)
|
<commit_before>import os, io
from setuptools import setup, find_packages
long_description = (
io.open('README.rst', encoding='utf-8').read()
+ '\n' +
io.open('CHANGES.txt', encoding='utf-8').read())
setup(name='more.chameleon',
version='0.3.dev0',
description="Chameleon template integration for Morepath",
long_description=long_description,
author="Martijn Faassen",
author_email="faassen@startifact.com",
keywords='morepath chameleon',
license="BSD",
url="http://pypi.python.org/pypi/more.chameleon",
namespace_packages=['more'],
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=[
'setuptools',
'morepath >= 0.10',
'chameleon >= 2.20'
],
extras_require = dict(
test=['pytest >= 2.6.0',
'pytest-cov',
'WebTest'],
),
)
<commit_msg>Use io.open with encoding='utf-8' and flake8 compliance<commit_after>
|
import io
from setuptools import setup, find_packages
long_description = '\n'.join((
io.open('README.rst', encoding='utf-8').read(),
io.open('CHANGES.txt', encoding='utf-8').read()
))
setup(name='more.chameleon',
version='0.3.dev0',
description="Chameleon template integration for Morepath",
long_description=long_description,
author="Martijn Faassen",
author_email="faassen@startifact.com",
keywords='morepath chameleon',
license="BSD",
url="http://pypi.python.org/pypi/more.chameleon",
namespace_packages=['more'],
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=[
'setuptools',
'morepath >= 0.10',
'chameleon >= 2.20'
],
extras_require=dict(
test=['pytest >= 2.6.0',
'pytest-cov',
'WebTest'],
),
)
|
import os, io
from setuptools import setup, find_packages
long_description = (
io.open('README.rst', encoding='utf-8').read()
+ '\n' +
io.open('CHANGES.txt', encoding='utf-8').read())
setup(name='more.chameleon',
version='0.3.dev0',
description="Chameleon template integration for Morepath",
long_description=long_description,
author="Martijn Faassen",
author_email="faassen@startifact.com",
keywords='morepath chameleon',
license="BSD",
url="http://pypi.python.org/pypi/more.chameleon",
namespace_packages=['more'],
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=[
'setuptools',
'morepath >= 0.10',
'chameleon >= 2.20'
],
extras_require = dict(
test=['pytest >= 2.6.0',
'pytest-cov',
'WebTest'],
),
)
Use io.open with encoding='utf-8' and flake8 complianceimport io
from setuptools import setup, find_packages
long_description = '\n'.join((
io.open('README.rst', encoding='utf-8').read(),
io.open('CHANGES.txt', encoding='utf-8').read()
))
setup(name='more.chameleon',
version='0.3.dev0',
description="Chameleon template integration for Morepath",
long_description=long_description,
author="Martijn Faassen",
author_email="faassen@startifact.com",
keywords='morepath chameleon',
license="BSD",
url="http://pypi.python.org/pypi/more.chameleon",
namespace_packages=['more'],
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=[
'setuptools',
'morepath >= 0.10',
'chameleon >= 2.20'
],
extras_require=dict(
test=['pytest >= 2.6.0',
'pytest-cov',
'WebTest'],
),
)
|
<commit_before>import os, io
from setuptools import setup, find_packages
long_description = (
io.open('README.rst', encoding='utf-8').read()
+ '\n' +
io.open('CHANGES.txt', encoding='utf-8').read())
setup(name='more.chameleon',
version='0.3.dev0',
description="Chameleon template integration for Morepath",
long_description=long_description,
author="Martijn Faassen",
author_email="faassen@startifact.com",
keywords='morepath chameleon',
license="BSD",
url="http://pypi.python.org/pypi/more.chameleon",
namespace_packages=['more'],
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=[
'setuptools',
'morepath >= 0.10',
'chameleon >= 2.20'
],
extras_require = dict(
test=['pytest >= 2.6.0',
'pytest-cov',
'WebTest'],
),
)
<commit_msg>Use io.open with encoding='utf-8' and flake8 compliance<commit_after>import io
from setuptools import setup, find_packages
long_description = '\n'.join((
io.open('README.rst', encoding='utf-8').read(),
io.open('CHANGES.txt', encoding='utf-8').read()
))
setup(name='more.chameleon',
version='0.3.dev0',
description="Chameleon template integration for Morepath",
long_description=long_description,
author="Martijn Faassen",
author_email="faassen@startifact.com",
keywords='morepath chameleon',
license="BSD",
url="http://pypi.python.org/pypi/more.chameleon",
namespace_packages=['more'],
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=[
'setuptools',
'morepath >= 0.10',
'chameleon >= 2.20'
],
extras_require=dict(
test=['pytest >= 2.6.0',
'pytest-cov',
'WebTest'],
),
)
|
d5d359c5ec0f1735e97355839f1a12c6ea45c460
|
polygamy/pygit2_git.py
|
polygamy/pygit2_git.py
|
from __future__ import absolute_import
import pygit2
from .base_git import NoSuchRemote
from .plain_git import PlainGit
class Pygit2Git(PlainGit):
@staticmethod
def is_on_branch(path):
repo = pygit2.Repository(path)
return not (repo.head_is_detached or repo.head_is_unborn)
@staticmethod
def get_remote_url(path, remote_name):
repo = pygit2.Repository(path)
for remote in repo.remotes:
if remote.name == remote_name:
break
else:
raise NoSuchRemote()
return remote.url
|
from __future__ import absolute_import
import pygit2
from .base_git import NoSuchRemote
from .plain_git import PlainGit
class Pygit2Git(PlainGit):
@staticmethod
def is_on_branch(path):
repo = pygit2.Repository(path)
return not (repo.head_is_detached or repo.head_is_unborn)
@staticmethod
def get_remote_url(path, remote_name):
repo = pygit2.Repository(path)
for remote in repo.remotes:
if remote.name == remote_name:
break
else:
raise NoSuchRemote()
return remote.url
@staticmethod
def add_remote(path, remote_name, remote_url):
repo = pygit2.Repository(path)
repo.create_remote(remote_name, remote_url)
|
Add add_remote to pygit2 implementation
|
Add add_remote to pygit2 implementation
|
Python
|
bsd-3-clause
|
solarnz/polygamy,solarnz/polygamy
|
from __future__ import absolute_import
import pygit2
from .base_git import NoSuchRemote
from .plain_git import PlainGit
class Pygit2Git(PlainGit):
@staticmethod
def is_on_branch(path):
repo = pygit2.Repository(path)
return not (repo.head_is_detached or repo.head_is_unborn)
@staticmethod
def get_remote_url(path, remote_name):
repo = pygit2.Repository(path)
for remote in repo.remotes:
if remote.name == remote_name:
break
else:
raise NoSuchRemote()
return remote.url
Add add_remote to pygit2 implementation
|
from __future__ import absolute_import
import pygit2
from .base_git import NoSuchRemote
from .plain_git import PlainGit
class Pygit2Git(PlainGit):
@staticmethod
def is_on_branch(path):
repo = pygit2.Repository(path)
return not (repo.head_is_detached or repo.head_is_unborn)
@staticmethod
def get_remote_url(path, remote_name):
repo = pygit2.Repository(path)
for remote in repo.remotes:
if remote.name == remote_name:
break
else:
raise NoSuchRemote()
return remote.url
@staticmethod
def add_remote(path, remote_name, remote_url):
repo = pygit2.Repository(path)
repo.create_remote(remote_name, remote_url)
|
<commit_before>from __future__ import absolute_import
import pygit2
from .base_git import NoSuchRemote
from .plain_git import PlainGit
class Pygit2Git(PlainGit):
@staticmethod
def is_on_branch(path):
repo = pygit2.Repository(path)
return not (repo.head_is_detached or repo.head_is_unborn)
@staticmethod
def get_remote_url(path, remote_name):
repo = pygit2.Repository(path)
for remote in repo.remotes:
if remote.name == remote_name:
break
else:
raise NoSuchRemote()
return remote.url
<commit_msg>Add add_remote to pygit2 implementation<commit_after>
|
from __future__ import absolute_import
import pygit2
from .base_git import NoSuchRemote
from .plain_git import PlainGit
class Pygit2Git(PlainGit):
@staticmethod
def is_on_branch(path):
repo = pygit2.Repository(path)
return not (repo.head_is_detached or repo.head_is_unborn)
@staticmethod
def get_remote_url(path, remote_name):
repo = pygit2.Repository(path)
for remote in repo.remotes:
if remote.name == remote_name:
break
else:
raise NoSuchRemote()
return remote.url
@staticmethod
def add_remote(path, remote_name, remote_url):
repo = pygit2.Repository(path)
repo.create_remote(remote_name, remote_url)
|
from __future__ import absolute_import
import pygit2
from .base_git import NoSuchRemote
from .plain_git import PlainGit
class Pygit2Git(PlainGit):
@staticmethod
def is_on_branch(path):
repo = pygit2.Repository(path)
return not (repo.head_is_detached or repo.head_is_unborn)
@staticmethod
def get_remote_url(path, remote_name):
repo = pygit2.Repository(path)
for remote in repo.remotes:
if remote.name == remote_name:
break
else:
raise NoSuchRemote()
return remote.url
Add add_remote to pygit2 implementationfrom __future__ import absolute_import
import pygit2
from .base_git import NoSuchRemote
from .plain_git import PlainGit
class Pygit2Git(PlainGit):
@staticmethod
def is_on_branch(path):
repo = pygit2.Repository(path)
return not (repo.head_is_detached or repo.head_is_unborn)
@staticmethod
def get_remote_url(path, remote_name):
repo = pygit2.Repository(path)
for remote in repo.remotes:
if remote.name == remote_name:
break
else:
raise NoSuchRemote()
return remote.url
@staticmethod
def add_remote(path, remote_name, remote_url):
repo = pygit2.Repository(path)
repo.create_remote(remote_name, remote_url)
|
<commit_before>from __future__ import absolute_import
import pygit2
from .base_git import NoSuchRemote
from .plain_git import PlainGit
class Pygit2Git(PlainGit):
@staticmethod
def is_on_branch(path):
repo = pygit2.Repository(path)
return not (repo.head_is_detached or repo.head_is_unborn)
@staticmethod
def get_remote_url(path, remote_name):
repo = pygit2.Repository(path)
for remote in repo.remotes:
if remote.name == remote_name:
break
else:
raise NoSuchRemote()
return remote.url
<commit_msg>Add add_remote to pygit2 implementation<commit_after>from __future__ import absolute_import
import pygit2
from .base_git import NoSuchRemote
from .plain_git import PlainGit
class Pygit2Git(PlainGit):
@staticmethod
def is_on_branch(path):
repo = pygit2.Repository(path)
return not (repo.head_is_detached or repo.head_is_unborn)
@staticmethod
def get_remote_url(path, remote_name):
repo = pygit2.Repository(path)
for remote in repo.remotes:
if remote.name == remote_name:
break
else:
raise NoSuchRemote()
return remote.url
@staticmethod
def add_remote(path, remote_name, remote_url):
repo = pygit2.Repository(path)
repo.create_remote(remote_name, remote_url)
|
5fc65183e40dd1d06bd6ae3e4e7ba0f0a0e2bdd6
|
alg_check_dag.py
|
alg_check_dag.py
|
from __future__ import absolute_import
from __future__ import print_function
from __future__ import division
def _previsit():
pass
def _postvisit():
pass
def _dfs_explore():
pass
def check_dag():
"""Check Directed Acyclic Graph (DAG)."""
pass
def main():
# DAG.
dag_adj_d = {
'A': ['D'],
'B': ['D'],
'C': ['D'],
'D': ['E', 'G'],
'E': ['J'],
'F': ['G'],
'G': ['I'],
'I': ['J'],
'J': []
}
if __name__ == '__main__':
main()
|
from __future__ import absolute_import
from __future__ import print_function
from __future__ import division
def _previsit():
pass
def _postvisit():
pass
def _dfs_explore():
pass
def check_dag():
"""Check Directed Acyclic Graph (DAG)."""
pass
def main():
# Graph adjacency dictionary for DAG.
dag_adj_d = {
'A': ['D'],
'B': ['D'],
'C': ['D'],
'D': ['E', 'G'],
'E': ['J'],
'F': ['G'],
'G': ['I'],
'I': ['J'],
'J': []
}
# Graph adjacency dictionary for non-DAG.
nondag_adj_d = {
'A': ['B'],
'B': ['C', 'E'],
'C': ['C', 'F'],
'D': ['B', 'G'],
'E': ['A', 'D'],
'F': ['H'],
'G': ['E'],
'H': ['I'],
'I': ['F']
}
if __name__ == '__main__':
main()
|
Add DAG & non-DAG adjacency dicts
|
Add DAG & non-DAG adjacency dicts
|
Python
|
bsd-2-clause
|
bowen0701/algorithms_data_structures
|
from __future__ import absolute_import
from __future__ import print_function
from __future__ import division
def _previsit():
pass
def _postvisit():
pass
def _dfs_explore():
pass
def check_dag():
"""Check Directed Acyclic Graph (DAG)."""
pass
def main():
# DAG.
dag_adj_d = {
'A': ['D'],
'B': ['D'],
'C': ['D'],
'D': ['E', 'G'],
'E': ['J'],
'F': ['G'],
'G': ['I'],
'I': ['J'],
'J': []
}
if __name__ == '__main__':
main()
Add DAG & non-DAG adjacency dicts
|
from __future__ import absolute_import
from __future__ import print_function
from __future__ import division
def _previsit():
pass
def _postvisit():
pass
def _dfs_explore():
pass
def check_dag():
"""Check Directed Acyclic Graph (DAG)."""
pass
def main():
# Graph adjacency dictionary for DAG.
dag_adj_d = {
'A': ['D'],
'B': ['D'],
'C': ['D'],
'D': ['E', 'G'],
'E': ['J'],
'F': ['G'],
'G': ['I'],
'I': ['J'],
'J': []
}
# Graph adjacency dictionary for non-DAG.
nondag_adj_d = {
'A': ['B'],
'B': ['C', 'E'],
'C': ['C', 'F'],
'D': ['B', 'G'],
'E': ['A', 'D'],
'F': ['H'],
'G': ['E'],
'H': ['I'],
'I': ['F']
}
if __name__ == '__main__':
main()
|
<commit_before>from __future__ import absolute_import
from __future__ import print_function
from __future__ import division
def _previsit():
pass
def _postvisit():
pass
def _dfs_explore():
pass
def check_dag():
"""Check Directed Acyclic Graph (DAG)."""
pass
def main():
# DAG.
dag_adj_d = {
'A': ['D'],
'B': ['D'],
'C': ['D'],
'D': ['E', 'G'],
'E': ['J'],
'F': ['G'],
'G': ['I'],
'I': ['J'],
'J': []
}
if __name__ == '__main__':
main()
<commit_msg>Add DAG & non-DAG adjacency dicts<commit_after>
|
from __future__ import absolute_import
from __future__ import print_function
from __future__ import division
def _previsit():
pass
def _postvisit():
pass
def _dfs_explore():
pass
def check_dag():
"""Check Directed Acyclic Graph (DAG)."""
pass
def main():
# Graph adjacency dictionary for DAG.
dag_adj_d = {
'A': ['D'],
'B': ['D'],
'C': ['D'],
'D': ['E', 'G'],
'E': ['J'],
'F': ['G'],
'G': ['I'],
'I': ['J'],
'J': []
}
# Graph adjacency dictionary for non-DAG.
nondag_adj_d = {
'A': ['B'],
'B': ['C', 'E'],
'C': ['C', 'F'],
'D': ['B', 'G'],
'E': ['A', 'D'],
'F': ['H'],
'G': ['E'],
'H': ['I'],
'I': ['F']
}
if __name__ == '__main__':
main()
|
from __future__ import absolute_import
from __future__ import print_function
from __future__ import division
def _previsit():
pass
def _postvisit():
pass
def _dfs_explore():
pass
def check_dag():
"""Check Directed Acyclic Graph (DAG)."""
pass
def main():
# DAG.
dag_adj_d = {
'A': ['D'],
'B': ['D'],
'C': ['D'],
'D': ['E', 'G'],
'E': ['J'],
'F': ['G'],
'G': ['I'],
'I': ['J'],
'J': []
}
if __name__ == '__main__':
main()
Add DAG & non-DAG adjacency dictsfrom __future__ import absolute_import
from __future__ import print_function
from __future__ import division
def _previsit():
pass
def _postvisit():
pass
def _dfs_explore():
pass
def check_dag():
"""Check Directed Acyclic Graph (DAG)."""
pass
def main():
# Graph adjacency dictionary for DAG.
dag_adj_d = {
'A': ['D'],
'B': ['D'],
'C': ['D'],
'D': ['E', 'G'],
'E': ['J'],
'F': ['G'],
'G': ['I'],
'I': ['J'],
'J': []
}
# Graph adjacency dictionary for non-DAG.
nondag_adj_d = {
'A': ['B'],
'B': ['C', 'E'],
'C': ['C', 'F'],
'D': ['B', 'G'],
'E': ['A', 'D'],
'F': ['H'],
'G': ['E'],
'H': ['I'],
'I': ['F']
}
if __name__ == '__main__':
main()
|
<commit_before>from __future__ import absolute_import
from __future__ import print_function
from __future__ import division
def _previsit():
pass
def _postvisit():
pass
def _dfs_explore():
pass
def check_dag():
"""Check Directed Acyclic Graph (DAG)."""
pass
def main():
# DAG.
dag_adj_d = {
'A': ['D'],
'B': ['D'],
'C': ['D'],
'D': ['E', 'G'],
'E': ['J'],
'F': ['G'],
'G': ['I'],
'I': ['J'],
'J': []
}
if __name__ == '__main__':
main()
<commit_msg>Add DAG & non-DAG adjacency dicts<commit_after>from __future__ import absolute_import
from __future__ import print_function
from __future__ import division
def _previsit():
pass
def _postvisit():
pass
def _dfs_explore():
pass
def check_dag():
"""Check Directed Acyclic Graph (DAG)."""
pass
def main():
# Graph adjacency dictionary for DAG.
dag_adj_d = {
'A': ['D'],
'B': ['D'],
'C': ['D'],
'D': ['E', 'G'],
'E': ['J'],
'F': ['G'],
'G': ['I'],
'I': ['J'],
'J': []
}
# Graph adjacency dictionary for non-DAG.
nondag_adj_d = {
'A': ['B'],
'B': ['C', 'E'],
'C': ['C', 'F'],
'D': ['B', 'G'],
'E': ['A', 'D'],
'F': ['H'],
'G': ['E'],
'H': ['I'],
'I': ['F']
}
if __name__ == '__main__':
main()
|
4b665bb2e85994e3df0324afacb2453b8f4998a1
|
contact_map/tests/test_dask_runner.py
|
contact_map/tests/test_dask_runner.py
|
# pylint: disable=wildcard-import, missing-docstring, protected-access
# pylint: disable=attribute-defined-outside-init, invalid-name, no-self-use
# pylint: disable=wrong-import-order, unused-wildcard-import
from .utils import *
from contact_map.dask_runner import *
class TestDaskContactFrequency(object):
def test_dask_integration(self):
# this is an integration test to check that dask works
dask = pytest.importorskip('dask') # pylint: disable=W0612
distributed = pytest.importorskip('dask.distributed')
# Explicitly set only 4 workers on Travis instead of 31
# Fix copied from https://github.com/spencerahill/aospy/pull/220/files
cluster = distributed.LocalCluster(n_workers=4)
client = distributed.Client(cluster)
filename = find_testfile("trajectory.pdb")
dask_freq = DaskContactFrequency(client, filename, cutoff=0.075,
n_neighbors_ignored=0)
client.close()
assert dask_freq.n_frames == 5
|
# pylint: disable=wildcard-import, missing-docstring, protected-access
# pylint: disable=attribute-defined-outside-init, invalid-name, no-self-use
# pylint: disable=wrong-import-order, unused-wildcard-import
from .utils import *
from contact_map.dask_runner import *
def dask_setup_test_cluster(distributed, n_workers=4, n_attempts=3):
"""Set up a test cluster using dask.distributed. Try up to n_attempts
times, and skip the test if all attempts fail.
"""
cluster = None
for _ in range(n_attempts):
try:
cluster = distributed.LocalCluster(n_workers=n_workers)
except distributed.TimeoutError:
continue
else:
return cluster
# only get here if all retries fail
pytest.skip("Failed to set up distributed LocalCluster")
class TestDaskContactFrequency(object):
def test_dask_integration(self):
# this is an integration test to check that dask works
dask = pytest.importorskip('dask') # pylint: disable=W0612
distributed = pytest.importorskip('dask.distributed')
# Explicitly set only 4 workers on Travis instead of 31
# Fix copied from https://github.com/spencerahill/aospy/pull/220/files
cluster = dask_setup_test_cluster(distributed)
client = distributed.Client(cluster)
filename = find_testfile("trajectory.pdb")
dask_freq = DaskContactFrequency(client, filename, cutoff=0.075,
n_neighbors_ignored=0)
client.close()
assert dask_freq.n_frames == 5
|
Handle dask TimeoutError exception in tests
|
Handle dask TimeoutError exception in tests
|
Python
|
lgpl-2.1
|
dwhswenson/contact_map,dwhswenson/contact_map
|
# pylint: disable=wildcard-import, missing-docstring, protected-access
# pylint: disable=attribute-defined-outside-init, invalid-name, no-self-use
# pylint: disable=wrong-import-order, unused-wildcard-import
from .utils import *
from contact_map.dask_runner import *
class TestDaskContactFrequency(object):
def test_dask_integration(self):
# this is an integration test to check that dask works
dask = pytest.importorskip('dask') # pylint: disable=W0612
distributed = pytest.importorskip('dask.distributed')
# Explicitly set only 4 workers on Travis instead of 31
# Fix copied from https://github.com/spencerahill/aospy/pull/220/files
cluster = distributed.LocalCluster(n_workers=4)
client = distributed.Client(cluster)
filename = find_testfile("trajectory.pdb")
dask_freq = DaskContactFrequency(client, filename, cutoff=0.075,
n_neighbors_ignored=0)
client.close()
assert dask_freq.n_frames == 5
Handle dask TimeoutError exception in tests
|
# pylint: disable=wildcard-import, missing-docstring, protected-access
# pylint: disable=attribute-defined-outside-init, invalid-name, no-self-use
# pylint: disable=wrong-import-order, unused-wildcard-import
from .utils import *
from contact_map.dask_runner import *
def dask_setup_test_cluster(distributed, n_workers=4, n_attempts=3):
"""Set up a test cluster using dask.distributed. Try up to n_attempts
times, and skip the test if all attempts fail.
"""
cluster = None
for _ in range(n_attempts):
try:
cluster = distributed.LocalCluster(n_workers=n_workers)
except distributed.TimeoutError:
continue
else:
return cluster
# only get here if all retries fail
pytest.skip("Failed to set up distributed LocalCluster")
class TestDaskContactFrequency(object):
def test_dask_integration(self):
# this is an integration test to check that dask works
dask = pytest.importorskip('dask') # pylint: disable=W0612
distributed = pytest.importorskip('dask.distributed')
# Explicitly set only 4 workers on Travis instead of 31
# Fix copied from https://github.com/spencerahill/aospy/pull/220/files
cluster = dask_setup_test_cluster(distributed)
client = distributed.Client(cluster)
filename = find_testfile("trajectory.pdb")
dask_freq = DaskContactFrequency(client, filename, cutoff=0.075,
n_neighbors_ignored=0)
client.close()
assert dask_freq.n_frames == 5
|
<commit_before>
# pylint: disable=wildcard-import, missing-docstring, protected-access
# pylint: disable=attribute-defined-outside-init, invalid-name, no-self-use
# pylint: disable=wrong-import-order, unused-wildcard-import
from .utils import *
from contact_map.dask_runner import *
class TestDaskContactFrequency(object):
def test_dask_integration(self):
# this is an integration test to check that dask works
dask = pytest.importorskip('dask') # pylint: disable=W0612
distributed = pytest.importorskip('dask.distributed')
# Explicitly set only 4 workers on Travis instead of 31
# Fix copied from https://github.com/spencerahill/aospy/pull/220/files
cluster = distributed.LocalCluster(n_workers=4)
client = distributed.Client(cluster)
filename = find_testfile("trajectory.pdb")
dask_freq = DaskContactFrequency(client, filename, cutoff=0.075,
n_neighbors_ignored=0)
client.close()
assert dask_freq.n_frames == 5
<commit_msg>Handle dask TimeoutError exception in tests<commit_after>
|
# pylint: disable=wildcard-import, missing-docstring, protected-access
# pylint: disable=attribute-defined-outside-init, invalid-name, no-self-use
# pylint: disable=wrong-import-order, unused-wildcard-import
from .utils import *
from contact_map.dask_runner import *
def dask_setup_test_cluster(distributed, n_workers=4, n_attempts=3):
"""Set up a test cluster using dask.distributed. Try up to n_attempts
times, and skip the test if all attempts fail.
"""
cluster = None
for _ in range(n_attempts):
try:
cluster = distributed.LocalCluster(n_workers=n_workers)
except distributed.TimeoutError:
continue
else:
return cluster
# only get here if all retries fail
pytest.skip("Failed to set up distributed LocalCluster")
class TestDaskContactFrequency(object):
def test_dask_integration(self):
# this is an integration test to check that dask works
dask = pytest.importorskip('dask') # pylint: disable=W0612
distributed = pytest.importorskip('dask.distributed')
# Explicitly set only 4 workers on Travis instead of 31
# Fix copied from https://github.com/spencerahill/aospy/pull/220/files
cluster = dask_setup_test_cluster(distributed)
client = distributed.Client(cluster)
filename = find_testfile("trajectory.pdb")
dask_freq = DaskContactFrequency(client, filename, cutoff=0.075,
n_neighbors_ignored=0)
client.close()
assert dask_freq.n_frames == 5
|
# pylint: disable=wildcard-import, missing-docstring, protected-access
# pylint: disable=attribute-defined-outside-init, invalid-name, no-self-use
# pylint: disable=wrong-import-order, unused-wildcard-import
from .utils import *
from contact_map.dask_runner import *
class TestDaskContactFrequency(object):
def test_dask_integration(self):
# this is an integration test to check that dask works
dask = pytest.importorskip('dask') # pylint: disable=W0612
distributed = pytest.importorskip('dask.distributed')
# Explicitly set only 4 workers on Travis instead of 31
# Fix copied from https://github.com/spencerahill/aospy/pull/220/files
cluster = distributed.LocalCluster(n_workers=4)
client = distributed.Client(cluster)
filename = find_testfile("trajectory.pdb")
dask_freq = DaskContactFrequency(client, filename, cutoff=0.075,
n_neighbors_ignored=0)
client.close()
assert dask_freq.n_frames == 5
Handle dask TimeoutError exception in tests
# pylint: disable=wildcard-import, missing-docstring, protected-access
# pylint: disable=attribute-defined-outside-init, invalid-name, no-self-use
# pylint: disable=wrong-import-order, unused-wildcard-import
from .utils import *
from contact_map.dask_runner import *
def dask_setup_test_cluster(distributed, n_workers=4, n_attempts=3):
"""Set up a test cluster using dask.distributed. Try up to n_attempts
times, and skip the test if all attempts fail.
"""
cluster = None
for _ in range(n_attempts):
try:
cluster = distributed.LocalCluster(n_workers=n_workers)
except distributed.TimeoutError:
continue
else:
return cluster
# only get here if all retries fail
pytest.skip("Failed to set up distributed LocalCluster")
class TestDaskContactFrequency(object):
def test_dask_integration(self):
# this is an integration test to check that dask works
dask = pytest.importorskip('dask') # pylint: disable=W0612
distributed = pytest.importorskip('dask.distributed')
# Explicitly set only 4 workers on Travis instead of 31
# Fix copied from https://github.com/spencerahill/aospy/pull/220/files
cluster = dask_setup_test_cluster(distributed)
client = distributed.Client(cluster)
filename = find_testfile("trajectory.pdb")
dask_freq = DaskContactFrequency(client, filename, cutoff=0.075,
n_neighbors_ignored=0)
client.close()
assert dask_freq.n_frames == 5
|
<commit_before>
# pylint: disable=wildcard-import, missing-docstring, protected-access
# pylint: disable=attribute-defined-outside-init, invalid-name, no-self-use
# pylint: disable=wrong-import-order, unused-wildcard-import
from .utils import *
from contact_map.dask_runner import *
class TestDaskContactFrequency(object):
def test_dask_integration(self):
# this is an integration test to check that dask works
dask = pytest.importorskip('dask') # pylint: disable=W0612
distributed = pytest.importorskip('dask.distributed')
# Explicitly set only 4 workers on Travis instead of 31
# Fix copied from https://github.com/spencerahill/aospy/pull/220/files
cluster = distributed.LocalCluster(n_workers=4)
client = distributed.Client(cluster)
filename = find_testfile("trajectory.pdb")
dask_freq = DaskContactFrequency(client, filename, cutoff=0.075,
n_neighbors_ignored=0)
client.close()
assert dask_freq.n_frames == 5
<commit_msg>Handle dask TimeoutError exception in tests<commit_after>
# pylint: disable=wildcard-import, missing-docstring, protected-access
# pylint: disable=attribute-defined-outside-init, invalid-name, no-self-use
# pylint: disable=wrong-import-order, unused-wildcard-import
from .utils import *
from contact_map.dask_runner import *
def dask_setup_test_cluster(distributed, n_workers=4, n_attempts=3):
"""Set up a test cluster using dask.distributed. Try up to n_attempts
times, and skip the test if all attempts fail.
"""
cluster = None
for _ in range(n_attempts):
try:
cluster = distributed.LocalCluster(n_workers=n_workers)
except distributed.TimeoutError:
continue
else:
return cluster
# only get here if all retries fail
pytest.skip("Failed to set up distributed LocalCluster")
class TestDaskContactFrequency(object):
def test_dask_integration(self):
# this is an integration test to check that dask works
dask = pytest.importorskip('dask') # pylint: disable=W0612
distributed = pytest.importorskip('dask.distributed')
# Explicitly set only 4 workers on Travis instead of 31
# Fix copied from https://github.com/spencerahill/aospy/pull/220/files
cluster = dask_setup_test_cluster(distributed)
client = distributed.Client(cluster)
filename = find_testfile("trajectory.pdb")
dask_freq = DaskContactFrequency(client, filename, cutoff=0.075,
n_neighbors_ignored=0)
client.close()
assert dask_freq.n_frames == 5
|
e4dd1da8f2fdfa2f4071ab1796b31147f12d00a0
|
setup.py
|
setup.py
|
from setuptools import setup
setup(name='covenant',
version='0.1.0',
description='Code contracts for Python 3',
author='Kamil Kisiel',
author_email='kamil@kamilkisiel.net',
url='http://pypi.python.org/pypi/covenant',
license="BSD License",
packages=["covenant"],
keywords="contract",
platforms=["All"],
install_requires=["decorator"],
classifiers=['Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Natural Language :: English',
'Operating System :: OS Independent',
'Programming Language :: Python :: 3',
'Topic :: Software Development :: Libraries',
'Topic :: Utilities'],
)
|
from setuptools import setup
setup(name='covenant',
version='0.1.0',
description='Code contracts for Python 3',
author='Kamil Kisiel',
author_email='kamil@kamilkisiel.net',
url='http://pypi.python.org/pypi/covenant',
license="BSD License",
packages=["covenant"],
keywords="contract",
platforms=["All"],
install_requires=["decorator"],
classifiers=['Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Natural Language :: English',
'Operating System :: OS Independent',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Topic :: Software Development :: Libraries',
'Topic :: Utilities'],
)
|
Add Python 3.2 trove classifier
|
Add Python 3.2 trove classifier
|
Python
|
mit
|
kisielk/covenant,kisielk/covenant
|
from setuptools import setup
setup(name='covenant',
version='0.1.0',
description='Code contracts for Python 3',
author='Kamil Kisiel',
author_email='kamil@kamilkisiel.net',
url='http://pypi.python.org/pypi/covenant',
license="BSD License",
packages=["covenant"],
keywords="contract",
platforms=["All"],
install_requires=["decorator"],
classifiers=['Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Natural Language :: English',
'Operating System :: OS Independent',
'Programming Language :: Python :: 3',
'Topic :: Software Development :: Libraries',
'Topic :: Utilities'],
)
Add Python 3.2 trove classifier
|
from setuptools import setup
setup(name='covenant',
version='0.1.0',
description='Code contracts for Python 3',
author='Kamil Kisiel',
author_email='kamil@kamilkisiel.net',
url='http://pypi.python.org/pypi/covenant',
license="BSD License",
packages=["covenant"],
keywords="contract",
platforms=["All"],
install_requires=["decorator"],
classifiers=['Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Natural Language :: English',
'Operating System :: OS Independent',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Topic :: Software Development :: Libraries',
'Topic :: Utilities'],
)
|
<commit_before>from setuptools import setup
setup(name='covenant',
version='0.1.0',
description='Code contracts for Python 3',
author='Kamil Kisiel',
author_email='kamil@kamilkisiel.net',
url='http://pypi.python.org/pypi/covenant',
license="BSD License",
packages=["covenant"],
keywords="contract",
platforms=["All"],
install_requires=["decorator"],
classifiers=['Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Natural Language :: English',
'Operating System :: OS Independent',
'Programming Language :: Python :: 3',
'Topic :: Software Development :: Libraries',
'Topic :: Utilities'],
)
<commit_msg>Add Python 3.2 trove classifier<commit_after>
|
from setuptools import setup
setup(name='covenant',
version='0.1.0',
description='Code contracts for Python 3',
author='Kamil Kisiel',
author_email='kamil@kamilkisiel.net',
url='http://pypi.python.org/pypi/covenant',
license="BSD License",
packages=["covenant"],
keywords="contract",
platforms=["All"],
install_requires=["decorator"],
classifiers=['Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Natural Language :: English',
'Operating System :: OS Independent',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Topic :: Software Development :: Libraries',
'Topic :: Utilities'],
)
|
from setuptools import setup
setup(name='covenant',
version='0.1.0',
description='Code contracts for Python 3',
author='Kamil Kisiel',
author_email='kamil@kamilkisiel.net',
url='http://pypi.python.org/pypi/covenant',
license="BSD License",
packages=["covenant"],
keywords="contract",
platforms=["All"],
install_requires=["decorator"],
classifiers=['Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Natural Language :: English',
'Operating System :: OS Independent',
'Programming Language :: Python :: 3',
'Topic :: Software Development :: Libraries',
'Topic :: Utilities'],
)
Add Python 3.2 trove classifierfrom setuptools import setup
setup(name='covenant',
version='0.1.0',
description='Code contracts for Python 3',
author='Kamil Kisiel',
author_email='kamil@kamilkisiel.net',
url='http://pypi.python.org/pypi/covenant',
license="BSD License",
packages=["covenant"],
keywords="contract",
platforms=["All"],
install_requires=["decorator"],
classifiers=['Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Natural Language :: English',
'Operating System :: OS Independent',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Topic :: Software Development :: Libraries',
'Topic :: Utilities'],
)
|
<commit_before>from setuptools import setup
setup(name='covenant',
version='0.1.0',
description='Code contracts for Python 3',
author='Kamil Kisiel',
author_email='kamil@kamilkisiel.net',
url='http://pypi.python.org/pypi/covenant',
license="BSD License",
packages=["covenant"],
keywords="contract",
platforms=["All"],
install_requires=["decorator"],
classifiers=['Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Natural Language :: English',
'Operating System :: OS Independent',
'Programming Language :: Python :: 3',
'Topic :: Software Development :: Libraries',
'Topic :: Utilities'],
)
<commit_msg>Add Python 3.2 trove classifier<commit_after>from setuptools import setup
setup(name='covenant',
version='0.1.0',
description='Code contracts for Python 3',
author='Kamil Kisiel',
author_email='kamil@kamilkisiel.net',
url='http://pypi.python.org/pypi/covenant',
license="BSD License",
packages=["covenant"],
keywords="contract",
platforms=["All"],
install_requires=["decorator"],
classifiers=['Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Natural Language :: English',
'Operating System :: OS Independent',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Topic :: Software Development :: Libraries',
'Topic :: Utilities'],
)
|
44b9649df3418713b8ef5ae8e1f2990a92a48907
|
setup.py
|
setup.py
|
from setuptools import setup
setup(
name='icapservice',
version='0.2.0',
description='ICAP service library for Python',
author='Giles Brown',
author_email='giles_brown@hotmail.com',
url='https://github.com/gilesbrown/icapservice',
license='MIT',
packages=['icapservice'],
zip_safe=False,
install_requires=['six', 'brotlipy'],
include_package_data=True,
package_data={'': ['LICENSE']},
classifiers=(
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'Natural Language :: English',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python',
'Programming Language :: Python :: 2.7',
#'Programming Language :: Python :: 3',
#'Programming Language :: Python :: 3.4',
#'Programming Language :: Python :: 3.5',
),
)
|
from setuptools import setup
setup(
name='icapservice',
version='0.2.1',
description='ICAP service library for Python',
author='Giles Brown',
author_email='giles_brown@hotmail.com',
url='https://github.com/gilesbrown/icapservice',
license='MIT',
packages=['icapservice'],
zip_safe=False,
install_requires=['six', 'brotlipy'],
include_package_data=True,
package_data={'': ['LICENSE']},
classifiers=(
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'Natural Language :: English',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python',
'Programming Language :: Python :: 2.7',
#'Programming Language :: Python :: 3',
#'Programming Language :: Python :: 3.4',
#'Programming Language :: Python :: 3.5',
),
)
|
Patch 0.2.1 to remove print
|
Patch 0.2.1 to remove print
|
Python
|
mit
|
gilesbrown/python-icapservice,gilesbrown/python-icapservice
|
from setuptools import setup
setup(
name='icapservice',
version='0.2.0',
description='ICAP service library for Python',
author='Giles Brown',
author_email='giles_brown@hotmail.com',
url='https://github.com/gilesbrown/icapservice',
license='MIT',
packages=['icapservice'],
zip_safe=False,
install_requires=['six', 'brotlipy'],
include_package_data=True,
package_data={'': ['LICENSE']},
classifiers=(
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'Natural Language :: English',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python',
'Programming Language :: Python :: 2.7',
#'Programming Language :: Python :: 3',
#'Programming Language :: Python :: 3.4',
#'Programming Language :: Python :: 3.5',
),
)
Patch 0.2.1 to remove print
|
from setuptools import setup
setup(
name='icapservice',
version='0.2.1',
description='ICAP service library for Python',
author='Giles Brown',
author_email='giles_brown@hotmail.com',
url='https://github.com/gilesbrown/icapservice',
license='MIT',
packages=['icapservice'],
zip_safe=False,
install_requires=['six', 'brotlipy'],
include_package_data=True,
package_data={'': ['LICENSE']},
classifiers=(
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'Natural Language :: English',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python',
'Programming Language :: Python :: 2.7',
#'Programming Language :: Python :: 3',
#'Programming Language :: Python :: 3.4',
#'Programming Language :: Python :: 3.5',
),
)
|
<commit_before>from setuptools import setup
setup(
name='icapservice',
version='0.2.0',
description='ICAP service library for Python',
author='Giles Brown',
author_email='giles_brown@hotmail.com',
url='https://github.com/gilesbrown/icapservice',
license='MIT',
packages=['icapservice'],
zip_safe=False,
install_requires=['six', 'brotlipy'],
include_package_data=True,
package_data={'': ['LICENSE']},
classifiers=(
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'Natural Language :: English',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python',
'Programming Language :: Python :: 2.7',
#'Programming Language :: Python :: 3',
#'Programming Language :: Python :: 3.4',
#'Programming Language :: Python :: 3.5',
),
)
<commit_msg>Patch 0.2.1 to remove print<commit_after>
|
from setuptools import setup
setup(
name='icapservice',
version='0.2.1',
description='ICAP service library for Python',
author='Giles Brown',
author_email='giles_brown@hotmail.com',
url='https://github.com/gilesbrown/icapservice',
license='MIT',
packages=['icapservice'],
zip_safe=False,
install_requires=['six', 'brotlipy'],
include_package_data=True,
package_data={'': ['LICENSE']},
classifiers=(
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'Natural Language :: English',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python',
'Programming Language :: Python :: 2.7',
#'Programming Language :: Python :: 3',
#'Programming Language :: Python :: 3.4',
#'Programming Language :: Python :: 3.5',
),
)
|
from setuptools import setup
setup(
name='icapservice',
version='0.2.0',
description='ICAP service library for Python',
author='Giles Brown',
author_email='giles_brown@hotmail.com',
url='https://github.com/gilesbrown/icapservice',
license='MIT',
packages=['icapservice'],
zip_safe=False,
install_requires=['six', 'brotlipy'],
include_package_data=True,
package_data={'': ['LICENSE']},
classifiers=(
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'Natural Language :: English',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python',
'Programming Language :: Python :: 2.7',
#'Programming Language :: Python :: 3',
#'Programming Language :: Python :: 3.4',
#'Programming Language :: Python :: 3.5',
),
)
Patch 0.2.1 to remove printfrom setuptools import setup
setup(
name='icapservice',
version='0.2.1',
description='ICAP service library for Python',
author='Giles Brown',
author_email='giles_brown@hotmail.com',
url='https://github.com/gilesbrown/icapservice',
license='MIT',
packages=['icapservice'],
zip_safe=False,
install_requires=['six', 'brotlipy'],
include_package_data=True,
package_data={'': ['LICENSE']},
classifiers=(
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'Natural Language :: English',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python',
'Programming Language :: Python :: 2.7',
#'Programming Language :: Python :: 3',
#'Programming Language :: Python :: 3.4',
#'Programming Language :: Python :: 3.5',
),
)
|
<commit_before>from setuptools import setup
setup(
name='icapservice',
version='0.2.0',
description='ICAP service library for Python',
author='Giles Brown',
author_email='giles_brown@hotmail.com',
url='https://github.com/gilesbrown/icapservice',
license='MIT',
packages=['icapservice'],
zip_safe=False,
install_requires=['six', 'brotlipy'],
include_package_data=True,
package_data={'': ['LICENSE']},
classifiers=(
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'Natural Language :: English',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python',
'Programming Language :: Python :: 2.7',
#'Programming Language :: Python :: 3',
#'Programming Language :: Python :: 3.4',
#'Programming Language :: Python :: 3.5',
),
)
<commit_msg>Patch 0.2.1 to remove print<commit_after>from setuptools import setup
setup(
name='icapservice',
version='0.2.1',
description='ICAP service library for Python',
author='Giles Brown',
author_email='giles_brown@hotmail.com',
url='https://github.com/gilesbrown/icapservice',
license='MIT',
packages=['icapservice'],
zip_safe=False,
install_requires=['six', 'brotlipy'],
include_package_data=True,
package_data={'': ['LICENSE']},
classifiers=(
'Development Status :: 3 - Alpha',
'Intended Audience :: Developers',
'Natural Language :: English',
'License :: OSI Approved :: MIT License',
'Programming Language :: Python',
'Programming Language :: Python :: 2.7',
#'Programming Language :: Python :: 3',
#'Programming Language :: Python :: 3.4',
#'Programming Language :: Python :: 3.5',
),
)
|
86103cbdc457c699f3a76eba914a8708c65bbbbc
|
setup.py
|
setup.py
|
from setuptools import setup
with open('README.md') as file:
# Try my best to have at least the intro in Markdown/reST.
long_description = file.read().partition('<!-- END long_description -->')[0]
setup(name='caniusepython3',
version='1.0',
description='Determine what projects are blocking you from porting to Python 3',
long_description=long_description,
author='Brett Cannon',
author_email='brett@python.org',
url='https://github.com/brettcannon/caniusepython3',
py_modules=['caniusepython3'],
setup_requires=['setuptools'],
install_requires=['distlib', 'pip'],
tests_require=['mock'],
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Console',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
],
entry_points={
'console_scripts': [
'caniusepython3=caniusepython3:main',
]
},
zip_safe=True,
)
|
from setuptools import setup
with open('README.md') as file:
# Try my best to have at least the intro in Markdown/reST.
long_description = file.read().partition('<!-- END long_description -->')[0]
setup(name='caniusepython3',
version='1.0',
description='Determine what projects are blocking you from porting to Python 3',
long_description=long_description,
author='Brett Cannon',
author_email='brett@python.org',
url='https://github.com/brettcannon/caniusepython3',
py_modules=['caniusepython3'],
setup_requires=['setuptools'],
install_requires=['distlib', 'pip'],
tests_require=['mock'],
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Console',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
],
entry_points={
'console_scripts': [
'caniusepython3=caniusepython3:main',
]
},
zip_safe=True,
)
|
Set classifier for Python 3.2
|
Set classifier for Python 3.2
For #6
|
Python
|
apache-2.0
|
ctismer/caniusepython3,dhamaniasad/caniusepython3,nett55/caniusepypy,brettcannon/caniusepython3,svisser/caniusepython3,public/caniusepypy
|
from setuptools import setup
with open('README.md') as file:
# Try my best to have at least the intro in Markdown/reST.
long_description = file.read().partition('<!-- END long_description -->')[0]
setup(name='caniusepython3',
version='1.0',
description='Determine what projects are blocking you from porting to Python 3',
long_description=long_description,
author='Brett Cannon',
author_email='brett@python.org',
url='https://github.com/brettcannon/caniusepython3',
py_modules=['caniusepython3'],
setup_requires=['setuptools'],
install_requires=['distlib', 'pip'],
tests_require=['mock'],
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Console',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
],
entry_points={
'console_scripts': [
'caniusepython3=caniusepython3:main',
]
},
zip_safe=True,
)
Set classifier for Python 3.2
For #6
|
from setuptools import setup
with open('README.md') as file:
# Try my best to have at least the intro in Markdown/reST.
long_description = file.read().partition('<!-- END long_description -->')[0]
setup(name='caniusepython3',
version='1.0',
description='Determine what projects are blocking you from porting to Python 3',
long_description=long_description,
author='Brett Cannon',
author_email='brett@python.org',
url='https://github.com/brettcannon/caniusepython3',
py_modules=['caniusepython3'],
setup_requires=['setuptools'],
install_requires=['distlib', 'pip'],
tests_require=['mock'],
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Console',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
],
entry_points={
'console_scripts': [
'caniusepython3=caniusepython3:main',
]
},
zip_safe=True,
)
|
<commit_before>from setuptools import setup
with open('README.md') as file:
# Try my best to have at least the intro in Markdown/reST.
long_description = file.read().partition('<!-- END long_description -->')[0]
setup(name='caniusepython3',
version='1.0',
description='Determine what projects are blocking you from porting to Python 3',
long_description=long_description,
author='Brett Cannon',
author_email='brett@python.org',
url='https://github.com/brettcannon/caniusepython3',
py_modules=['caniusepython3'],
setup_requires=['setuptools'],
install_requires=['distlib', 'pip'],
tests_require=['mock'],
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Console',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
],
entry_points={
'console_scripts': [
'caniusepython3=caniusepython3:main',
]
},
zip_safe=True,
)
<commit_msg>Set classifier for Python 3.2
For #6<commit_after>
|
from setuptools import setup
with open('README.md') as file:
# Try my best to have at least the intro in Markdown/reST.
long_description = file.read().partition('<!-- END long_description -->')[0]
setup(name='caniusepython3',
version='1.0',
description='Determine what projects are blocking you from porting to Python 3',
long_description=long_description,
author='Brett Cannon',
author_email='brett@python.org',
url='https://github.com/brettcannon/caniusepython3',
py_modules=['caniusepython3'],
setup_requires=['setuptools'],
install_requires=['distlib', 'pip'],
tests_require=['mock'],
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Console',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
],
entry_points={
'console_scripts': [
'caniusepython3=caniusepython3:main',
]
},
zip_safe=True,
)
|
from setuptools import setup
with open('README.md') as file:
# Try my best to have at least the intro in Markdown/reST.
long_description = file.read().partition('<!-- END long_description -->')[0]
setup(name='caniusepython3',
version='1.0',
description='Determine what projects are blocking you from porting to Python 3',
long_description=long_description,
author='Brett Cannon',
author_email='brett@python.org',
url='https://github.com/brettcannon/caniusepython3',
py_modules=['caniusepython3'],
setup_requires=['setuptools'],
install_requires=['distlib', 'pip'],
tests_require=['mock'],
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Console',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
],
entry_points={
'console_scripts': [
'caniusepython3=caniusepython3:main',
]
},
zip_safe=True,
)
Set classifier for Python 3.2
For #6from setuptools import setup
with open('README.md') as file:
# Try my best to have at least the intro in Markdown/reST.
long_description = file.read().partition('<!-- END long_description -->')[0]
setup(name='caniusepython3',
version='1.0',
description='Determine what projects are blocking you from porting to Python 3',
long_description=long_description,
author='Brett Cannon',
author_email='brett@python.org',
url='https://github.com/brettcannon/caniusepython3',
py_modules=['caniusepython3'],
setup_requires=['setuptools'],
install_requires=['distlib', 'pip'],
tests_require=['mock'],
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Console',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
],
entry_points={
'console_scripts': [
'caniusepython3=caniusepython3:main',
]
},
zip_safe=True,
)
|
<commit_before>from setuptools import setup
with open('README.md') as file:
# Try my best to have at least the intro in Markdown/reST.
long_description = file.read().partition('<!-- END long_description -->')[0]
setup(name='caniusepython3',
version='1.0',
description='Determine what projects are blocking you from porting to Python 3',
long_description=long_description,
author='Brett Cannon',
author_email='brett@python.org',
url='https://github.com/brettcannon/caniusepython3',
py_modules=['caniusepython3'],
setup_requires=['setuptools'],
install_requires=['distlib', 'pip'],
tests_require=['mock'],
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Console',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
],
entry_points={
'console_scripts': [
'caniusepython3=caniusepython3:main',
]
},
zip_safe=True,
)
<commit_msg>Set classifier for Python 3.2
For #6<commit_after>from setuptools import setup
with open('README.md') as file:
# Try my best to have at least the intro in Markdown/reST.
long_description = file.read().partition('<!-- END long_description -->')[0]
setup(name='caniusepython3',
version='1.0',
description='Determine what projects are blocking you from porting to Python 3',
long_description=long_description,
author='Brett Cannon',
author_email='brett@python.org',
url='https://github.com/brettcannon/caniusepython3',
py_modules=['caniusepython3'],
setup_requires=['setuptools'],
install_requires=['distlib', 'pip'],
tests_require=['mock'],
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Console',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.2',
'Programming Language :: Python :: 3.3',
],
entry_points={
'console_scripts': [
'caniusepython3=caniusepython3:main',
]
},
zip_safe=True,
)
|
b984060f9e3455df56953580dd3ee4e0717f916b
|
setup.py
|
setup.py
|
#!/usr/bin/env python
from setuptools import setup
setup(
name='rinse',
version='0.1.2',
description='Python3 SOAP client built with lxml and requests.',
long_description=open('README.rst').read(),
author='Tyson Clugg',
author_email='tyson@clugg.net',
url='http://github.com/tysonclugg/rinse',
license='MIT',
packages=['rinse'],
test_suite='rinse.tests',
install_requires=[
'defusedxml',
'lxml',
'requests',
],
classifiers=[
"Programming Language :: Python :: 3",
"Development Status :: 3 - Alpha",
"License :: OSI Approved :: MIT License",
],
)
|
#!/usr/bin/env python
from setuptools import setup
setup(
name='rinse',
version='0.1.2',
description='Python3 SOAP client built with lxml and requests.',
long_description=open('README.rst').read(),
author='Tyson Clugg',
author_email='tyson@clugg.net',
url='http://github.com/tysonclugg/rinse',
license='MIT',
packages=['rinse'],
test_suite='rinse.tests',
install_requires=[
'defusedxml',
'lxml',
'requests',
],
classifiers=[
"Programming Language :: Python :: 2",
"Programming Language :: Python :: 3",
"Development Status :: 3 - Alpha",
"License :: OSI Approved :: MIT License",
],
)
|
Add "Python 2" classifier since it works (according to Travis CI).
|
Add "Python 2" classifier since it works (according to Travis CI).
|
Python
|
mit
|
MarkusH/rinse,simudream/rinse,simudream/rinse,funkybob/rinse,tysonclugg/rinse,thedrow/rinse,tysonclugg/rinse,MarkusH/rinse
|
#!/usr/bin/env python
from setuptools import setup
setup(
name='rinse',
version='0.1.2',
description='Python3 SOAP client built with lxml and requests.',
long_description=open('README.rst').read(),
author='Tyson Clugg',
author_email='tyson@clugg.net',
url='http://github.com/tysonclugg/rinse',
license='MIT',
packages=['rinse'],
test_suite='rinse.tests',
install_requires=[
'defusedxml',
'lxml',
'requests',
],
classifiers=[
"Programming Language :: Python :: 3",
"Development Status :: 3 - Alpha",
"License :: OSI Approved :: MIT License",
],
)
Add "Python 2" classifier since it works (according to Travis CI).
|
#!/usr/bin/env python
from setuptools import setup
setup(
name='rinse',
version='0.1.2',
description='Python3 SOAP client built with lxml and requests.',
long_description=open('README.rst').read(),
author='Tyson Clugg',
author_email='tyson@clugg.net',
url='http://github.com/tysonclugg/rinse',
license='MIT',
packages=['rinse'],
test_suite='rinse.tests',
install_requires=[
'defusedxml',
'lxml',
'requests',
],
classifiers=[
"Programming Language :: Python :: 2",
"Programming Language :: Python :: 3",
"Development Status :: 3 - Alpha",
"License :: OSI Approved :: MIT License",
],
)
|
<commit_before>#!/usr/bin/env python
from setuptools import setup
setup(
name='rinse',
version='0.1.2',
description='Python3 SOAP client built with lxml and requests.',
long_description=open('README.rst').read(),
author='Tyson Clugg',
author_email='tyson@clugg.net',
url='http://github.com/tysonclugg/rinse',
license='MIT',
packages=['rinse'],
test_suite='rinse.tests',
install_requires=[
'defusedxml',
'lxml',
'requests',
],
classifiers=[
"Programming Language :: Python :: 3",
"Development Status :: 3 - Alpha",
"License :: OSI Approved :: MIT License",
],
)
<commit_msg>Add "Python 2" classifier since it works (according to Travis CI).<commit_after>
|
#!/usr/bin/env python
from setuptools import setup
setup(
name='rinse',
version='0.1.2',
description='Python3 SOAP client built with lxml and requests.',
long_description=open('README.rst').read(),
author='Tyson Clugg',
author_email='tyson@clugg.net',
url='http://github.com/tysonclugg/rinse',
license='MIT',
packages=['rinse'],
test_suite='rinse.tests',
install_requires=[
'defusedxml',
'lxml',
'requests',
],
classifiers=[
"Programming Language :: Python :: 2",
"Programming Language :: Python :: 3",
"Development Status :: 3 - Alpha",
"License :: OSI Approved :: MIT License",
],
)
|
#!/usr/bin/env python
from setuptools import setup
setup(
name='rinse',
version='0.1.2',
description='Python3 SOAP client built with lxml and requests.',
long_description=open('README.rst').read(),
author='Tyson Clugg',
author_email='tyson@clugg.net',
url='http://github.com/tysonclugg/rinse',
license='MIT',
packages=['rinse'],
test_suite='rinse.tests',
install_requires=[
'defusedxml',
'lxml',
'requests',
],
classifiers=[
"Programming Language :: Python :: 3",
"Development Status :: 3 - Alpha",
"License :: OSI Approved :: MIT License",
],
)
Add "Python 2" classifier since it works (according to Travis CI).#!/usr/bin/env python
from setuptools import setup
setup(
name='rinse',
version='0.1.2',
description='Python3 SOAP client built with lxml and requests.',
long_description=open('README.rst').read(),
author='Tyson Clugg',
author_email='tyson@clugg.net',
url='http://github.com/tysonclugg/rinse',
license='MIT',
packages=['rinse'],
test_suite='rinse.tests',
install_requires=[
'defusedxml',
'lxml',
'requests',
],
classifiers=[
"Programming Language :: Python :: 2",
"Programming Language :: Python :: 3",
"Development Status :: 3 - Alpha",
"License :: OSI Approved :: MIT License",
],
)
|
<commit_before>#!/usr/bin/env python
from setuptools import setup
setup(
name='rinse',
version='0.1.2',
description='Python3 SOAP client built with lxml and requests.',
long_description=open('README.rst').read(),
author='Tyson Clugg',
author_email='tyson@clugg.net',
url='http://github.com/tysonclugg/rinse',
license='MIT',
packages=['rinse'],
test_suite='rinse.tests',
install_requires=[
'defusedxml',
'lxml',
'requests',
],
classifiers=[
"Programming Language :: Python :: 3",
"Development Status :: 3 - Alpha",
"License :: OSI Approved :: MIT License",
],
)
<commit_msg>Add "Python 2" classifier since it works (according to Travis CI).<commit_after>#!/usr/bin/env python
from setuptools import setup
setup(
name='rinse',
version='0.1.2',
description='Python3 SOAP client built with lxml and requests.',
long_description=open('README.rst').read(),
author='Tyson Clugg',
author_email='tyson@clugg.net',
url='http://github.com/tysonclugg/rinse',
license='MIT',
packages=['rinse'],
test_suite='rinse.tests',
install_requires=[
'defusedxml',
'lxml',
'requests',
],
classifiers=[
"Programming Language :: Python :: 2",
"Programming Language :: Python :: 3",
"Development Status :: 3 - Alpha",
"License :: OSI Approved :: MIT License",
],
)
|
ee8355e8c3d06c6add56ce6962d5ccb9432c98c5
|
setup.py
|
setup.py
|
from setuptools import setup
def readme():
with open('README.rst') as f:
return f.read()
setup(name='housecanary',
version='0.5',
description='Client Wrapper for the HouseCanary API',
long_description=readme(),
url='http://github.com/housecanary/hc-api-python',
author='HouseCanary',
author_email='techops@housecanary.com',
license='MIT',
packages=['housecanary'],
install_requires=['requests','docopt', 'openpyxl'],
zip_safe=False,
test_suite='nose.collector',
tests_require=['nose', 'mock'],
scripts=['bin/hc_api_export'])
|
from setuptools import setup
def readme():
with open('README.rst') as f:
return f.read()
setup(name='housecanary',
version='0.5.1',
description='Client Wrapper for the HouseCanary API',
long_description=readme(),
url='http://github.com/housecanary/hc-api-python',
author='HouseCanary',
author_email='techops@housecanary.com',
license='MIT',
packages=['housecanary', 'housecanary.excel'],
install_requires=['requests','docopt', 'openpyxl'],
zip_safe=False,
test_suite='nose.collector',
tests_require=['nose', 'mock'],
scripts=['bin/hc_api_export'])
|
Add missing package name and bump version
|
Add missing package name and bump version
|
Python
|
mit
|
housecanary/hc-api-python
|
from setuptools import setup
def readme():
with open('README.rst') as f:
return f.read()
setup(name='housecanary',
version='0.5',
description='Client Wrapper for the HouseCanary API',
long_description=readme(),
url='http://github.com/housecanary/hc-api-python',
author='HouseCanary',
author_email='techops@housecanary.com',
license='MIT',
packages=['housecanary'],
install_requires=['requests','docopt', 'openpyxl'],
zip_safe=False,
test_suite='nose.collector',
tests_require=['nose', 'mock'],
scripts=['bin/hc_api_export'])Add missing package name and bump version
|
from setuptools import setup
def readme():
with open('README.rst') as f:
return f.read()
setup(name='housecanary',
version='0.5.1',
description='Client Wrapper for the HouseCanary API',
long_description=readme(),
url='http://github.com/housecanary/hc-api-python',
author='HouseCanary',
author_email='techops@housecanary.com',
license='MIT',
packages=['housecanary', 'housecanary.excel'],
install_requires=['requests','docopt', 'openpyxl'],
zip_safe=False,
test_suite='nose.collector',
tests_require=['nose', 'mock'],
scripts=['bin/hc_api_export'])
|
<commit_before>from setuptools import setup
def readme():
with open('README.rst') as f:
return f.read()
setup(name='housecanary',
version='0.5',
description='Client Wrapper for the HouseCanary API',
long_description=readme(),
url='http://github.com/housecanary/hc-api-python',
author='HouseCanary',
author_email='techops@housecanary.com',
license='MIT',
packages=['housecanary'],
install_requires=['requests','docopt', 'openpyxl'],
zip_safe=False,
test_suite='nose.collector',
tests_require=['nose', 'mock'],
scripts=['bin/hc_api_export'])<commit_msg>Add missing package name and bump version<commit_after>
|
from setuptools import setup
def readme():
with open('README.rst') as f:
return f.read()
setup(name='housecanary',
version='0.5.1',
description='Client Wrapper for the HouseCanary API',
long_description=readme(),
url='http://github.com/housecanary/hc-api-python',
author='HouseCanary',
author_email='techops@housecanary.com',
license='MIT',
packages=['housecanary', 'housecanary.excel'],
install_requires=['requests','docopt', 'openpyxl'],
zip_safe=False,
test_suite='nose.collector',
tests_require=['nose', 'mock'],
scripts=['bin/hc_api_export'])
|
from setuptools import setup
def readme():
with open('README.rst') as f:
return f.read()
setup(name='housecanary',
version='0.5',
description='Client Wrapper for the HouseCanary API',
long_description=readme(),
url='http://github.com/housecanary/hc-api-python',
author='HouseCanary',
author_email='techops@housecanary.com',
license='MIT',
packages=['housecanary'],
install_requires=['requests','docopt', 'openpyxl'],
zip_safe=False,
test_suite='nose.collector',
tests_require=['nose', 'mock'],
scripts=['bin/hc_api_export'])Add missing package name and bump versionfrom setuptools import setup
def readme():
with open('README.rst') as f:
return f.read()
setup(name='housecanary',
version='0.5.1',
description='Client Wrapper for the HouseCanary API',
long_description=readme(),
url='http://github.com/housecanary/hc-api-python',
author='HouseCanary',
author_email='techops@housecanary.com',
license='MIT',
packages=['housecanary', 'housecanary.excel'],
install_requires=['requests','docopt', 'openpyxl'],
zip_safe=False,
test_suite='nose.collector',
tests_require=['nose', 'mock'],
scripts=['bin/hc_api_export'])
|
<commit_before>from setuptools import setup
def readme():
with open('README.rst') as f:
return f.read()
setup(name='housecanary',
version='0.5',
description='Client Wrapper for the HouseCanary API',
long_description=readme(),
url='http://github.com/housecanary/hc-api-python',
author='HouseCanary',
author_email='techops@housecanary.com',
license='MIT',
packages=['housecanary'],
install_requires=['requests','docopt', 'openpyxl'],
zip_safe=False,
test_suite='nose.collector',
tests_require=['nose', 'mock'],
scripts=['bin/hc_api_export'])<commit_msg>Add missing package name and bump version<commit_after>from setuptools import setup
def readme():
with open('README.rst') as f:
return f.read()
setup(name='housecanary',
version='0.5.1',
description='Client Wrapper for the HouseCanary API',
long_description=readme(),
url='http://github.com/housecanary/hc-api-python',
author='HouseCanary',
author_email='techops@housecanary.com',
license='MIT',
packages=['housecanary', 'housecanary.excel'],
install_requires=['requests','docopt', 'openpyxl'],
zip_safe=False,
test_suite='nose.collector',
tests_require=['nose', 'mock'],
scripts=['bin/hc_api_export'])
|
98c80428b35e2a79b4481c5eb9180613266e9698
|
setup.py
|
setup.py
|
from setuptools import find_packages, setup
version = '6.0.0'
install_requires = (
'djangorestframework>=2.4.3,<3',
'incuna_mail>=2.0.0,<3',
)
extras_require = {
'avatar': [
'django-imagekit>=3.2',
],
'utils': [
'raven>=5.1.1',
],
}
setup(
name='django-user-management',
packages=find_packages(),
include_package_data=True,
version=version,
description='User management model mixins and api views.',
long_description='',
keywords='django rest framework user management api',
author='Incuna',
author_email='admin@incuna.com',
url='https://github.com/incuna/django-user-management/',
install_requires=install_requires,
extras_require=extras_require,
zip_safe=False,
license='BSD',
classifiers=[
'Development Status :: 5 - Production/Stable',
'Environment :: Web Environment',
'Framework :: Django',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Topic :: Software Development',
'Topic :: Utilities',
],
)
|
from setuptools import find_packages, setup
version = '6.0.0'
install_requires = (
'djangorestframework>=2.4.4,<3',
'incuna_mail>=2.0.0,<3',
)
extras_require = {
'avatar': [
'django-imagekit>=3.2',
],
'utils': [
'raven>=5.1.1',
],
}
setup(
name='django-user-management',
packages=find_packages(),
include_package_data=True,
version=version,
description='User management model mixins and api views.',
long_description='',
keywords='django rest framework user management api',
author='Incuna',
author_email='admin@incuna.com',
url='https://github.com/incuna/django-user-management/',
install_requires=install_requires,
extras_require=extras_require,
zip_safe=False,
license='BSD',
classifiers=[
'Development Status :: 5 - Production/Stable',
'Environment :: Web Environment',
'Framework :: Django',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Topic :: Software Development',
'Topic :: Utilities',
],
)
|
Update `djangorestframework` from 2.4.3 to 2.4.4
|
Update `djangorestframework` from 2.4.3 to 2.4.4
|
Python
|
bsd-2-clause
|
incuna/django-user-management,incuna/django-user-management
|
from setuptools import find_packages, setup
version = '6.0.0'
install_requires = (
'djangorestframework>=2.4.3,<3',
'incuna_mail>=2.0.0,<3',
)
extras_require = {
'avatar': [
'django-imagekit>=3.2',
],
'utils': [
'raven>=5.1.1',
],
}
setup(
name='django-user-management',
packages=find_packages(),
include_package_data=True,
version=version,
description='User management model mixins and api views.',
long_description='',
keywords='django rest framework user management api',
author='Incuna',
author_email='admin@incuna.com',
url='https://github.com/incuna/django-user-management/',
install_requires=install_requires,
extras_require=extras_require,
zip_safe=False,
license='BSD',
classifiers=[
'Development Status :: 5 - Production/Stable',
'Environment :: Web Environment',
'Framework :: Django',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Topic :: Software Development',
'Topic :: Utilities',
],
)
Update `djangorestframework` from 2.4.3 to 2.4.4
|
from setuptools import find_packages, setup
version = '6.0.0'
install_requires = (
'djangorestframework>=2.4.4,<3',
'incuna_mail>=2.0.0,<3',
)
extras_require = {
'avatar': [
'django-imagekit>=3.2',
],
'utils': [
'raven>=5.1.1',
],
}
setup(
name='django-user-management',
packages=find_packages(),
include_package_data=True,
version=version,
description='User management model mixins and api views.',
long_description='',
keywords='django rest framework user management api',
author='Incuna',
author_email='admin@incuna.com',
url='https://github.com/incuna/django-user-management/',
install_requires=install_requires,
extras_require=extras_require,
zip_safe=False,
license='BSD',
classifiers=[
'Development Status :: 5 - Production/Stable',
'Environment :: Web Environment',
'Framework :: Django',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Topic :: Software Development',
'Topic :: Utilities',
],
)
|
<commit_before>from setuptools import find_packages, setup
version = '6.0.0'
install_requires = (
'djangorestframework>=2.4.3,<3',
'incuna_mail>=2.0.0,<3',
)
extras_require = {
'avatar': [
'django-imagekit>=3.2',
],
'utils': [
'raven>=5.1.1',
],
}
setup(
name='django-user-management',
packages=find_packages(),
include_package_data=True,
version=version,
description='User management model mixins and api views.',
long_description='',
keywords='django rest framework user management api',
author='Incuna',
author_email='admin@incuna.com',
url='https://github.com/incuna/django-user-management/',
install_requires=install_requires,
extras_require=extras_require,
zip_safe=False,
license='BSD',
classifiers=[
'Development Status :: 5 - Production/Stable',
'Environment :: Web Environment',
'Framework :: Django',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Topic :: Software Development',
'Topic :: Utilities',
],
)
<commit_msg>Update `djangorestframework` from 2.4.3 to 2.4.4<commit_after>
|
from setuptools import find_packages, setup
version = '6.0.0'
install_requires = (
'djangorestframework>=2.4.4,<3',
'incuna_mail>=2.0.0,<3',
)
extras_require = {
'avatar': [
'django-imagekit>=3.2',
],
'utils': [
'raven>=5.1.1',
],
}
setup(
name='django-user-management',
packages=find_packages(),
include_package_data=True,
version=version,
description='User management model mixins and api views.',
long_description='',
keywords='django rest framework user management api',
author='Incuna',
author_email='admin@incuna.com',
url='https://github.com/incuna/django-user-management/',
install_requires=install_requires,
extras_require=extras_require,
zip_safe=False,
license='BSD',
classifiers=[
'Development Status :: 5 - Production/Stable',
'Environment :: Web Environment',
'Framework :: Django',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Topic :: Software Development',
'Topic :: Utilities',
],
)
|
from setuptools import find_packages, setup
version = '6.0.0'
install_requires = (
'djangorestframework>=2.4.3,<3',
'incuna_mail>=2.0.0,<3',
)
extras_require = {
'avatar': [
'django-imagekit>=3.2',
],
'utils': [
'raven>=5.1.1',
],
}
setup(
name='django-user-management',
packages=find_packages(),
include_package_data=True,
version=version,
description='User management model mixins and api views.',
long_description='',
keywords='django rest framework user management api',
author='Incuna',
author_email='admin@incuna.com',
url='https://github.com/incuna/django-user-management/',
install_requires=install_requires,
extras_require=extras_require,
zip_safe=False,
license='BSD',
classifiers=[
'Development Status :: 5 - Production/Stable',
'Environment :: Web Environment',
'Framework :: Django',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Topic :: Software Development',
'Topic :: Utilities',
],
)
Update `djangorestframework` from 2.4.3 to 2.4.4from setuptools import find_packages, setup
version = '6.0.0'
install_requires = (
'djangorestframework>=2.4.4,<3',
'incuna_mail>=2.0.0,<3',
)
extras_require = {
'avatar': [
'django-imagekit>=3.2',
],
'utils': [
'raven>=5.1.1',
],
}
setup(
name='django-user-management',
packages=find_packages(),
include_package_data=True,
version=version,
description='User management model mixins and api views.',
long_description='',
keywords='django rest framework user management api',
author='Incuna',
author_email='admin@incuna.com',
url='https://github.com/incuna/django-user-management/',
install_requires=install_requires,
extras_require=extras_require,
zip_safe=False,
license='BSD',
classifiers=[
'Development Status :: 5 - Production/Stable',
'Environment :: Web Environment',
'Framework :: Django',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Topic :: Software Development',
'Topic :: Utilities',
],
)
|
<commit_before>from setuptools import find_packages, setup
version = '6.0.0'
install_requires = (
'djangorestframework>=2.4.3,<3',
'incuna_mail>=2.0.0,<3',
)
extras_require = {
'avatar': [
'django-imagekit>=3.2',
],
'utils': [
'raven>=5.1.1',
],
}
setup(
name='django-user-management',
packages=find_packages(),
include_package_data=True,
version=version,
description='User management model mixins and api views.',
long_description='',
keywords='django rest framework user management api',
author='Incuna',
author_email='admin@incuna.com',
url='https://github.com/incuna/django-user-management/',
install_requires=install_requires,
extras_require=extras_require,
zip_safe=False,
license='BSD',
classifiers=[
'Development Status :: 5 - Production/Stable',
'Environment :: Web Environment',
'Framework :: Django',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Topic :: Software Development',
'Topic :: Utilities',
],
)
<commit_msg>Update `djangorestframework` from 2.4.3 to 2.4.4<commit_after>from setuptools import find_packages, setup
version = '6.0.0'
install_requires = (
'djangorestframework>=2.4.4,<3',
'incuna_mail>=2.0.0,<3',
)
extras_require = {
'avatar': [
'django-imagekit>=3.2',
],
'utils': [
'raven>=5.1.1',
],
}
setup(
name='django-user-management',
packages=find_packages(),
include_package_data=True,
version=version,
description='User management model mixins and api views.',
long_description='',
keywords='django rest framework user management api',
author='Incuna',
author_email='admin@incuna.com',
url='https://github.com/incuna/django-user-management/',
install_requires=install_requires,
extras_require=extras_require,
zip_safe=False,
license='BSD',
classifiers=[
'Development Status :: 5 - Production/Stable',
'Environment :: Web Environment',
'Framework :: Django',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: 3.4',
'Topic :: Software Development',
'Topic :: Utilities',
],
)
|
350f3bb3431d451f6bf6f2fac2e696b9122d65a6
|
setup.py
|
setup.py
|
#!/usr/bin/env python
import os
import sys
from setuptools import setup
readme = open('README.rst').read()
doclink = """
Documentation
-------------
The full documentation is at http://gryaml.rtfd.org."""
history = open('HISTORY.rst').read().replace('.. :changelog:', '')
setup(
name='gryaml',
use_scm_version=True,
description='Represent Neo4j graph data as YAML.',
long_description=readme + '\n\n' + doclink + '\n\n' + history,
author='Wil Cooley',
author_email='wcooley@nakedape.cc',
url='https://github.com/wcooley/python-gryaml',
packages=[
'gryaml',
],
package_dir={'gryaml': 'gryaml'},
include_package_data=True,
install_requires=[
'py2neo>=2.0,<3',
'pyyaml',
],
setup_requires=['setuptools_scm'],
license='MIT',
zip_safe=False,
keywords='gryaml',
classifiers=[
'Development Status :: 2 - Pre-Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Natural Language :: English',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: Implementation :: PyPy',
],
)
|
#!/usr/bin/env python
"""Setuptools setup."""
from setuptools import setup
readme = open('README.rst').read()
doclink = """
Documentation
-------------
The full documentation is at http://gryaml.rtfd.org."""
history = open('HISTORY.rst').read().replace('.. :changelog:', '')
setup(
name='gryaml',
use_scm_version=True,
description='Represent Neo4j graph data as YAML.',
long_description=readme + '\n\n' + doclink + '\n\n' + history,
author='Wil Cooley',
author_email='wcooley@nakedape.cc',
url='https://github.com/wcooley/python-gryaml',
packages=[
'gryaml',
],
package_dir={'gryaml': 'gryaml'},
include_package_data=True,
install_requires=[
'boltons',
'py2neo>=2.0,<3',
'pyyaml',
],
setup_requires=['setuptools_scm'],
license='MIT',
zip_safe=False,
keywords='gryaml',
classifiers=[
'Development Status :: 2 - Pre-Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Natural Language :: English',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: Implementation :: PyPy',
],
)
|
Add missing 'boltons' package & clean up
|
Add missing 'boltons' package & clean up
|
Python
|
mit
|
wcooley/python-gryaml
|
#!/usr/bin/env python
import os
import sys
from setuptools import setup
readme = open('README.rst').read()
doclink = """
Documentation
-------------
The full documentation is at http://gryaml.rtfd.org."""
history = open('HISTORY.rst').read().replace('.. :changelog:', '')
setup(
name='gryaml',
use_scm_version=True,
description='Represent Neo4j graph data as YAML.',
long_description=readme + '\n\n' + doclink + '\n\n' + history,
author='Wil Cooley',
author_email='wcooley@nakedape.cc',
url='https://github.com/wcooley/python-gryaml',
packages=[
'gryaml',
],
package_dir={'gryaml': 'gryaml'},
include_package_data=True,
install_requires=[
'py2neo>=2.0,<3',
'pyyaml',
],
setup_requires=['setuptools_scm'],
license='MIT',
zip_safe=False,
keywords='gryaml',
classifiers=[
'Development Status :: 2 - Pre-Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Natural Language :: English',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: Implementation :: PyPy',
],
)
Add missing 'boltons' package & clean up
|
#!/usr/bin/env python
"""Setuptools setup."""
from setuptools import setup
readme = open('README.rst').read()
doclink = """
Documentation
-------------
The full documentation is at http://gryaml.rtfd.org."""
history = open('HISTORY.rst').read().replace('.. :changelog:', '')
setup(
name='gryaml',
use_scm_version=True,
description='Represent Neo4j graph data as YAML.',
long_description=readme + '\n\n' + doclink + '\n\n' + history,
author='Wil Cooley',
author_email='wcooley@nakedape.cc',
url='https://github.com/wcooley/python-gryaml',
packages=[
'gryaml',
],
package_dir={'gryaml': 'gryaml'},
include_package_data=True,
install_requires=[
'boltons',
'py2neo>=2.0,<3',
'pyyaml',
],
setup_requires=['setuptools_scm'],
license='MIT',
zip_safe=False,
keywords='gryaml',
classifiers=[
'Development Status :: 2 - Pre-Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Natural Language :: English',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: Implementation :: PyPy',
],
)
|
<commit_before>#!/usr/bin/env python
import os
import sys
from setuptools import setup
readme = open('README.rst').read()
doclink = """
Documentation
-------------
The full documentation is at http://gryaml.rtfd.org."""
history = open('HISTORY.rst').read().replace('.. :changelog:', '')
setup(
name='gryaml',
use_scm_version=True,
description='Represent Neo4j graph data as YAML.',
long_description=readme + '\n\n' + doclink + '\n\n' + history,
author='Wil Cooley',
author_email='wcooley@nakedape.cc',
url='https://github.com/wcooley/python-gryaml',
packages=[
'gryaml',
],
package_dir={'gryaml': 'gryaml'},
include_package_data=True,
install_requires=[
'py2neo>=2.0,<3',
'pyyaml',
],
setup_requires=['setuptools_scm'],
license='MIT',
zip_safe=False,
keywords='gryaml',
classifiers=[
'Development Status :: 2 - Pre-Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Natural Language :: English',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: Implementation :: PyPy',
],
)
<commit_msg>Add missing 'boltons' package & clean up<commit_after>
|
#!/usr/bin/env python
"""Setuptools setup."""
from setuptools import setup
readme = open('README.rst').read()
doclink = """
Documentation
-------------
The full documentation is at http://gryaml.rtfd.org."""
history = open('HISTORY.rst').read().replace('.. :changelog:', '')
setup(
name='gryaml',
use_scm_version=True,
description='Represent Neo4j graph data as YAML.',
long_description=readme + '\n\n' + doclink + '\n\n' + history,
author='Wil Cooley',
author_email='wcooley@nakedape.cc',
url='https://github.com/wcooley/python-gryaml',
packages=[
'gryaml',
],
package_dir={'gryaml': 'gryaml'},
include_package_data=True,
install_requires=[
'boltons',
'py2neo>=2.0,<3',
'pyyaml',
],
setup_requires=['setuptools_scm'],
license='MIT',
zip_safe=False,
keywords='gryaml',
classifiers=[
'Development Status :: 2 - Pre-Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Natural Language :: English',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: Implementation :: PyPy',
],
)
|
#!/usr/bin/env python
import os
import sys
from setuptools import setup
readme = open('README.rst').read()
doclink = """
Documentation
-------------
The full documentation is at http://gryaml.rtfd.org."""
history = open('HISTORY.rst').read().replace('.. :changelog:', '')
setup(
name='gryaml',
use_scm_version=True,
description='Represent Neo4j graph data as YAML.',
long_description=readme + '\n\n' + doclink + '\n\n' + history,
author='Wil Cooley',
author_email='wcooley@nakedape.cc',
url='https://github.com/wcooley/python-gryaml',
packages=[
'gryaml',
],
package_dir={'gryaml': 'gryaml'},
include_package_data=True,
install_requires=[
'py2neo>=2.0,<3',
'pyyaml',
],
setup_requires=['setuptools_scm'],
license='MIT',
zip_safe=False,
keywords='gryaml',
classifiers=[
'Development Status :: 2 - Pre-Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Natural Language :: English',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: Implementation :: PyPy',
],
)
Add missing 'boltons' package & clean up#!/usr/bin/env python
"""Setuptools setup."""
from setuptools import setup
readme = open('README.rst').read()
doclink = """
Documentation
-------------
The full documentation is at http://gryaml.rtfd.org."""
history = open('HISTORY.rst').read().replace('.. :changelog:', '')
setup(
name='gryaml',
use_scm_version=True,
description='Represent Neo4j graph data as YAML.',
long_description=readme + '\n\n' + doclink + '\n\n' + history,
author='Wil Cooley',
author_email='wcooley@nakedape.cc',
url='https://github.com/wcooley/python-gryaml',
packages=[
'gryaml',
],
package_dir={'gryaml': 'gryaml'},
include_package_data=True,
install_requires=[
'boltons',
'py2neo>=2.0,<3',
'pyyaml',
],
setup_requires=['setuptools_scm'],
license='MIT',
zip_safe=False,
keywords='gryaml',
classifiers=[
'Development Status :: 2 - Pre-Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Natural Language :: English',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: Implementation :: PyPy',
],
)
|
<commit_before>#!/usr/bin/env python
import os
import sys
from setuptools import setup
readme = open('README.rst').read()
doclink = """
Documentation
-------------
The full documentation is at http://gryaml.rtfd.org."""
history = open('HISTORY.rst').read().replace('.. :changelog:', '')
setup(
name='gryaml',
use_scm_version=True,
description='Represent Neo4j graph data as YAML.',
long_description=readme + '\n\n' + doclink + '\n\n' + history,
author='Wil Cooley',
author_email='wcooley@nakedape.cc',
url='https://github.com/wcooley/python-gryaml',
packages=[
'gryaml',
],
package_dir={'gryaml': 'gryaml'},
include_package_data=True,
install_requires=[
'py2neo>=2.0,<3',
'pyyaml',
],
setup_requires=['setuptools_scm'],
license='MIT',
zip_safe=False,
keywords='gryaml',
classifiers=[
'Development Status :: 2 - Pre-Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Natural Language :: English',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: Implementation :: PyPy',
],
)
<commit_msg>Add missing 'boltons' package & clean up<commit_after>#!/usr/bin/env python
"""Setuptools setup."""
from setuptools import setup
readme = open('README.rst').read()
doclink = """
Documentation
-------------
The full documentation is at http://gryaml.rtfd.org."""
history = open('HISTORY.rst').read().replace('.. :changelog:', '')
setup(
name='gryaml',
use_scm_version=True,
description='Represent Neo4j graph data as YAML.',
long_description=readme + '\n\n' + doclink + '\n\n' + history,
author='Wil Cooley',
author_email='wcooley@nakedape.cc',
url='https://github.com/wcooley/python-gryaml',
packages=[
'gryaml',
],
package_dir={'gryaml': 'gryaml'},
include_package_data=True,
install_requires=[
'boltons',
'py2neo>=2.0,<3',
'pyyaml',
],
setup_requires=['setuptools_scm'],
license='MIT',
zip_safe=False,
keywords='gryaml',
classifiers=[
'Development Status :: 2 - Pre-Alpha',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Natural Language :: English',
'Programming Language :: Python :: 2',
'Programming Language :: Python :: 2.6',
'Programming Language :: Python :: 2.7',
'Programming Language :: Python :: 3',
'Programming Language :: Python :: 3.3',
'Programming Language :: Python :: Implementation :: PyPy',
],
)
|
672d6cb517198bc37c4126c997ba984901a14b55
|
setup.py
|
setup.py
|
import os, io
from setuptools import setup, find_packages
long_description = (
io.open('README.rst', encoding='utf-8').read()
+ '\n' +
io.open('CHANGES.txt', encoding='utf-8').read())
setup(name='more.jinja2',
version='0.3.dev0',
description="Jinja2 template integration for Morepath",
long_description=long_description,
author="Martijn Faassen",
author_email="faassen@startifact.com",
keywords='morepath jinja2',
license="BSD",
url="http://pypi.python.org/pypi/more.jinja2",
namespace_packages=['more'],
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=[
'setuptools',
'morepath >= 0.10',
'Jinja2 >= 2.7.3'
],
extras_require = dict(
test=['pytest >= 2.6.0',
'pytest-cov',
'WebTest'],
),
)
|
import io
from setuptools import setup, find_packages
long_description = '\n'.join((
io.open('README.rst', encoding='utf-8').read(),
io.open('CHANGES.txt', encoding='utf-8').read()
))
setup(name='more.jinja2',
version='0.3.dev0',
description="Jinja2 template integration for Morepath",
long_description=long_description,
author="Martijn Faassen",
author_email="faassen@startifact.com",
keywords='morepath jinja2',
license="BSD",
url="http://pypi.python.org/pypi/more.jinja2",
namespace_packages=['more'],
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=[
'setuptools',
'morepath >= 0.10',
'Jinja2 >= 2.7.3'
],
extras_require=dict(
test=['pytest >= 2.6.0',
'pytest-cov',
'WebTest'],
),
)
|
Use io.open with encoding='utf-8' and flake8 compliance
|
Use io.open with encoding='utf-8' and flake8 compliance
|
Python
|
bsd-3-clause
|
morepath/more.jinja2
|
import os, io
from setuptools import setup, find_packages
long_description = (
io.open('README.rst', encoding='utf-8').read()
+ '\n' +
io.open('CHANGES.txt', encoding='utf-8').read())
setup(name='more.jinja2',
version='0.3.dev0',
description="Jinja2 template integration for Morepath",
long_description=long_description,
author="Martijn Faassen",
author_email="faassen@startifact.com",
keywords='morepath jinja2',
license="BSD",
url="http://pypi.python.org/pypi/more.jinja2",
namespace_packages=['more'],
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=[
'setuptools',
'morepath >= 0.10',
'Jinja2 >= 2.7.3'
],
extras_require = dict(
test=['pytest >= 2.6.0',
'pytest-cov',
'WebTest'],
),
)
Use io.open with encoding='utf-8' and flake8 compliance
|
import io
from setuptools import setup, find_packages
long_description = '\n'.join((
io.open('README.rst', encoding='utf-8').read(),
io.open('CHANGES.txt', encoding='utf-8').read()
))
setup(name='more.jinja2',
version='0.3.dev0',
description="Jinja2 template integration for Morepath",
long_description=long_description,
author="Martijn Faassen",
author_email="faassen@startifact.com",
keywords='morepath jinja2',
license="BSD",
url="http://pypi.python.org/pypi/more.jinja2",
namespace_packages=['more'],
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=[
'setuptools',
'morepath >= 0.10',
'Jinja2 >= 2.7.3'
],
extras_require=dict(
test=['pytest >= 2.6.0',
'pytest-cov',
'WebTest'],
),
)
|
<commit_before>import os, io
from setuptools import setup, find_packages
long_description = (
io.open('README.rst', encoding='utf-8').read()
+ '\n' +
io.open('CHANGES.txt', encoding='utf-8').read())
setup(name='more.jinja2',
version='0.3.dev0',
description="Jinja2 template integration for Morepath",
long_description=long_description,
author="Martijn Faassen",
author_email="faassen@startifact.com",
keywords='morepath jinja2',
license="BSD",
url="http://pypi.python.org/pypi/more.jinja2",
namespace_packages=['more'],
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=[
'setuptools',
'morepath >= 0.10',
'Jinja2 >= 2.7.3'
],
extras_require = dict(
test=['pytest >= 2.6.0',
'pytest-cov',
'WebTest'],
),
)
<commit_msg>Use io.open with encoding='utf-8' and flake8 compliance<commit_after>
|
import io
from setuptools import setup, find_packages
long_description = '\n'.join((
io.open('README.rst', encoding='utf-8').read(),
io.open('CHANGES.txt', encoding='utf-8').read()
))
setup(name='more.jinja2',
version='0.3.dev0',
description="Jinja2 template integration for Morepath",
long_description=long_description,
author="Martijn Faassen",
author_email="faassen@startifact.com",
keywords='morepath jinja2',
license="BSD",
url="http://pypi.python.org/pypi/more.jinja2",
namespace_packages=['more'],
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=[
'setuptools',
'morepath >= 0.10',
'Jinja2 >= 2.7.3'
],
extras_require=dict(
test=['pytest >= 2.6.0',
'pytest-cov',
'WebTest'],
),
)
|
import os, io
from setuptools import setup, find_packages
long_description = (
io.open('README.rst', encoding='utf-8').read()
+ '\n' +
io.open('CHANGES.txt', encoding='utf-8').read())
setup(name='more.jinja2',
version='0.3.dev0',
description="Jinja2 template integration for Morepath",
long_description=long_description,
author="Martijn Faassen",
author_email="faassen@startifact.com",
keywords='morepath jinja2',
license="BSD",
url="http://pypi.python.org/pypi/more.jinja2",
namespace_packages=['more'],
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=[
'setuptools',
'morepath >= 0.10',
'Jinja2 >= 2.7.3'
],
extras_require = dict(
test=['pytest >= 2.6.0',
'pytest-cov',
'WebTest'],
),
)
Use io.open with encoding='utf-8' and flake8 complianceimport io
from setuptools import setup, find_packages
long_description = '\n'.join((
io.open('README.rst', encoding='utf-8').read(),
io.open('CHANGES.txt', encoding='utf-8').read()
))
setup(name='more.jinja2',
version='0.3.dev0',
description="Jinja2 template integration for Morepath",
long_description=long_description,
author="Martijn Faassen",
author_email="faassen@startifact.com",
keywords='morepath jinja2',
license="BSD",
url="http://pypi.python.org/pypi/more.jinja2",
namespace_packages=['more'],
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=[
'setuptools',
'morepath >= 0.10',
'Jinja2 >= 2.7.3'
],
extras_require=dict(
test=['pytest >= 2.6.0',
'pytest-cov',
'WebTest'],
),
)
|
<commit_before>import os, io
from setuptools import setup, find_packages
long_description = (
io.open('README.rst', encoding='utf-8').read()
+ '\n' +
io.open('CHANGES.txt', encoding='utf-8').read())
setup(name='more.jinja2',
version='0.3.dev0',
description="Jinja2 template integration for Morepath",
long_description=long_description,
author="Martijn Faassen",
author_email="faassen@startifact.com",
keywords='morepath jinja2',
license="BSD",
url="http://pypi.python.org/pypi/more.jinja2",
namespace_packages=['more'],
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=[
'setuptools',
'morepath >= 0.10',
'Jinja2 >= 2.7.3'
],
extras_require = dict(
test=['pytest >= 2.6.0',
'pytest-cov',
'WebTest'],
),
)
<commit_msg>Use io.open with encoding='utf-8' and flake8 compliance<commit_after>import io
from setuptools import setup, find_packages
long_description = '\n'.join((
io.open('README.rst', encoding='utf-8').read(),
io.open('CHANGES.txt', encoding='utf-8').read()
))
setup(name='more.jinja2',
version='0.3.dev0',
description="Jinja2 template integration for Morepath",
long_description=long_description,
author="Martijn Faassen",
author_email="faassen@startifact.com",
keywords='morepath jinja2',
license="BSD",
url="http://pypi.python.org/pypi/more.jinja2",
namespace_packages=['more'],
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=[
'setuptools',
'morepath >= 0.10',
'Jinja2 >= 2.7.3'
],
extras_require=dict(
test=['pytest >= 2.6.0',
'pytest-cov',
'WebTest'],
),
)
|
7d475a44f2584396890360d1545abfa585dfdda8
|
setup.py
|
setup.py
|
from setuptools import setup, find_packages
setup(name='django-forge',
version=__import__('forge').__version__,
author='Justin Bronn',
author_email='jbronn@gmail.com',
description='A Django implementation of the Puppet Forge web API.',
url='https://github.com/jbronn/django-forge',
download_url='http://pypi.python.org/pypi/django-forge/',
install_requires=[
'Django>=1.4',
'requests>=2',
'semantic_version>=2.1.2',
],
packages=find_packages(),
package_data={'apache': ['forge/apache'],
'templates': ['forge/templates']},
include_package_data=True,
zip_safe=False,
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Web Environment',
'Framework :: Django',
'Intended Audience :: Developers',
'Intended Audience :: System Administrators',
'License :: OSI Approved :: Apache Software License',
],
)
|
from setuptools import setup, find_packages
setup(name='django-forge',
version=__import__('forge').__version__,
author='Justin Bronn',
author_email='jbronn@gmail.com',
description='A Django implementation of the Puppet Forge API.',
url='https://github.com/jbronn/django-forge',
download_url='http://pypi.python.org/pypi/django-forge/',
install_requires=[
'Django>=1.4',
'requests>=2',
'semantic_version>=2.1.2',
],
packages=find_packages(),
include_package_data=True,
zip_safe=False,
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Web Environment',
'Framework :: Django',
'Intended Audience :: Developers',
'Intended Audience :: System Administrators',
'License :: OSI Approved :: Apache Software License',
],
)
|
Simplify description and MANIFEST.in suits me fine.
|
Simplify description and MANIFEST.in suits me fine.
|
Python
|
apache-2.0
|
jbronn/django-forge,jbronn/django-forge,ocadotechnology/django-forge,ocadotechnology/django-forge
|
from setuptools import setup, find_packages
setup(name='django-forge',
version=__import__('forge').__version__,
author='Justin Bronn',
author_email='jbronn@gmail.com',
description='A Django implementation of the Puppet Forge web API.',
url='https://github.com/jbronn/django-forge',
download_url='http://pypi.python.org/pypi/django-forge/',
install_requires=[
'Django>=1.4',
'requests>=2',
'semantic_version>=2.1.2',
],
packages=find_packages(),
package_data={'apache': ['forge/apache'],
'templates': ['forge/templates']},
include_package_data=True,
zip_safe=False,
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Web Environment',
'Framework :: Django',
'Intended Audience :: Developers',
'Intended Audience :: System Administrators',
'License :: OSI Approved :: Apache Software License',
],
)
Simplify description and MANIFEST.in suits me fine.
|
from setuptools import setup, find_packages
setup(name='django-forge',
version=__import__('forge').__version__,
author='Justin Bronn',
author_email='jbronn@gmail.com',
description='A Django implementation of the Puppet Forge API.',
url='https://github.com/jbronn/django-forge',
download_url='http://pypi.python.org/pypi/django-forge/',
install_requires=[
'Django>=1.4',
'requests>=2',
'semantic_version>=2.1.2',
],
packages=find_packages(),
include_package_data=True,
zip_safe=False,
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Web Environment',
'Framework :: Django',
'Intended Audience :: Developers',
'Intended Audience :: System Administrators',
'License :: OSI Approved :: Apache Software License',
],
)
|
<commit_before>from setuptools import setup, find_packages
setup(name='django-forge',
version=__import__('forge').__version__,
author='Justin Bronn',
author_email='jbronn@gmail.com',
description='A Django implementation of the Puppet Forge web API.',
url='https://github.com/jbronn/django-forge',
download_url='http://pypi.python.org/pypi/django-forge/',
install_requires=[
'Django>=1.4',
'requests>=2',
'semantic_version>=2.1.2',
],
packages=find_packages(),
package_data={'apache': ['forge/apache'],
'templates': ['forge/templates']},
include_package_data=True,
zip_safe=False,
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Web Environment',
'Framework :: Django',
'Intended Audience :: Developers',
'Intended Audience :: System Administrators',
'License :: OSI Approved :: Apache Software License',
],
)
<commit_msg>Simplify description and MANIFEST.in suits me fine.<commit_after>
|
from setuptools import setup, find_packages
setup(name='django-forge',
version=__import__('forge').__version__,
author='Justin Bronn',
author_email='jbronn@gmail.com',
description='A Django implementation of the Puppet Forge API.',
url='https://github.com/jbronn/django-forge',
download_url='http://pypi.python.org/pypi/django-forge/',
install_requires=[
'Django>=1.4',
'requests>=2',
'semantic_version>=2.1.2',
],
packages=find_packages(),
include_package_data=True,
zip_safe=False,
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Web Environment',
'Framework :: Django',
'Intended Audience :: Developers',
'Intended Audience :: System Administrators',
'License :: OSI Approved :: Apache Software License',
],
)
|
from setuptools import setup, find_packages
setup(name='django-forge',
version=__import__('forge').__version__,
author='Justin Bronn',
author_email='jbronn@gmail.com',
description='A Django implementation of the Puppet Forge web API.',
url='https://github.com/jbronn/django-forge',
download_url='http://pypi.python.org/pypi/django-forge/',
install_requires=[
'Django>=1.4',
'requests>=2',
'semantic_version>=2.1.2',
],
packages=find_packages(),
package_data={'apache': ['forge/apache'],
'templates': ['forge/templates']},
include_package_data=True,
zip_safe=False,
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Web Environment',
'Framework :: Django',
'Intended Audience :: Developers',
'Intended Audience :: System Administrators',
'License :: OSI Approved :: Apache Software License',
],
)
Simplify description and MANIFEST.in suits me fine.from setuptools import setup, find_packages
setup(name='django-forge',
version=__import__('forge').__version__,
author='Justin Bronn',
author_email='jbronn@gmail.com',
description='A Django implementation of the Puppet Forge API.',
url='https://github.com/jbronn/django-forge',
download_url='http://pypi.python.org/pypi/django-forge/',
install_requires=[
'Django>=1.4',
'requests>=2',
'semantic_version>=2.1.2',
],
packages=find_packages(),
include_package_data=True,
zip_safe=False,
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Web Environment',
'Framework :: Django',
'Intended Audience :: Developers',
'Intended Audience :: System Administrators',
'License :: OSI Approved :: Apache Software License',
],
)
|
<commit_before>from setuptools import setup, find_packages
setup(name='django-forge',
version=__import__('forge').__version__,
author='Justin Bronn',
author_email='jbronn@gmail.com',
description='A Django implementation of the Puppet Forge web API.',
url='https://github.com/jbronn/django-forge',
download_url='http://pypi.python.org/pypi/django-forge/',
install_requires=[
'Django>=1.4',
'requests>=2',
'semantic_version>=2.1.2',
],
packages=find_packages(),
package_data={'apache': ['forge/apache'],
'templates': ['forge/templates']},
include_package_data=True,
zip_safe=False,
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Web Environment',
'Framework :: Django',
'Intended Audience :: Developers',
'Intended Audience :: System Administrators',
'License :: OSI Approved :: Apache Software License',
],
)
<commit_msg>Simplify description and MANIFEST.in suits me fine.<commit_after>from setuptools import setup, find_packages
setup(name='django-forge',
version=__import__('forge').__version__,
author='Justin Bronn',
author_email='jbronn@gmail.com',
description='A Django implementation of the Puppet Forge API.',
url='https://github.com/jbronn/django-forge',
download_url='http://pypi.python.org/pypi/django-forge/',
install_requires=[
'Django>=1.4',
'requests>=2',
'semantic_version>=2.1.2',
],
packages=find_packages(),
include_package_data=True,
zip_safe=False,
classifiers=[
'Development Status :: 4 - Beta',
'Environment :: Web Environment',
'Framework :: Django',
'Intended Audience :: Developers',
'Intended Audience :: System Administrators',
'License :: OSI Approved :: Apache Software License',
],
)
|
2e850a22d0fcf8441d47928f5d758e3cb6b6bbaa
|
setup.py
|
setup.py
|
#!/usr/bin/env python
from setuptools import find_packages, setup
setup(
name='Flask-uWSGI-WebSocket',
version='0.6.1',
url='https://github.com/zeekay/flask_uwsgi_websocket',
license='MIT',
author='Zach Kelling',
author_email='zk@monoid.io',
description='High-performance WebSockets for your Flask apps powered by uWSGI.',
long_description=open('README.rst').read(),
py_modules=['flask_uwsgi_websocket'],
zip_safe=False,
include_package_data=True,
packages=find_packages(),
platforms='any',
install_requires=[
'Flask',
'uwsgi',
],
classifiers=[
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
'Topic :: Software Development :: Libraries :: Python Modules'
],
keywords='uwsgi flask websockets'
)
|
#!/usr/bin/env python
from setuptools import find_packages, setup
setup(
name='Flask-uWSGI-WebSocket',
version='0.6.1',
url='https://github.com/zeekay/flask-uwsgi-websocket',
license='MIT',
author='Zach Kelling',
author_email='zk@monoid.io',
description='High-performance WebSockets for your Flask apps powered by uWSGI.',
long_description=open('README.rst').read(),
py_modules=['flask_uwsgi_websocket'],
zip_safe=False,
include_package_data=True,
packages=find_packages(),
platforms='any',
install_requires=[
'Flask',
'uwsgi',
],
classifiers=[
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
'Topic :: Software Development :: Libraries :: Python Modules'
],
keywords='uwsgi flask websockets'
)
|
Fix homepage url to point to the correct repo
|
Fix homepage url to point to the correct repo
|
Python
|
mit
|
zeekay/flask-uwsgi-websocket
|
#!/usr/bin/env python
from setuptools import find_packages, setup
setup(
name='Flask-uWSGI-WebSocket',
version='0.6.1',
url='https://github.com/zeekay/flask_uwsgi_websocket',
license='MIT',
author='Zach Kelling',
author_email='zk@monoid.io',
description='High-performance WebSockets for your Flask apps powered by uWSGI.',
long_description=open('README.rst').read(),
py_modules=['flask_uwsgi_websocket'],
zip_safe=False,
include_package_data=True,
packages=find_packages(),
platforms='any',
install_requires=[
'Flask',
'uwsgi',
],
classifiers=[
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
'Topic :: Software Development :: Libraries :: Python Modules'
],
keywords='uwsgi flask websockets'
)
Fix homepage url to point to the correct repo
|
#!/usr/bin/env python
from setuptools import find_packages, setup
setup(
name='Flask-uWSGI-WebSocket',
version='0.6.1',
url='https://github.com/zeekay/flask-uwsgi-websocket',
license='MIT',
author='Zach Kelling',
author_email='zk@monoid.io',
description='High-performance WebSockets for your Flask apps powered by uWSGI.',
long_description=open('README.rst').read(),
py_modules=['flask_uwsgi_websocket'],
zip_safe=False,
include_package_data=True,
packages=find_packages(),
platforms='any',
install_requires=[
'Flask',
'uwsgi',
],
classifiers=[
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
'Topic :: Software Development :: Libraries :: Python Modules'
],
keywords='uwsgi flask websockets'
)
|
<commit_before>#!/usr/bin/env python
from setuptools import find_packages, setup
setup(
name='Flask-uWSGI-WebSocket',
version='0.6.1',
url='https://github.com/zeekay/flask_uwsgi_websocket',
license='MIT',
author='Zach Kelling',
author_email='zk@monoid.io',
description='High-performance WebSockets for your Flask apps powered by uWSGI.',
long_description=open('README.rst').read(),
py_modules=['flask_uwsgi_websocket'],
zip_safe=False,
include_package_data=True,
packages=find_packages(),
platforms='any',
install_requires=[
'Flask',
'uwsgi',
],
classifiers=[
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
'Topic :: Software Development :: Libraries :: Python Modules'
],
keywords='uwsgi flask websockets'
)
<commit_msg>Fix homepage url to point to the correct repo<commit_after>
|
#!/usr/bin/env python
from setuptools import find_packages, setup
setup(
name='Flask-uWSGI-WebSocket',
version='0.6.1',
url='https://github.com/zeekay/flask-uwsgi-websocket',
license='MIT',
author='Zach Kelling',
author_email='zk@monoid.io',
description='High-performance WebSockets for your Flask apps powered by uWSGI.',
long_description=open('README.rst').read(),
py_modules=['flask_uwsgi_websocket'],
zip_safe=False,
include_package_data=True,
packages=find_packages(),
platforms='any',
install_requires=[
'Flask',
'uwsgi',
],
classifiers=[
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
'Topic :: Software Development :: Libraries :: Python Modules'
],
keywords='uwsgi flask websockets'
)
|
#!/usr/bin/env python
from setuptools import find_packages, setup
setup(
name='Flask-uWSGI-WebSocket',
version='0.6.1',
url='https://github.com/zeekay/flask_uwsgi_websocket',
license='MIT',
author='Zach Kelling',
author_email='zk@monoid.io',
description='High-performance WebSockets for your Flask apps powered by uWSGI.',
long_description=open('README.rst').read(),
py_modules=['flask_uwsgi_websocket'],
zip_safe=False,
include_package_data=True,
packages=find_packages(),
platforms='any',
install_requires=[
'Flask',
'uwsgi',
],
classifiers=[
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
'Topic :: Software Development :: Libraries :: Python Modules'
],
keywords='uwsgi flask websockets'
)
Fix homepage url to point to the correct repo#!/usr/bin/env python
from setuptools import find_packages, setup
setup(
name='Flask-uWSGI-WebSocket',
version='0.6.1',
url='https://github.com/zeekay/flask-uwsgi-websocket',
license='MIT',
author='Zach Kelling',
author_email='zk@monoid.io',
description='High-performance WebSockets for your Flask apps powered by uWSGI.',
long_description=open('README.rst').read(),
py_modules=['flask_uwsgi_websocket'],
zip_safe=False,
include_package_data=True,
packages=find_packages(),
platforms='any',
install_requires=[
'Flask',
'uwsgi',
],
classifiers=[
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
'Topic :: Software Development :: Libraries :: Python Modules'
],
keywords='uwsgi flask websockets'
)
|
<commit_before>#!/usr/bin/env python
from setuptools import find_packages, setup
setup(
name='Flask-uWSGI-WebSocket',
version='0.6.1',
url='https://github.com/zeekay/flask_uwsgi_websocket',
license='MIT',
author='Zach Kelling',
author_email='zk@monoid.io',
description='High-performance WebSockets for your Flask apps powered by uWSGI.',
long_description=open('README.rst').read(),
py_modules=['flask_uwsgi_websocket'],
zip_safe=False,
include_package_data=True,
packages=find_packages(),
platforms='any',
install_requires=[
'Flask',
'uwsgi',
],
classifiers=[
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
'Topic :: Software Development :: Libraries :: Python Modules'
],
keywords='uwsgi flask websockets'
)
<commit_msg>Fix homepage url to point to the correct repo<commit_after>#!/usr/bin/env python
from setuptools import find_packages, setup
setup(
name='Flask-uWSGI-WebSocket',
version='0.6.1',
url='https://github.com/zeekay/flask-uwsgi-websocket',
license='MIT',
author='Zach Kelling',
author_email='zk@monoid.io',
description='High-performance WebSockets for your Flask apps powered by uWSGI.',
long_description=open('README.rst').read(),
py_modules=['flask_uwsgi_websocket'],
zip_safe=False,
include_package_data=True,
packages=find_packages(),
platforms='any',
install_requires=[
'Flask',
'uwsgi',
],
classifiers=[
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
'Topic :: Software Development :: Libraries :: Python Modules'
],
keywords='uwsgi flask websockets'
)
|
fbb0708aebf437de8a5d2e8faf6334fc46d89b45
|
setup.py
|
setup.py
|
from setuptools import setup
from rohrpost import __version__
def read(filepath):
with open(filepath, "r", encoding="utf-8") as f:
return f.read()
setup(
name="rohrpost",
version=__version__,
description="rohrpost WebSocket protocol for ASGI",
long_description=read("README.rst"),
url="https://github.com/axsemantics/rohrpost",
author="Tobias Kunze",
author_email="tobias.kunze@ax-semantics.com",
license="MIT",
python_requires=">=3.5",
install_requires=["channels>=2.0"],
classifiers=[
"Development Status :: 5 - Production/Stable",
"Environment :: Web Environment",
"Intended Audience :: Developers",
"Operating System :: OS Independent",
"License :: OSI Approved :: MIT License",
"Programming Language :: Python :: 3",
"Programming Language :: Python :: 3.5",
"Programming Language :: Python :: 3.6",
"Programming Language :: Python :: 3.7",
"Topic :: Internet :: WWW/HTTP",
],
packages=["rohrpost"],
project_urls={
"Documentation": "https://rohrpost.readthedocs.io/en/stable/",
"Source": "https://github.com/axsemantics/rohrpost",
"Tracker": "https://github.com/axsemantics/rohrpost/issues",
},
)
|
from setuptools import setup
from rohrpost import __version__
def read(filepath):
with open(filepath, "r", encoding="utf-8") as f:
return f.read()
setup(
name="rohrpost",
version=__version__,
description="rohrpost WebSocket protocol for ASGI",
long_description=read("README.rst"),
url="https://github.com/axsemantics/rohrpost",
author="Tobias Kunze",
author_email="tobias.kunze@ax-semantics.com",
license="MIT",
python_requires=">=3.5",
install_requires=["channels>=2.0"],
classifiers=[
"Development Status :: 5 - Production/Stable",
"Environment :: Web Environment",
"Intended Audience :: Developers",
"Operating System :: OS Independent",
"License :: OSI Approved :: MIT License",
"Programming Language :: Python :: 3",
"Programming Language :: Python :: 3.5",
"Programming Language :: Python :: 3.6",
"Programming Language :: Python :: 3.7",
"Programming Language :: Python :: 3 :: Only",
"Topic :: Internet :: WWW/HTTP",
],
packages=["rohrpost"],
project_urls={
"Documentation": "https://rohrpost.readthedocs.io/en/stable/",
"Source": "https://github.com/axsemantics/rohrpost",
"Tracker": "https://github.com/axsemantics/rohrpost/issues",
},
)
|
Add 'Python :: 3 :: Only' classifier
|
Add 'Python :: 3 :: Only' classifier
|
Python
|
mit
|
axsemantics/rohrpost,axsemantics/rohrpost
|
from setuptools import setup
from rohrpost import __version__
def read(filepath):
with open(filepath, "r", encoding="utf-8") as f:
return f.read()
setup(
name="rohrpost",
version=__version__,
description="rohrpost WebSocket protocol for ASGI",
long_description=read("README.rst"),
url="https://github.com/axsemantics/rohrpost",
author="Tobias Kunze",
author_email="tobias.kunze@ax-semantics.com",
license="MIT",
python_requires=">=3.5",
install_requires=["channels>=2.0"],
classifiers=[
"Development Status :: 5 - Production/Stable",
"Environment :: Web Environment",
"Intended Audience :: Developers",
"Operating System :: OS Independent",
"License :: OSI Approved :: MIT License",
"Programming Language :: Python :: 3",
"Programming Language :: Python :: 3.5",
"Programming Language :: Python :: 3.6",
"Programming Language :: Python :: 3.7",
"Topic :: Internet :: WWW/HTTP",
],
packages=["rohrpost"],
project_urls={
"Documentation": "https://rohrpost.readthedocs.io/en/stable/",
"Source": "https://github.com/axsemantics/rohrpost",
"Tracker": "https://github.com/axsemantics/rohrpost/issues",
},
)
Add 'Python :: 3 :: Only' classifier
|
from setuptools import setup
from rohrpost import __version__
def read(filepath):
with open(filepath, "r", encoding="utf-8") as f:
return f.read()
setup(
name="rohrpost",
version=__version__,
description="rohrpost WebSocket protocol for ASGI",
long_description=read("README.rst"),
url="https://github.com/axsemantics/rohrpost",
author="Tobias Kunze",
author_email="tobias.kunze@ax-semantics.com",
license="MIT",
python_requires=">=3.5",
install_requires=["channels>=2.0"],
classifiers=[
"Development Status :: 5 - Production/Stable",
"Environment :: Web Environment",
"Intended Audience :: Developers",
"Operating System :: OS Independent",
"License :: OSI Approved :: MIT License",
"Programming Language :: Python :: 3",
"Programming Language :: Python :: 3.5",
"Programming Language :: Python :: 3.6",
"Programming Language :: Python :: 3.7",
"Programming Language :: Python :: 3 :: Only",
"Topic :: Internet :: WWW/HTTP",
],
packages=["rohrpost"],
project_urls={
"Documentation": "https://rohrpost.readthedocs.io/en/stable/",
"Source": "https://github.com/axsemantics/rohrpost",
"Tracker": "https://github.com/axsemantics/rohrpost/issues",
},
)
|
<commit_before>from setuptools import setup
from rohrpost import __version__
def read(filepath):
with open(filepath, "r", encoding="utf-8") as f:
return f.read()
setup(
name="rohrpost",
version=__version__,
description="rohrpost WebSocket protocol for ASGI",
long_description=read("README.rst"),
url="https://github.com/axsemantics/rohrpost",
author="Tobias Kunze",
author_email="tobias.kunze@ax-semantics.com",
license="MIT",
python_requires=">=3.5",
install_requires=["channels>=2.0"],
classifiers=[
"Development Status :: 5 - Production/Stable",
"Environment :: Web Environment",
"Intended Audience :: Developers",
"Operating System :: OS Independent",
"License :: OSI Approved :: MIT License",
"Programming Language :: Python :: 3",
"Programming Language :: Python :: 3.5",
"Programming Language :: Python :: 3.6",
"Programming Language :: Python :: 3.7",
"Topic :: Internet :: WWW/HTTP",
],
packages=["rohrpost"],
project_urls={
"Documentation": "https://rohrpost.readthedocs.io/en/stable/",
"Source": "https://github.com/axsemantics/rohrpost",
"Tracker": "https://github.com/axsemantics/rohrpost/issues",
},
)
<commit_msg>Add 'Python :: 3 :: Only' classifier<commit_after>
|
from setuptools import setup
from rohrpost import __version__
def read(filepath):
with open(filepath, "r", encoding="utf-8") as f:
return f.read()
setup(
name="rohrpost",
version=__version__,
description="rohrpost WebSocket protocol for ASGI",
long_description=read("README.rst"),
url="https://github.com/axsemantics/rohrpost",
author="Tobias Kunze",
author_email="tobias.kunze@ax-semantics.com",
license="MIT",
python_requires=">=3.5",
install_requires=["channels>=2.0"],
classifiers=[
"Development Status :: 5 - Production/Stable",
"Environment :: Web Environment",
"Intended Audience :: Developers",
"Operating System :: OS Independent",
"License :: OSI Approved :: MIT License",
"Programming Language :: Python :: 3",
"Programming Language :: Python :: 3.5",
"Programming Language :: Python :: 3.6",
"Programming Language :: Python :: 3.7",
"Programming Language :: Python :: 3 :: Only",
"Topic :: Internet :: WWW/HTTP",
],
packages=["rohrpost"],
project_urls={
"Documentation": "https://rohrpost.readthedocs.io/en/stable/",
"Source": "https://github.com/axsemantics/rohrpost",
"Tracker": "https://github.com/axsemantics/rohrpost/issues",
},
)
|
from setuptools import setup
from rohrpost import __version__
def read(filepath):
with open(filepath, "r", encoding="utf-8") as f:
return f.read()
setup(
name="rohrpost",
version=__version__,
description="rohrpost WebSocket protocol for ASGI",
long_description=read("README.rst"),
url="https://github.com/axsemantics/rohrpost",
author="Tobias Kunze",
author_email="tobias.kunze@ax-semantics.com",
license="MIT",
python_requires=">=3.5",
install_requires=["channels>=2.0"],
classifiers=[
"Development Status :: 5 - Production/Stable",
"Environment :: Web Environment",
"Intended Audience :: Developers",
"Operating System :: OS Independent",
"License :: OSI Approved :: MIT License",
"Programming Language :: Python :: 3",
"Programming Language :: Python :: 3.5",
"Programming Language :: Python :: 3.6",
"Programming Language :: Python :: 3.7",
"Topic :: Internet :: WWW/HTTP",
],
packages=["rohrpost"],
project_urls={
"Documentation": "https://rohrpost.readthedocs.io/en/stable/",
"Source": "https://github.com/axsemantics/rohrpost",
"Tracker": "https://github.com/axsemantics/rohrpost/issues",
},
)
Add 'Python :: 3 :: Only' classifierfrom setuptools import setup
from rohrpost import __version__
def read(filepath):
with open(filepath, "r", encoding="utf-8") as f:
return f.read()
setup(
name="rohrpost",
version=__version__,
description="rohrpost WebSocket protocol for ASGI",
long_description=read("README.rst"),
url="https://github.com/axsemantics/rohrpost",
author="Tobias Kunze",
author_email="tobias.kunze@ax-semantics.com",
license="MIT",
python_requires=">=3.5",
install_requires=["channels>=2.0"],
classifiers=[
"Development Status :: 5 - Production/Stable",
"Environment :: Web Environment",
"Intended Audience :: Developers",
"Operating System :: OS Independent",
"License :: OSI Approved :: MIT License",
"Programming Language :: Python :: 3",
"Programming Language :: Python :: 3.5",
"Programming Language :: Python :: 3.6",
"Programming Language :: Python :: 3.7",
"Programming Language :: Python :: 3 :: Only",
"Topic :: Internet :: WWW/HTTP",
],
packages=["rohrpost"],
project_urls={
"Documentation": "https://rohrpost.readthedocs.io/en/stable/",
"Source": "https://github.com/axsemantics/rohrpost",
"Tracker": "https://github.com/axsemantics/rohrpost/issues",
},
)
|
<commit_before>from setuptools import setup
from rohrpost import __version__
def read(filepath):
with open(filepath, "r", encoding="utf-8") as f:
return f.read()
setup(
name="rohrpost",
version=__version__,
description="rohrpost WebSocket protocol for ASGI",
long_description=read("README.rst"),
url="https://github.com/axsemantics/rohrpost",
author="Tobias Kunze",
author_email="tobias.kunze@ax-semantics.com",
license="MIT",
python_requires=">=3.5",
install_requires=["channels>=2.0"],
classifiers=[
"Development Status :: 5 - Production/Stable",
"Environment :: Web Environment",
"Intended Audience :: Developers",
"Operating System :: OS Independent",
"License :: OSI Approved :: MIT License",
"Programming Language :: Python :: 3",
"Programming Language :: Python :: 3.5",
"Programming Language :: Python :: 3.6",
"Programming Language :: Python :: 3.7",
"Topic :: Internet :: WWW/HTTP",
],
packages=["rohrpost"],
project_urls={
"Documentation": "https://rohrpost.readthedocs.io/en/stable/",
"Source": "https://github.com/axsemantics/rohrpost",
"Tracker": "https://github.com/axsemantics/rohrpost/issues",
},
)
<commit_msg>Add 'Python :: 3 :: Only' classifier<commit_after>from setuptools import setup
from rohrpost import __version__
def read(filepath):
with open(filepath, "r", encoding="utf-8") as f:
return f.read()
setup(
name="rohrpost",
version=__version__,
description="rohrpost WebSocket protocol for ASGI",
long_description=read("README.rst"),
url="https://github.com/axsemantics/rohrpost",
author="Tobias Kunze",
author_email="tobias.kunze@ax-semantics.com",
license="MIT",
python_requires=">=3.5",
install_requires=["channels>=2.0"],
classifiers=[
"Development Status :: 5 - Production/Stable",
"Environment :: Web Environment",
"Intended Audience :: Developers",
"Operating System :: OS Independent",
"License :: OSI Approved :: MIT License",
"Programming Language :: Python :: 3",
"Programming Language :: Python :: 3.5",
"Programming Language :: Python :: 3.6",
"Programming Language :: Python :: 3.7",
"Programming Language :: Python :: 3 :: Only",
"Topic :: Internet :: WWW/HTTP",
],
packages=["rohrpost"],
project_urls={
"Documentation": "https://rohrpost.readthedocs.io/en/stable/",
"Source": "https://github.com/axsemantics/rohrpost",
"Tracker": "https://github.com/axsemantics/rohrpost/issues",
},
)
|
fab75e9c6c6e4ebfffaeb5594f206022cadd5f31
|
setup.py
|
setup.py
|
#!/usr/bin/python
# -*- coding: utf-8 -*-
from setuptools import setup
setup(
name='pymisp',
version='1.1.2',
author='RaphaΓ«l Vinot',
author_email='raphael.vinot@circl.lu',
maintainer='RaphaΓ«l Vinot',
url='https://github.com/MISP/PyMISP',
description='Python API for MISP.',
packages=['pymisp'],
classifiers=[
'License :: OSI Approved :: GNU General Public License v3 (GPLv3)',
'Development Status :: 5 - Production/Stable',
'Environment :: Console',
'Intended Audience :: Science/Research',
'Intended Audience :: Telecommunications Industry',
'Programming Language :: Python',
'Topic :: Security',
'Topic :: Internet',
],
install_requires=['requests'],
)
|
#!/usr/bin/python
# -*- coding: utf-8 -*-
from setuptools import setup
setup(
name='pymisp',
version='1.2',
author='RaphaΓ«l Vinot',
author_email='raphael.vinot@circl.lu',
maintainer='RaphaΓ«l Vinot',
url='https://github.com/MISP/PyMISP',
description='Python API for MISP.',
packages=['pymisp'],
classifiers=[
'License :: OSI Approved :: GNU General Public License v3 (GPLv3)',
'Development Status :: 5 - Production/Stable',
'Environment :: Console',
'Intended Audience :: Science/Research',
'Intended Audience :: Telecommunications Industry',
'Programming Language :: Python',
'Topic :: Security',
'Topic :: Internet',
],
install_requires=['requests'],
)
|
Increase flexibility of upload sample
|
Increase flexibility of upload sample
|
Python
|
bsd-2-clause
|
pombredanne/PyMISP,grolinet/PyMISP,iglocska/PyMISP
|
#!/usr/bin/python
# -*- coding: utf-8 -*-
from setuptools import setup
setup(
name='pymisp',
version='1.1.2',
author='RaphaΓ«l Vinot',
author_email='raphael.vinot@circl.lu',
maintainer='RaphaΓ«l Vinot',
url='https://github.com/MISP/PyMISP',
description='Python API for MISP.',
packages=['pymisp'],
classifiers=[
'License :: OSI Approved :: GNU General Public License v3 (GPLv3)',
'Development Status :: 5 - Production/Stable',
'Environment :: Console',
'Intended Audience :: Science/Research',
'Intended Audience :: Telecommunications Industry',
'Programming Language :: Python',
'Topic :: Security',
'Topic :: Internet',
],
install_requires=['requests'],
)
Increase flexibility of upload sample
|
#!/usr/bin/python
# -*- coding: utf-8 -*-
from setuptools import setup
setup(
name='pymisp',
version='1.2',
author='RaphaΓ«l Vinot',
author_email='raphael.vinot@circl.lu',
maintainer='RaphaΓ«l Vinot',
url='https://github.com/MISP/PyMISP',
description='Python API for MISP.',
packages=['pymisp'],
classifiers=[
'License :: OSI Approved :: GNU General Public License v3 (GPLv3)',
'Development Status :: 5 - Production/Stable',
'Environment :: Console',
'Intended Audience :: Science/Research',
'Intended Audience :: Telecommunications Industry',
'Programming Language :: Python',
'Topic :: Security',
'Topic :: Internet',
],
install_requires=['requests'],
)
|
<commit_before>#!/usr/bin/python
# -*- coding: utf-8 -*-
from setuptools import setup
setup(
name='pymisp',
version='1.1.2',
author='RaphaΓ«l Vinot',
author_email='raphael.vinot@circl.lu',
maintainer='RaphaΓ«l Vinot',
url='https://github.com/MISP/PyMISP',
description='Python API for MISP.',
packages=['pymisp'],
classifiers=[
'License :: OSI Approved :: GNU General Public License v3 (GPLv3)',
'Development Status :: 5 - Production/Stable',
'Environment :: Console',
'Intended Audience :: Science/Research',
'Intended Audience :: Telecommunications Industry',
'Programming Language :: Python',
'Topic :: Security',
'Topic :: Internet',
],
install_requires=['requests'],
)
<commit_msg>Increase flexibility of upload sample<commit_after>
|
#!/usr/bin/python
# -*- coding: utf-8 -*-
from setuptools import setup
setup(
name='pymisp',
version='1.2',
author='RaphaΓ«l Vinot',
author_email='raphael.vinot@circl.lu',
maintainer='RaphaΓ«l Vinot',
url='https://github.com/MISP/PyMISP',
description='Python API for MISP.',
packages=['pymisp'],
classifiers=[
'License :: OSI Approved :: GNU General Public License v3 (GPLv3)',
'Development Status :: 5 - Production/Stable',
'Environment :: Console',
'Intended Audience :: Science/Research',
'Intended Audience :: Telecommunications Industry',
'Programming Language :: Python',
'Topic :: Security',
'Topic :: Internet',
],
install_requires=['requests'],
)
|
#!/usr/bin/python
# -*- coding: utf-8 -*-
from setuptools import setup
setup(
name='pymisp',
version='1.1.2',
author='RaphaΓ«l Vinot',
author_email='raphael.vinot@circl.lu',
maintainer='RaphaΓ«l Vinot',
url='https://github.com/MISP/PyMISP',
description='Python API for MISP.',
packages=['pymisp'],
classifiers=[
'License :: OSI Approved :: GNU General Public License v3 (GPLv3)',
'Development Status :: 5 - Production/Stable',
'Environment :: Console',
'Intended Audience :: Science/Research',
'Intended Audience :: Telecommunications Industry',
'Programming Language :: Python',
'Topic :: Security',
'Topic :: Internet',
],
install_requires=['requests'],
)
Increase flexibility of upload sample#!/usr/bin/python
# -*- coding: utf-8 -*-
from setuptools import setup
setup(
name='pymisp',
version='1.2',
author='RaphaΓ«l Vinot',
author_email='raphael.vinot@circl.lu',
maintainer='RaphaΓ«l Vinot',
url='https://github.com/MISP/PyMISP',
description='Python API for MISP.',
packages=['pymisp'],
classifiers=[
'License :: OSI Approved :: GNU General Public License v3 (GPLv3)',
'Development Status :: 5 - Production/Stable',
'Environment :: Console',
'Intended Audience :: Science/Research',
'Intended Audience :: Telecommunications Industry',
'Programming Language :: Python',
'Topic :: Security',
'Topic :: Internet',
],
install_requires=['requests'],
)
|
<commit_before>#!/usr/bin/python
# -*- coding: utf-8 -*-
from setuptools import setup
setup(
name='pymisp',
version='1.1.2',
author='RaphaΓ«l Vinot',
author_email='raphael.vinot@circl.lu',
maintainer='RaphaΓ«l Vinot',
url='https://github.com/MISP/PyMISP',
description='Python API for MISP.',
packages=['pymisp'],
classifiers=[
'License :: OSI Approved :: GNU General Public License v3 (GPLv3)',
'Development Status :: 5 - Production/Stable',
'Environment :: Console',
'Intended Audience :: Science/Research',
'Intended Audience :: Telecommunications Industry',
'Programming Language :: Python',
'Topic :: Security',
'Topic :: Internet',
],
install_requires=['requests'],
)
<commit_msg>Increase flexibility of upload sample<commit_after>#!/usr/bin/python
# -*- coding: utf-8 -*-
from setuptools import setup
setup(
name='pymisp',
version='1.2',
author='RaphaΓ«l Vinot',
author_email='raphael.vinot@circl.lu',
maintainer='RaphaΓ«l Vinot',
url='https://github.com/MISP/PyMISP',
description='Python API for MISP.',
packages=['pymisp'],
classifiers=[
'License :: OSI Approved :: GNU General Public License v3 (GPLv3)',
'Development Status :: 5 - Production/Stable',
'Environment :: Console',
'Intended Audience :: Science/Research',
'Intended Audience :: Telecommunications Industry',
'Programming Language :: Python',
'Topic :: Security',
'Topic :: Internet',
],
install_requires=['requests'],
)
|
8c228a79450c49ee1d494ca1e3cf61ea7bcc8177
|
setup.py
|
setup.py
|
"""
Copyright (c) 2010-2013, Anthony Garcia <anthony@lagg.me>
Distributed under the ISC License (see LICENSE)
"""
from distutils.core import setup, Command
from distutils.errors import DistutilsOptionError
from unittest import TestLoader, TextTestRunner
import os
import steam
class run_tests(Command):
description = "Run the steamodd unit tests"
user_options = [
("key=", 'k', "Your API key")
]
def initialize_options(self):
self.key = None
def finalize_options(self):
if not self.key:
raise DistutilsOptionError("API key is required")
else:
os.environ["STEAM_API_KEY"] = self.key
def run(self):
tests = TestLoader().discover("tests")
TextTestRunner(verbosity = 2).run(tests)
setup(name = "steamodd",
version = steam.__version__,
description = "High level Steam API implementation with low level reusable core",
packages = ["steam"],
author = steam.__author__,
author_email = steam.__contact__,
url = "https://github.com/Lagg/steamodd",
classifiers = [
"License :: OSI Approved :: ISC License (ISCL)",
"Intended Audience :: Developers",
"Operating System :: OS Independent",
"Programming Language :: Python"
],
license = steam.__license__,
cmdclass = {"run_tests": run_tests})
|
"""
Copyright (c) 2010-2013, Anthony Garcia <anthony@lagg.me>
Distributed under the ISC License (see LICENSE)
"""
from distutils.core import setup, Command
from distutils.errors import DistutilsOptionError
from unittest import TestLoader, TextTestRunner
import steam
class run_tests(Command):
description = "Run the steamodd unit tests"
user_options = [
("key=", 'k', "Your API key")
]
def initialize_options(self):
self.key = None
def finalize_options(self):
if not self.key:
raise DistutilsOptionError("API key is required")
else:
steam.api.key.set(self.key)
def run(self):
tests = TestLoader().discover("tests")
TextTestRunner(verbosity = 2).run(tests)
setup(name = "steamodd",
version = steam.__version__,
description = "High level Steam API implementation with low level reusable core",
packages = ["steam"],
author = steam.__author__,
author_email = steam.__contact__,
url = "https://github.com/Lagg/steamodd",
classifiers = [
"License :: OSI Approved :: ISC License (ISCL)",
"Intended Audience :: Developers",
"Operating System :: OS Independent",
"Programming Language :: Python"
],
license = steam.__license__,
cmdclass = {"run_tests": run_tests})
|
Set API key directly in test runner
|
Set API key directly in test runner
|
Python
|
isc
|
miedzinski/steamodd,Lagg/steamodd
|
"""
Copyright (c) 2010-2013, Anthony Garcia <anthony@lagg.me>
Distributed under the ISC License (see LICENSE)
"""
from distutils.core import setup, Command
from distutils.errors import DistutilsOptionError
from unittest import TestLoader, TextTestRunner
import os
import steam
class run_tests(Command):
description = "Run the steamodd unit tests"
user_options = [
("key=", 'k', "Your API key")
]
def initialize_options(self):
self.key = None
def finalize_options(self):
if not self.key:
raise DistutilsOptionError("API key is required")
else:
os.environ["STEAM_API_KEY"] = self.key
def run(self):
tests = TestLoader().discover("tests")
TextTestRunner(verbosity = 2).run(tests)
setup(name = "steamodd",
version = steam.__version__,
description = "High level Steam API implementation with low level reusable core",
packages = ["steam"],
author = steam.__author__,
author_email = steam.__contact__,
url = "https://github.com/Lagg/steamodd",
classifiers = [
"License :: OSI Approved :: ISC License (ISCL)",
"Intended Audience :: Developers",
"Operating System :: OS Independent",
"Programming Language :: Python"
],
license = steam.__license__,
cmdclass = {"run_tests": run_tests})
Set API key directly in test runner
|
"""
Copyright (c) 2010-2013, Anthony Garcia <anthony@lagg.me>
Distributed under the ISC License (see LICENSE)
"""
from distutils.core import setup, Command
from distutils.errors import DistutilsOptionError
from unittest import TestLoader, TextTestRunner
import steam
class run_tests(Command):
description = "Run the steamodd unit tests"
user_options = [
("key=", 'k', "Your API key")
]
def initialize_options(self):
self.key = None
def finalize_options(self):
if not self.key:
raise DistutilsOptionError("API key is required")
else:
steam.api.key.set(self.key)
def run(self):
tests = TestLoader().discover("tests")
TextTestRunner(verbosity = 2).run(tests)
setup(name = "steamodd",
version = steam.__version__,
description = "High level Steam API implementation with low level reusable core",
packages = ["steam"],
author = steam.__author__,
author_email = steam.__contact__,
url = "https://github.com/Lagg/steamodd",
classifiers = [
"License :: OSI Approved :: ISC License (ISCL)",
"Intended Audience :: Developers",
"Operating System :: OS Independent",
"Programming Language :: Python"
],
license = steam.__license__,
cmdclass = {"run_tests": run_tests})
|
<commit_before>"""
Copyright (c) 2010-2013, Anthony Garcia <anthony@lagg.me>
Distributed under the ISC License (see LICENSE)
"""
from distutils.core import setup, Command
from distutils.errors import DistutilsOptionError
from unittest import TestLoader, TextTestRunner
import os
import steam
class run_tests(Command):
description = "Run the steamodd unit tests"
user_options = [
("key=", 'k', "Your API key")
]
def initialize_options(self):
self.key = None
def finalize_options(self):
if not self.key:
raise DistutilsOptionError("API key is required")
else:
os.environ["STEAM_API_KEY"] = self.key
def run(self):
tests = TestLoader().discover("tests")
TextTestRunner(verbosity = 2).run(tests)
setup(name = "steamodd",
version = steam.__version__,
description = "High level Steam API implementation with low level reusable core",
packages = ["steam"],
author = steam.__author__,
author_email = steam.__contact__,
url = "https://github.com/Lagg/steamodd",
classifiers = [
"License :: OSI Approved :: ISC License (ISCL)",
"Intended Audience :: Developers",
"Operating System :: OS Independent",
"Programming Language :: Python"
],
license = steam.__license__,
cmdclass = {"run_tests": run_tests})
<commit_msg>Set API key directly in test runner<commit_after>
|
"""
Copyright (c) 2010-2013, Anthony Garcia <anthony@lagg.me>
Distributed under the ISC License (see LICENSE)
"""
from distutils.core import setup, Command
from distutils.errors import DistutilsOptionError
from unittest import TestLoader, TextTestRunner
import steam
class run_tests(Command):
description = "Run the steamodd unit tests"
user_options = [
("key=", 'k', "Your API key")
]
def initialize_options(self):
self.key = None
def finalize_options(self):
if not self.key:
raise DistutilsOptionError("API key is required")
else:
steam.api.key.set(self.key)
def run(self):
tests = TestLoader().discover("tests")
TextTestRunner(verbosity = 2).run(tests)
setup(name = "steamodd",
version = steam.__version__,
description = "High level Steam API implementation with low level reusable core",
packages = ["steam"],
author = steam.__author__,
author_email = steam.__contact__,
url = "https://github.com/Lagg/steamodd",
classifiers = [
"License :: OSI Approved :: ISC License (ISCL)",
"Intended Audience :: Developers",
"Operating System :: OS Independent",
"Programming Language :: Python"
],
license = steam.__license__,
cmdclass = {"run_tests": run_tests})
|
"""
Copyright (c) 2010-2013, Anthony Garcia <anthony@lagg.me>
Distributed under the ISC License (see LICENSE)
"""
from distutils.core import setup, Command
from distutils.errors import DistutilsOptionError
from unittest import TestLoader, TextTestRunner
import os
import steam
class run_tests(Command):
description = "Run the steamodd unit tests"
user_options = [
("key=", 'k', "Your API key")
]
def initialize_options(self):
self.key = None
def finalize_options(self):
if not self.key:
raise DistutilsOptionError("API key is required")
else:
os.environ["STEAM_API_KEY"] = self.key
def run(self):
tests = TestLoader().discover("tests")
TextTestRunner(verbosity = 2).run(tests)
setup(name = "steamodd",
version = steam.__version__,
description = "High level Steam API implementation with low level reusable core",
packages = ["steam"],
author = steam.__author__,
author_email = steam.__contact__,
url = "https://github.com/Lagg/steamodd",
classifiers = [
"License :: OSI Approved :: ISC License (ISCL)",
"Intended Audience :: Developers",
"Operating System :: OS Independent",
"Programming Language :: Python"
],
license = steam.__license__,
cmdclass = {"run_tests": run_tests})
Set API key directly in test runner"""
Copyright (c) 2010-2013, Anthony Garcia <anthony@lagg.me>
Distributed under the ISC License (see LICENSE)
"""
from distutils.core import setup, Command
from distutils.errors import DistutilsOptionError
from unittest import TestLoader, TextTestRunner
import steam
class run_tests(Command):
description = "Run the steamodd unit tests"
user_options = [
("key=", 'k', "Your API key")
]
def initialize_options(self):
self.key = None
def finalize_options(self):
if not self.key:
raise DistutilsOptionError("API key is required")
else:
steam.api.key.set(self.key)
def run(self):
tests = TestLoader().discover("tests")
TextTestRunner(verbosity = 2).run(tests)
setup(name = "steamodd",
version = steam.__version__,
description = "High level Steam API implementation with low level reusable core",
packages = ["steam"],
author = steam.__author__,
author_email = steam.__contact__,
url = "https://github.com/Lagg/steamodd",
classifiers = [
"License :: OSI Approved :: ISC License (ISCL)",
"Intended Audience :: Developers",
"Operating System :: OS Independent",
"Programming Language :: Python"
],
license = steam.__license__,
cmdclass = {"run_tests": run_tests})
|
<commit_before>"""
Copyright (c) 2010-2013, Anthony Garcia <anthony@lagg.me>
Distributed under the ISC License (see LICENSE)
"""
from distutils.core import setup, Command
from distutils.errors import DistutilsOptionError
from unittest import TestLoader, TextTestRunner
import os
import steam
class run_tests(Command):
description = "Run the steamodd unit tests"
user_options = [
("key=", 'k', "Your API key")
]
def initialize_options(self):
self.key = None
def finalize_options(self):
if not self.key:
raise DistutilsOptionError("API key is required")
else:
os.environ["STEAM_API_KEY"] = self.key
def run(self):
tests = TestLoader().discover("tests")
TextTestRunner(verbosity = 2).run(tests)
setup(name = "steamodd",
version = steam.__version__,
description = "High level Steam API implementation with low level reusable core",
packages = ["steam"],
author = steam.__author__,
author_email = steam.__contact__,
url = "https://github.com/Lagg/steamodd",
classifiers = [
"License :: OSI Approved :: ISC License (ISCL)",
"Intended Audience :: Developers",
"Operating System :: OS Independent",
"Programming Language :: Python"
],
license = steam.__license__,
cmdclass = {"run_tests": run_tests})
<commit_msg>Set API key directly in test runner<commit_after>"""
Copyright (c) 2010-2013, Anthony Garcia <anthony@lagg.me>
Distributed under the ISC License (see LICENSE)
"""
from distutils.core import setup, Command
from distutils.errors import DistutilsOptionError
from unittest import TestLoader, TextTestRunner
import steam
class run_tests(Command):
description = "Run the steamodd unit tests"
user_options = [
("key=", 'k', "Your API key")
]
def initialize_options(self):
self.key = None
def finalize_options(self):
if not self.key:
raise DistutilsOptionError("API key is required")
else:
steam.api.key.set(self.key)
def run(self):
tests = TestLoader().discover("tests")
TextTestRunner(verbosity = 2).run(tests)
setup(name = "steamodd",
version = steam.__version__,
description = "High level Steam API implementation with low level reusable core",
packages = ["steam"],
author = steam.__author__,
author_email = steam.__contact__,
url = "https://github.com/Lagg/steamodd",
classifiers = [
"License :: OSI Approved :: ISC License (ISCL)",
"Intended Audience :: Developers",
"Operating System :: OS Independent",
"Programming Language :: Python"
],
license = steam.__license__,
cmdclass = {"run_tests": run_tests})
|
17956eb2b8089432ff2a0fcec5ce56884f904db1
|
setup.py
|
setup.py
|
#
# Copyright 2013 Cisco Systems
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from setuptools import setup, find_packages
setup(
name='django-cloudslave',
version='0.2.0',
description='Create, use and destroy temporary slaves in the cloud',
author='Soren Hansen',
author_email='sorhanse@cisco.com',
url='http://github.com/sorenh/python-django-cloudslave',
packages=find_packages(),
include_package_data=True,
license='Apache 2.0',
keywords='django openstack cloud',
install_requires=[
'django',
'python-novaclient'
],
test_suite='tests.main',
classifiers=[
'Development Status :: 2 - Pre-Alpha',
'Environment :: Web Environment',
'Framework :: Django',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Operating System :: POSIX :: Linux',
'Programming Language :: Python',
'Topic :: Software Development',
]
)
|
#
# Copyright 2013 Cisco Systems
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from setuptools import setup, find_packages
setup(
name='django-cloudslave',
version='0.2.0',
description='Create, use and destroy temporary slaves in the cloud',
author='Soren Hansen',
author_email='sorhanse@cisco.com',
url='http://github.com/sorenh/python-django-cloudslave',
packages=find_packages(),
include_package_data=True,
license='Apache 2.0',
keywords='django openstack cloud',
install_requires=[
'django',
'python-novaclient',
'south'
],
test_suite='tests.main',
classifiers=[
'Development Status :: 2 - Pre-Alpha',
'Environment :: Web Environment',
'Framework :: Django',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Operating System :: POSIX :: Linux',
'Programming Language :: Python',
'Topic :: Software Development',
]
)
|
Add south as a dependency.
|
Add south as a dependency.
|
Python
|
apache-2.0
|
sorenh/python-django-cloudslave
|
#
# Copyright 2013 Cisco Systems
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from setuptools import setup, find_packages
setup(
name='django-cloudslave',
version='0.2.0',
description='Create, use and destroy temporary slaves in the cloud',
author='Soren Hansen',
author_email='sorhanse@cisco.com',
url='http://github.com/sorenh/python-django-cloudslave',
packages=find_packages(),
include_package_data=True,
license='Apache 2.0',
keywords='django openstack cloud',
install_requires=[
'django',
'python-novaclient'
],
test_suite='tests.main',
classifiers=[
'Development Status :: 2 - Pre-Alpha',
'Environment :: Web Environment',
'Framework :: Django',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Operating System :: POSIX :: Linux',
'Programming Language :: Python',
'Topic :: Software Development',
]
)
Add south as a dependency.
|
#
# Copyright 2013 Cisco Systems
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from setuptools import setup, find_packages
setup(
name='django-cloudslave',
version='0.2.0',
description='Create, use and destroy temporary slaves in the cloud',
author='Soren Hansen',
author_email='sorhanse@cisco.com',
url='http://github.com/sorenh/python-django-cloudslave',
packages=find_packages(),
include_package_data=True,
license='Apache 2.0',
keywords='django openstack cloud',
install_requires=[
'django',
'python-novaclient',
'south'
],
test_suite='tests.main',
classifiers=[
'Development Status :: 2 - Pre-Alpha',
'Environment :: Web Environment',
'Framework :: Django',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Operating System :: POSIX :: Linux',
'Programming Language :: Python',
'Topic :: Software Development',
]
)
|
<commit_before>#
# Copyright 2013 Cisco Systems
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from setuptools import setup, find_packages
setup(
name='django-cloudslave',
version='0.2.0',
description='Create, use and destroy temporary slaves in the cloud',
author='Soren Hansen',
author_email='sorhanse@cisco.com',
url='http://github.com/sorenh/python-django-cloudslave',
packages=find_packages(),
include_package_data=True,
license='Apache 2.0',
keywords='django openstack cloud',
install_requires=[
'django',
'python-novaclient'
],
test_suite='tests.main',
classifiers=[
'Development Status :: 2 - Pre-Alpha',
'Environment :: Web Environment',
'Framework :: Django',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Operating System :: POSIX :: Linux',
'Programming Language :: Python',
'Topic :: Software Development',
]
)
<commit_msg>Add south as a dependency.<commit_after>
|
#
# Copyright 2013 Cisco Systems
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from setuptools import setup, find_packages
setup(
name='django-cloudslave',
version='0.2.0',
description='Create, use and destroy temporary slaves in the cloud',
author='Soren Hansen',
author_email='sorhanse@cisco.com',
url='http://github.com/sorenh/python-django-cloudslave',
packages=find_packages(),
include_package_data=True,
license='Apache 2.0',
keywords='django openstack cloud',
install_requires=[
'django',
'python-novaclient',
'south'
],
test_suite='tests.main',
classifiers=[
'Development Status :: 2 - Pre-Alpha',
'Environment :: Web Environment',
'Framework :: Django',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Operating System :: POSIX :: Linux',
'Programming Language :: Python',
'Topic :: Software Development',
]
)
|
#
# Copyright 2013 Cisco Systems
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from setuptools import setup, find_packages
setup(
name='django-cloudslave',
version='0.2.0',
description='Create, use and destroy temporary slaves in the cloud',
author='Soren Hansen',
author_email='sorhanse@cisco.com',
url='http://github.com/sorenh/python-django-cloudslave',
packages=find_packages(),
include_package_data=True,
license='Apache 2.0',
keywords='django openstack cloud',
install_requires=[
'django',
'python-novaclient'
],
test_suite='tests.main',
classifiers=[
'Development Status :: 2 - Pre-Alpha',
'Environment :: Web Environment',
'Framework :: Django',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Operating System :: POSIX :: Linux',
'Programming Language :: Python',
'Topic :: Software Development',
]
)
Add south as a dependency.#
# Copyright 2013 Cisco Systems
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from setuptools import setup, find_packages
setup(
name='django-cloudslave',
version='0.2.0',
description='Create, use and destroy temporary slaves in the cloud',
author='Soren Hansen',
author_email='sorhanse@cisco.com',
url='http://github.com/sorenh/python-django-cloudslave',
packages=find_packages(),
include_package_data=True,
license='Apache 2.0',
keywords='django openstack cloud',
install_requires=[
'django',
'python-novaclient',
'south'
],
test_suite='tests.main',
classifiers=[
'Development Status :: 2 - Pre-Alpha',
'Environment :: Web Environment',
'Framework :: Django',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Operating System :: POSIX :: Linux',
'Programming Language :: Python',
'Topic :: Software Development',
]
)
|
<commit_before>#
# Copyright 2013 Cisco Systems
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from setuptools import setup, find_packages
setup(
name='django-cloudslave',
version='0.2.0',
description='Create, use and destroy temporary slaves in the cloud',
author='Soren Hansen',
author_email='sorhanse@cisco.com',
url='http://github.com/sorenh/python-django-cloudslave',
packages=find_packages(),
include_package_data=True,
license='Apache 2.0',
keywords='django openstack cloud',
install_requires=[
'django',
'python-novaclient'
],
test_suite='tests.main',
classifiers=[
'Development Status :: 2 - Pre-Alpha',
'Environment :: Web Environment',
'Framework :: Django',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Operating System :: POSIX :: Linux',
'Programming Language :: Python',
'Topic :: Software Development',
]
)
<commit_msg>Add south as a dependency.<commit_after>#
# Copyright 2013 Cisco Systems
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# http://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
from setuptools import setup, find_packages
setup(
name='django-cloudslave',
version='0.2.0',
description='Create, use and destroy temporary slaves in the cloud',
author='Soren Hansen',
author_email='sorhanse@cisco.com',
url='http://github.com/sorenh/python-django-cloudslave',
packages=find_packages(),
include_package_data=True,
license='Apache 2.0',
keywords='django openstack cloud',
install_requires=[
'django',
'python-novaclient',
'south'
],
test_suite='tests.main',
classifiers=[
'Development Status :: 2 - Pre-Alpha',
'Environment :: Web Environment',
'Framework :: Django',
'Intended Audience :: Developers',
'License :: OSI Approved :: Apache Software License',
'Operating System :: POSIX :: Linux',
'Programming Language :: Python',
'Topic :: Software Development',
]
)
|
5f58d9f73b3b674d1d39ea7027a2e6de6dc8ff44
|
setup.py
|
setup.py
|
# -*- coding: utf-8 -*-
from distutils.core import setup
setup(
name='nutshell',
packages=['nutshell'],
version='0.1.1',
description='A minimal python library to access Nutshell CRM:s JSON-RPC API.',
author=u'Emil StenstrΓΆm',
author_email='em@kth.se',
url='https://github.com/EmilStenstrom/python-nutshell',
download_url='https://github.com/EmilStenstrom/python-nutshell/tarball/0.1.1',
keywords=['nutshell', 'nutshell-crm', 'json-rpc'],
classifiers=[],
)
|
# -*- coding: utf-8 -*-
from distutils.core import setup
setup(
name='nutshell',
packages=['nutshell'],
version='0.1.1',
description='A minimal python library to access Nutshell CRM:s JSON-RPC API.',
author=u'Emil StenstrΓΆm',
author_email='em@kth.se',
url='https://github.com/EmilStenstrom/python-nutshell',
download_url='https://github.com/EmilStenstrom/python-nutshell/tarball/0.1.1',
install_requires=["requests"],
keywords=['nutshell', 'nutshell-crm', 'json-rpc'],
classifiers=[],
)
|
Add requests as installation dependency.
|
Add requests as installation dependency.
|
Python
|
mit
|
EmilStenstrom/python-nutshell
|
# -*- coding: utf-8 -*-
from distutils.core import setup
setup(
name='nutshell',
packages=['nutshell'],
version='0.1.1',
description='A minimal python library to access Nutshell CRM:s JSON-RPC API.',
author=u'Emil StenstrΓΆm',
author_email='em@kth.se',
url='https://github.com/EmilStenstrom/python-nutshell',
download_url='https://github.com/EmilStenstrom/python-nutshell/tarball/0.1.1',
keywords=['nutshell', 'nutshell-crm', 'json-rpc'],
classifiers=[],
)
Add requests as installation dependency.
|
# -*- coding: utf-8 -*-
from distutils.core import setup
setup(
name='nutshell',
packages=['nutshell'],
version='0.1.1',
description='A minimal python library to access Nutshell CRM:s JSON-RPC API.',
author=u'Emil StenstrΓΆm',
author_email='em@kth.se',
url='https://github.com/EmilStenstrom/python-nutshell',
download_url='https://github.com/EmilStenstrom/python-nutshell/tarball/0.1.1',
install_requires=["requests"],
keywords=['nutshell', 'nutshell-crm', 'json-rpc'],
classifiers=[],
)
|
<commit_before># -*- coding: utf-8 -*-
from distutils.core import setup
setup(
name='nutshell',
packages=['nutshell'],
version='0.1.1',
description='A minimal python library to access Nutshell CRM:s JSON-RPC API.',
author=u'Emil StenstrΓΆm',
author_email='em@kth.se',
url='https://github.com/EmilStenstrom/python-nutshell',
download_url='https://github.com/EmilStenstrom/python-nutshell/tarball/0.1.1',
keywords=['nutshell', 'nutshell-crm', 'json-rpc'],
classifiers=[],
)
<commit_msg>Add requests as installation dependency.<commit_after>
|
# -*- coding: utf-8 -*-
from distutils.core import setup
setup(
name='nutshell',
packages=['nutshell'],
version='0.1.1',
description='A minimal python library to access Nutshell CRM:s JSON-RPC API.',
author=u'Emil StenstrΓΆm',
author_email='em@kth.se',
url='https://github.com/EmilStenstrom/python-nutshell',
download_url='https://github.com/EmilStenstrom/python-nutshell/tarball/0.1.1',
install_requires=["requests"],
keywords=['nutshell', 'nutshell-crm', 'json-rpc'],
classifiers=[],
)
|
# -*- coding: utf-8 -*-
from distutils.core import setup
setup(
name='nutshell',
packages=['nutshell'],
version='0.1.1',
description='A minimal python library to access Nutshell CRM:s JSON-RPC API.',
author=u'Emil StenstrΓΆm',
author_email='em@kth.se',
url='https://github.com/EmilStenstrom/python-nutshell',
download_url='https://github.com/EmilStenstrom/python-nutshell/tarball/0.1.1',
keywords=['nutshell', 'nutshell-crm', 'json-rpc'],
classifiers=[],
)
Add requests as installation dependency.# -*- coding: utf-8 -*-
from distutils.core import setup
setup(
name='nutshell',
packages=['nutshell'],
version='0.1.1',
description='A minimal python library to access Nutshell CRM:s JSON-RPC API.',
author=u'Emil StenstrΓΆm',
author_email='em@kth.se',
url='https://github.com/EmilStenstrom/python-nutshell',
download_url='https://github.com/EmilStenstrom/python-nutshell/tarball/0.1.1',
install_requires=["requests"],
keywords=['nutshell', 'nutshell-crm', 'json-rpc'],
classifiers=[],
)
|
<commit_before># -*- coding: utf-8 -*-
from distutils.core import setup
setup(
name='nutshell',
packages=['nutshell'],
version='0.1.1',
description='A minimal python library to access Nutshell CRM:s JSON-RPC API.',
author=u'Emil StenstrΓΆm',
author_email='em@kth.se',
url='https://github.com/EmilStenstrom/python-nutshell',
download_url='https://github.com/EmilStenstrom/python-nutshell/tarball/0.1.1',
keywords=['nutshell', 'nutshell-crm', 'json-rpc'],
classifiers=[],
)
<commit_msg>Add requests as installation dependency.<commit_after># -*- coding: utf-8 -*-
from distutils.core import setup
setup(
name='nutshell',
packages=['nutshell'],
version='0.1.1',
description='A minimal python library to access Nutshell CRM:s JSON-RPC API.',
author=u'Emil StenstrΓΆm',
author_email='em@kth.se',
url='https://github.com/EmilStenstrom/python-nutshell',
download_url='https://github.com/EmilStenstrom/python-nutshell/tarball/0.1.1',
install_requires=["requests"],
keywords=['nutshell', 'nutshell-crm', 'json-rpc'],
classifiers=[],
)
|
c04e740db80792076c5a7299da8e552dfa3603bf
|
setup.py
|
setup.py
|
import os
from setuptools import setup, find_packages
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
setup(
name = 'django-helpline-faq',
version = '1.1',
description = 'A simple FAQ application for Django sites.',
long_description = read('README.rst'),
license = "BSD",
author ='Kevin Fricovsky',
author_email = 'kfricovsky@gmail.com',
url = 'http://django-fack.rtfd.org/',
packages = find_packages(exclude=['example']),
zip_safe = False,
classifiers = [
'Development Status :: 3 - Alpha',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Framework :: Django',
],
install_requires = ['Django >= 1.3'],
test_suite = "faq._testrunner.runtests",
tests_require = ["mock"],
)
|
import os
from setuptools import setup, find_packages
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
setup(
name = 'django-helpline-faq',
version = '1.1',
description = 'A simple FAQ application for Django sites.',
long_description = read('README.rst'),
license = "BSD",
author ='Kevin Fricovsky',
author_email = 'kfricovsky@gmail.com',
url = 'http://django-fack.rtfd.org/',
packages = find_packages(exclude=['example']),
include_package_data=True,
zip_safe = False,
classifiers = [
'Development Status :: 3 - Alpha',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Framework :: Django',
],
install_requires = ['Django >= 1.3'],
test_suite = "faq._testrunner.runtests",
tests_require = ["mock"],
)
|
Include Package Data was missing.
|
Include Package Data was missing.
|
Python
|
bsd-3-clause
|
Kirembu/django-helpline-faq,Kirembu/django-helpline-faq
|
import os
from setuptools import setup, find_packages
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
setup(
name = 'django-helpline-faq',
version = '1.1',
description = 'A simple FAQ application for Django sites.',
long_description = read('README.rst'),
license = "BSD",
author ='Kevin Fricovsky',
author_email = 'kfricovsky@gmail.com',
url = 'http://django-fack.rtfd.org/',
packages = find_packages(exclude=['example']),
zip_safe = False,
classifiers = [
'Development Status :: 3 - Alpha',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Framework :: Django',
],
install_requires = ['Django >= 1.3'],
test_suite = "faq._testrunner.runtests",
tests_require = ["mock"],
)
Include Package Data was missing.
|
import os
from setuptools import setup, find_packages
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
setup(
name = 'django-helpline-faq',
version = '1.1',
description = 'A simple FAQ application for Django sites.',
long_description = read('README.rst'),
license = "BSD",
author ='Kevin Fricovsky',
author_email = 'kfricovsky@gmail.com',
url = 'http://django-fack.rtfd.org/',
packages = find_packages(exclude=['example']),
include_package_data=True,
zip_safe = False,
classifiers = [
'Development Status :: 3 - Alpha',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Framework :: Django',
],
install_requires = ['Django >= 1.3'],
test_suite = "faq._testrunner.runtests",
tests_require = ["mock"],
)
|
<commit_before>import os
from setuptools import setup, find_packages
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
setup(
name = 'django-helpline-faq',
version = '1.1',
description = 'A simple FAQ application for Django sites.',
long_description = read('README.rst'),
license = "BSD",
author ='Kevin Fricovsky',
author_email = 'kfricovsky@gmail.com',
url = 'http://django-fack.rtfd.org/',
packages = find_packages(exclude=['example']),
zip_safe = False,
classifiers = [
'Development Status :: 3 - Alpha',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Framework :: Django',
],
install_requires = ['Django >= 1.3'],
test_suite = "faq._testrunner.runtests",
tests_require = ["mock"],
)
<commit_msg>Include Package Data was missing.<commit_after>
|
import os
from setuptools import setup, find_packages
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
setup(
name = 'django-helpline-faq',
version = '1.1',
description = 'A simple FAQ application for Django sites.',
long_description = read('README.rst'),
license = "BSD",
author ='Kevin Fricovsky',
author_email = 'kfricovsky@gmail.com',
url = 'http://django-fack.rtfd.org/',
packages = find_packages(exclude=['example']),
include_package_data=True,
zip_safe = False,
classifiers = [
'Development Status :: 3 - Alpha',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Framework :: Django',
],
install_requires = ['Django >= 1.3'],
test_suite = "faq._testrunner.runtests",
tests_require = ["mock"],
)
|
import os
from setuptools import setup, find_packages
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
setup(
name = 'django-helpline-faq',
version = '1.1',
description = 'A simple FAQ application for Django sites.',
long_description = read('README.rst'),
license = "BSD",
author ='Kevin Fricovsky',
author_email = 'kfricovsky@gmail.com',
url = 'http://django-fack.rtfd.org/',
packages = find_packages(exclude=['example']),
zip_safe = False,
classifiers = [
'Development Status :: 3 - Alpha',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Framework :: Django',
],
install_requires = ['Django >= 1.3'],
test_suite = "faq._testrunner.runtests",
tests_require = ["mock"],
)
Include Package Data was missing.import os
from setuptools import setup, find_packages
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
setup(
name = 'django-helpline-faq',
version = '1.1',
description = 'A simple FAQ application for Django sites.',
long_description = read('README.rst'),
license = "BSD",
author ='Kevin Fricovsky',
author_email = 'kfricovsky@gmail.com',
url = 'http://django-fack.rtfd.org/',
packages = find_packages(exclude=['example']),
include_package_data=True,
zip_safe = False,
classifiers = [
'Development Status :: 3 - Alpha',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Framework :: Django',
],
install_requires = ['Django >= 1.3'],
test_suite = "faq._testrunner.runtests",
tests_require = ["mock"],
)
|
<commit_before>import os
from setuptools import setup, find_packages
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
setup(
name = 'django-helpline-faq',
version = '1.1',
description = 'A simple FAQ application for Django sites.',
long_description = read('README.rst'),
license = "BSD",
author ='Kevin Fricovsky',
author_email = 'kfricovsky@gmail.com',
url = 'http://django-fack.rtfd.org/',
packages = find_packages(exclude=['example']),
zip_safe = False,
classifiers = [
'Development Status :: 3 - Alpha',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Framework :: Django',
],
install_requires = ['Django >= 1.3'],
test_suite = "faq._testrunner.runtests",
tests_require = ["mock"],
)
<commit_msg>Include Package Data was missing.<commit_after>import os
from setuptools import setup, find_packages
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
setup(
name = 'django-helpline-faq',
version = '1.1',
description = 'A simple FAQ application for Django sites.',
long_description = read('README.rst'),
license = "BSD",
author ='Kevin Fricovsky',
author_email = 'kfricovsky@gmail.com',
url = 'http://django-fack.rtfd.org/',
packages = find_packages(exclude=['example']),
include_package_data=True,
zip_safe = False,
classifiers = [
'Development Status :: 3 - Alpha',
'Environment :: Web Environment',
'Intended Audience :: Developers',
'License :: OSI Approved :: BSD License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Framework :: Django',
],
install_requires = ['Django >= 1.3'],
test_suite = "faq._testrunner.runtests",
tests_require = ["mock"],
)
|
67a0f6c0aa8015f5dea7dcc8c7bc6cae809016f5
|
setup.py
|
setup.py
|
import os
from setuptools import setup
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
setup(name='windpowerlib',
version='0.1.2dev',
description='Creating time series of wind power plants.',
url='http://github.com/wind-python/windpowerlib',
author='oemof developer group',
author_email='windpowerlib@rl-institut.de',
license=None,
packages=['windpowerlib'],
package_data={
'windpowerlib': [os.path.join('data', '*.csv')]},
long_description=read('README.rst'),
zip_safe=False,
install_requires=['pandas >= 0.19.1',
'requests'],
extras_require={
'dev': ['pytest', 'jupyter', 'sphinx_rtd_theme', 'nbformat']})
|
import os
from setuptools import setup
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
setup(name='windpowerlib',
version='0.1.2dev',
description='Creating time series of wind power plants.',
url='http://github.com/wind-python/windpowerlib',
author='oemof developer group',
author_email='windpowerlib@rl-institut.de',
license=None,
packages=['windpowerlib'],
package_data={
'windpowerlib': [os.path.join('data', '*.csv')]},
long_description=read('README.rst'),
zip_safe=False,
install_requires=['pandas >= 0.19.1, < 0.25.0',
'requests'],
extras_require={
'dev': ['pytest', 'jupyter', 'sphinx_rtd_theme', 'nbformat']})
|
Add upper limit of pandas
|
Add upper limit of pandas
|
Python
|
mit
|
wind-python/windpowerlib
|
import os
from setuptools import setup
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
setup(name='windpowerlib',
version='0.1.2dev',
description='Creating time series of wind power plants.',
url='http://github.com/wind-python/windpowerlib',
author='oemof developer group',
author_email='windpowerlib@rl-institut.de',
license=None,
packages=['windpowerlib'],
package_data={
'windpowerlib': [os.path.join('data', '*.csv')]},
long_description=read('README.rst'),
zip_safe=False,
install_requires=['pandas >= 0.19.1',
'requests'],
extras_require={
'dev': ['pytest', 'jupyter', 'sphinx_rtd_theme', 'nbformat']})
Add upper limit of pandas
|
import os
from setuptools import setup
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
setup(name='windpowerlib',
version='0.1.2dev',
description='Creating time series of wind power plants.',
url='http://github.com/wind-python/windpowerlib',
author='oemof developer group',
author_email='windpowerlib@rl-institut.de',
license=None,
packages=['windpowerlib'],
package_data={
'windpowerlib': [os.path.join('data', '*.csv')]},
long_description=read('README.rst'),
zip_safe=False,
install_requires=['pandas >= 0.19.1, < 0.25.0',
'requests'],
extras_require={
'dev': ['pytest', 'jupyter', 'sphinx_rtd_theme', 'nbformat']})
|
<commit_before>import os
from setuptools import setup
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
setup(name='windpowerlib',
version='0.1.2dev',
description='Creating time series of wind power plants.',
url='http://github.com/wind-python/windpowerlib',
author='oemof developer group',
author_email='windpowerlib@rl-institut.de',
license=None,
packages=['windpowerlib'],
package_data={
'windpowerlib': [os.path.join('data', '*.csv')]},
long_description=read('README.rst'),
zip_safe=False,
install_requires=['pandas >= 0.19.1',
'requests'],
extras_require={
'dev': ['pytest', 'jupyter', 'sphinx_rtd_theme', 'nbformat']})
<commit_msg>Add upper limit of pandas<commit_after>
|
import os
from setuptools import setup
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
setup(name='windpowerlib',
version='0.1.2dev',
description='Creating time series of wind power plants.',
url='http://github.com/wind-python/windpowerlib',
author='oemof developer group',
author_email='windpowerlib@rl-institut.de',
license=None,
packages=['windpowerlib'],
package_data={
'windpowerlib': [os.path.join('data', '*.csv')]},
long_description=read('README.rst'),
zip_safe=False,
install_requires=['pandas >= 0.19.1, < 0.25.0',
'requests'],
extras_require={
'dev': ['pytest', 'jupyter', 'sphinx_rtd_theme', 'nbformat']})
|
import os
from setuptools import setup
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
setup(name='windpowerlib',
version='0.1.2dev',
description='Creating time series of wind power plants.',
url='http://github.com/wind-python/windpowerlib',
author='oemof developer group',
author_email='windpowerlib@rl-institut.de',
license=None,
packages=['windpowerlib'],
package_data={
'windpowerlib': [os.path.join('data', '*.csv')]},
long_description=read('README.rst'),
zip_safe=False,
install_requires=['pandas >= 0.19.1',
'requests'],
extras_require={
'dev': ['pytest', 'jupyter', 'sphinx_rtd_theme', 'nbformat']})
Add upper limit of pandasimport os
from setuptools import setup
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
setup(name='windpowerlib',
version='0.1.2dev',
description='Creating time series of wind power plants.',
url='http://github.com/wind-python/windpowerlib',
author='oemof developer group',
author_email='windpowerlib@rl-institut.de',
license=None,
packages=['windpowerlib'],
package_data={
'windpowerlib': [os.path.join('data', '*.csv')]},
long_description=read('README.rst'),
zip_safe=False,
install_requires=['pandas >= 0.19.1, < 0.25.0',
'requests'],
extras_require={
'dev': ['pytest', 'jupyter', 'sphinx_rtd_theme', 'nbformat']})
|
<commit_before>import os
from setuptools import setup
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
setup(name='windpowerlib',
version='0.1.2dev',
description='Creating time series of wind power plants.',
url='http://github.com/wind-python/windpowerlib',
author='oemof developer group',
author_email='windpowerlib@rl-institut.de',
license=None,
packages=['windpowerlib'],
package_data={
'windpowerlib': [os.path.join('data', '*.csv')]},
long_description=read('README.rst'),
zip_safe=False,
install_requires=['pandas >= 0.19.1',
'requests'],
extras_require={
'dev': ['pytest', 'jupyter', 'sphinx_rtd_theme', 'nbformat']})
<commit_msg>Add upper limit of pandas<commit_after>import os
from setuptools import setup
def read(fname):
return open(os.path.join(os.path.dirname(__file__), fname)).read()
setup(name='windpowerlib',
version='0.1.2dev',
description='Creating time series of wind power plants.',
url='http://github.com/wind-python/windpowerlib',
author='oemof developer group',
author_email='windpowerlib@rl-institut.de',
license=None,
packages=['windpowerlib'],
package_data={
'windpowerlib': [os.path.join('data', '*.csv')]},
long_description=read('README.rst'),
zip_safe=False,
install_requires=['pandas >= 0.19.1, < 0.25.0',
'requests'],
extras_require={
'dev': ['pytest', 'jupyter', 'sphinx_rtd_theme', 'nbformat']})
|
199c421742c054129bc07ecfdf8b255482c4ad28
|
setup.py
|
setup.py
|
from setuptools import setup, find_packages
import sys, os
version = '0.1.0'
description = "CKAN extension for creating/distributing budget data packages"
with open('README.rst') as readme:
long_description = readme.read()
setup(
name='ckanext-budgets',
version=version,
description=description,
long_description=long_description,
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Other Audience',
'License :: OSI Approved :: GNU Affero General Public License v3',
'Programming Language :: Python :: 2.7',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
],
keywords='ckan',
author='Tryggvi Bjorgvinsson',
author_email='tryggvi.bjorgvinsson@okfn.org',
url='https://github.com/tryggvib/ckanext-budgets',
license='AGPLv3',
packages=find_packages(exclude=['ez_setup', 'examples', 'tests']),
namespace_packages=['ckanext', 'ckanext.budgets'],
include_package_data=True,
zip_safe=False,
install_requires=[
# -*- Extra requirements: -*-
'unicodecsv',
'requests',
'six',
'py-dateutil',
],
entry_points='''
[ckan.plugins]
budgets=ckanext.budgets.plugin:BudgetDataPackagePlugin
''',
)
|
from setuptools import setup, find_packages
import sys, os
version = '0.1.1'
description = "CKAN extension for creating/distributing budget data packages"
with open('README.rst') as readme:
long_description = readme.read()
setup(
name='ckanext-budgets',
version=version,
description=description,
long_description=long_description,
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Other Audience',
'License :: OSI Approved :: GNU Affero General Public License v3',
'Programming Language :: Python :: 2.7',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
],
keywords='ckan',
author='Tryggvi Bjorgvinsson',
author_email='tryggvi.bjorgvinsson@okfn.org',
url='https://github.com/tryggvib/ckanext-budgets',
license='AGPLv3',
packages=find_packages(exclude=['ez_setup', 'examples', 'tests']),
namespace_packages=['ckanext', 'ckanext.budgets'],
include_package_data=True,
zip_safe=False,
install_requires=[
# -*- Extra requirements: -*-
'unicodecsv',
'requests',
'six',
'py-dateutil',
'budgetdatapackage'
],
entry_points='''
[ckan.plugins]
budgets=ckanext.budgets.plugin:BudgetDataPackagePlugin
''',
)
|
Update dependencies and bump patch version
|
Update dependencies and bump patch version
budgetdatapackage was missing from the dependencies and has now
been added.
|
Python
|
agpl-3.0
|
openspending/ckanext-budgets,openspending/ckanext-budgets
|
from setuptools import setup, find_packages
import sys, os
version = '0.1.0'
description = "CKAN extension for creating/distributing budget data packages"
with open('README.rst') as readme:
long_description = readme.read()
setup(
name='ckanext-budgets',
version=version,
description=description,
long_description=long_description,
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Other Audience',
'License :: OSI Approved :: GNU Affero General Public License v3',
'Programming Language :: Python :: 2.7',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
],
keywords='ckan',
author='Tryggvi Bjorgvinsson',
author_email='tryggvi.bjorgvinsson@okfn.org',
url='https://github.com/tryggvib/ckanext-budgets',
license='AGPLv3',
packages=find_packages(exclude=['ez_setup', 'examples', 'tests']),
namespace_packages=['ckanext', 'ckanext.budgets'],
include_package_data=True,
zip_safe=False,
install_requires=[
# -*- Extra requirements: -*-
'unicodecsv',
'requests',
'six',
'py-dateutil',
],
entry_points='''
[ckan.plugins]
budgets=ckanext.budgets.plugin:BudgetDataPackagePlugin
''',
)
Update dependencies and bump patch version
budgetdatapackage was missing from the dependencies and has now
been added.
|
from setuptools import setup, find_packages
import sys, os
version = '0.1.1'
description = "CKAN extension for creating/distributing budget data packages"
with open('README.rst') as readme:
long_description = readme.read()
setup(
name='ckanext-budgets',
version=version,
description=description,
long_description=long_description,
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Other Audience',
'License :: OSI Approved :: GNU Affero General Public License v3',
'Programming Language :: Python :: 2.7',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
],
keywords='ckan',
author='Tryggvi Bjorgvinsson',
author_email='tryggvi.bjorgvinsson@okfn.org',
url='https://github.com/tryggvib/ckanext-budgets',
license='AGPLv3',
packages=find_packages(exclude=['ez_setup', 'examples', 'tests']),
namespace_packages=['ckanext', 'ckanext.budgets'],
include_package_data=True,
zip_safe=False,
install_requires=[
# -*- Extra requirements: -*-
'unicodecsv',
'requests',
'six',
'py-dateutil',
'budgetdatapackage'
],
entry_points='''
[ckan.plugins]
budgets=ckanext.budgets.plugin:BudgetDataPackagePlugin
''',
)
|
<commit_before>from setuptools import setup, find_packages
import sys, os
version = '0.1.0'
description = "CKAN extension for creating/distributing budget data packages"
with open('README.rst') as readme:
long_description = readme.read()
setup(
name='ckanext-budgets',
version=version,
description=description,
long_description=long_description,
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Other Audience',
'License :: OSI Approved :: GNU Affero General Public License v3',
'Programming Language :: Python :: 2.7',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
],
keywords='ckan',
author='Tryggvi Bjorgvinsson',
author_email='tryggvi.bjorgvinsson@okfn.org',
url='https://github.com/tryggvib/ckanext-budgets',
license='AGPLv3',
packages=find_packages(exclude=['ez_setup', 'examples', 'tests']),
namespace_packages=['ckanext', 'ckanext.budgets'],
include_package_data=True,
zip_safe=False,
install_requires=[
# -*- Extra requirements: -*-
'unicodecsv',
'requests',
'six',
'py-dateutil',
],
entry_points='''
[ckan.plugins]
budgets=ckanext.budgets.plugin:BudgetDataPackagePlugin
''',
)
<commit_msg>Update dependencies and bump patch version
budgetdatapackage was missing from the dependencies and has now
been added.<commit_after>
|
from setuptools import setup, find_packages
import sys, os
version = '0.1.1'
description = "CKAN extension for creating/distributing budget data packages"
with open('README.rst') as readme:
long_description = readme.read()
setup(
name='ckanext-budgets',
version=version,
description=description,
long_description=long_description,
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Other Audience',
'License :: OSI Approved :: GNU Affero General Public License v3',
'Programming Language :: Python :: 2.7',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
],
keywords='ckan',
author='Tryggvi Bjorgvinsson',
author_email='tryggvi.bjorgvinsson@okfn.org',
url='https://github.com/tryggvib/ckanext-budgets',
license='AGPLv3',
packages=find_packages(exclude=['ez_setup', 'examples', 'tests']),
namespace_packages=['ckanext', 'ckanext.budgets'],
include_package_data=True,
zip_safe=False,
install_requires=[
# -*- Extra requirements: -*-
'unicodecsv',
'requests',
'six',
'py-dateutil',
'budgetdatapackage'
],
entry_points='''
[ckan.plugins]
budgets=ckanext.budgets.plugin:BudgetDataPackagePlugin
''',
)
|
from setuptools import setup, find_packages
import sys, os
version = '0.1.0'
description = "CKAN extension for creating/distributing budget data packages"
with open('README.rst') as readme:
long_description = readme.read()
setup(
name='ckanext-budgets',
version=version,
description=description,
long_description=long_description,
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Other Audience',
'License :: OSI Approved :: GNU Affero General Public License v3',
'Programming Language :: Python :: 2.7',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
],
keywords='ckan',
author='Tryggvi Bjorgvinsson',
author_email='tryggvi.bjorgvinsson@okfn.org',
url='https://github.com/tryggvib/ckanext-budgets',
license='AGPLv3',
packages=find_packages(exclude=['ez_setup', 'examples', 'tests']),
namespace_packages=['ckanext', 'ckanext.budgets'],
include_package_data=True,
zip_safe=False,
install_requires=[
# -*- Extra requirements: -*-
'unicodecsv',
'requests',
'six',
'py-dateutil',
],
entry_points='''
[ckan.plugins]
budgets=ckanext.budgets.plugin:BudgetDataPackagePlugin
''',
)
Update dependencies and bump patch version
budgetdatapackage was missing from the dependencies and has now
been added.from setuptools import setup, find_packages
import sys, os
version = '0.1.1'
description = "CKAN extension for creating/distributing budget data packages"
with open('README.rst') as readme:
long_description = readme.read()
setup(
name='ckanext-budgets',
version=version,
description=description,
long_description=long_description,
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Other Audience',
'License :: OSI Approved :: GNU Affero General Public License v3',
'Programming Language :: Python :: 2.7',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
],
keywords='ckan',
author='Tryggvi Bjorgvinsson',
author_email='tryggvi.bjorgvinsson@okfn.org',
url='https://github.com/tryggvib/ckanext-budgets',
license='AGPLv3',
packages=find_packages(exclude=['ez_setup', 'examples', 'tests']),
namespace_packages=['ckanext', 'ckanext.budgets'],
include_package_data=True,
zip_safe=False,
install_requires=[
# -*- Extra requirements: -*-
'unicodecsv',
'requests',
'six',
'py-dateutil',
'budgetdatapackage'
],
entry_points='''
[ckan.plugins]
budgets=ckanext.budgets.plugin:BudgetDataPackagePlugin
''',
)
|
<commit_before>from setuptools import setup, find_packages
import sys, os
version = '0.1.0'
description = "CKAN extension for creating/distributing budget data packages"
with open('README.rst') as readme:
long_description = readme.read()
setup(
name='ckanext-budgets',
version=version,
description=description,
long_description=long_description,
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Other Audience',
'License :: OSI Approved :: GNU Affero General Public License v3',
'Programming Language :: Python :: 2.7',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
],
keywords='ckan',
author='Tryggvi Bjorgvinsson',
author_email='tryggvi.bjorgvinsson@okfn.org',
url='https://github.com/tryggvib/ckanext-budgets',
license='AGPLv3',
packages=find_packages(exclude=['ez_setup', 'examples', 'tests']),
namespace_packages=['ckanext', 'ckanext.budgets'],
include_package_data=True,
zip_safe=False,
install_requires=[
# -*- Extra requirements: -*-
'unicodecsv',
'requests',
'six',
'py-dateutil',
],
entry_points='''
[ckan.plugins]
budgets=ckanext.budgets.plugin:BudgetDataPackagePlugin
''',
)
<commit_msg>Update dependencies and bump patch version
budgetdatapackage was missing from the dependencies and has now
been added.<commit_after>from setuptools import setup, find_packages
import sys, os
version = '0.1.1'
description = "CKAN extension for creating/distributing budget data packages"
with open('README.rst') as readme:
long_description = readme.read()
setup(
name='ckanext-budgets',
version=version,
description=description,
long_description=long_description,
classifiers=[
'Development Status :: 4 - Beta',
'Intended Audience :: Other Audience',
'License :: OSI Approved :: GNU Affero General Public License v3',
'Programming Language :: Python :: 2.7',
'Topic :: Internet :: WWW/HTTP :: Dynamic Content',
],
keywords='ckan',
author='Tryggvi Bjorgvinsson',
author_email='tryggvi.bjorgvinsson@okfn.org',
url='https://github.com/tryggvib/ckanext-budgets',
license='AGPLv3',
packages=find_packages(exclude=['ez_setup', 'examples', 'tests']),
namespace_packages=['ckanext', 'ckanext.budgets'],
include_package_data=True,
zip_safe=False,
install_requires=[
# -*- Extra requirements: -*-
'unicodecsv',
'requests',
'six',
'py-dateutil',
'budgetdatapackage'
],
entry_points='''
[ckan.plugins]
budgets=ckanext.budgets.plugin:BudgetDataPackagePlugin
''',
)
|
72d01a13761afbdcdfeb4fabb9095fa786403b18
|
setup.py
|
setup.py
|
from setuptools import find_packages, setup
setup(
name='ngx-task',
version='0.1',
description='Testimonial for candidates to show up their code-foo',
author='Dmitry Shulyak',
author_email='dmitri.shulyak@gmail.com',
url='https://github.com/shudmi/ngx-task',
classifiers=[
'License :: Apache License 2.0',
'Programming Language :: Python',
'Programming Language :: Python 3',
'Programming Language :: Python 3.4',
],
packages=find_packages(exclude=['tests', 'tests.*']),
install_requires=[],
entry_points="""
[console_scripts]
ngx_generate=ngx_task.cli.generate_data
ngx_process=ngx_task.cli.process_data
"""
)
|
from setuptools import find_packages, setup
setup(
name='ngx-task',
version='0.2',
description='Testimonial for candidates to show up their code-foo',
author='Dmitry Shulyak',
author_email='dmitri.shulyak@gmail.com',
url='https://github.com/shudmi/ngx-task',
classifiers=[
'License :: Apache License 2.0',
'Programming Language :: Python',
'Programming Language :: Python 3',
'Programming Language :: Python 3.4',
],
packages=find_packages(exclude=['tests', 'tests.*']),
install_requires=[],
tests_require=[
"pytest==3.0.7",
],
entry_points="""
[console_scripts]
ngx_generate=ngx_task.cli:generate_data
ngx_process=ngx_task.cli:process_data
"""
)
|
Add tests requirements, fix console scripts definitions
|
Add tests requirements, fix console scripts definitions
|
Python
|
apache-2.0
|
shudmi/ngx-task
|
from setuptools import find_packages, setup
setup(
name='ngx-task',
version='0.1',
description='Testimonial for candidates to show up their code-foo',
author='Dmitry Shulyak',
author_email='dmitri.shulyak@gmail.com',
url='https://github.com/shudmi/ngx-task',
classifiers=[
'License :: Apache License 2.0',
'Programming Language :: Python',
'Programming Language :: Python 3',
'Programming Language :: Python 3.4',
],
packages=find_packages(exclude=['tests', 'tests.*']),
install_requires=[],
entry_points="""
[console_scripts]
ngx_generate=ngx_task.cli.generate_data
ngx_process=ngx_task.cli.process_data
"""
)
Add tests requirements, fix console scripts definitions
|
from setuptools import find_packages, setup
setup(
name='ngx-task',
version='0.2',
description='Testimonial for candidates to show up their code-foo',
author='Dmitry Shulyak',
author_email='dmitri.shulyak@gmail.com',
url='https://github.com/shudmi/ngx-task',
classifiers=[
'License :: Apache License 2.0',
'Programming Language :: Python',
'Programming Language :: Python 3',
'Programming Language :: Python 3.4',
],
packages=find_packages(exclude=['tests', 'tests.*']),
install_requires=[],
tests_require=[
"pytest==3.0.7",
],
entry_points="""
[console_scripts]
ngx_generate=ngx_task.cli:generate_data
ngx_process=ngx_task.cli:process_data
"""
)
|
<commit_before>from setuptools import find_packages, setup
setup(
name='ngx-task',
version='0.1',
description='Testimonial for candidates to show up their code-foo',
author='Dmitry Shulyak',
author_email='dmitri.shulyak@gmail.com',
url='https://github.com/shudmi/ngx-task',
classifiers=[
'License :: Apache License 2.0',
'Programming Language :: Python',
'Programming Language :: Python 3',
'Programming Language :: Python 3.4',
],
packages=find_packages(exclude=['tests', 'tests.*']),
install_requires=[],
entry_points="""
[console_scripts]
ngx_generate=ngx_task.cli.generate_data
ngx_process=ngx_task.cli.process_data
"""
)
<commit_msg>Add tests requirements, fix console scripts definitions<commit_after>
|
from setuptools import find_packages, setup
setup(
name='ngx-task',
version='0.2',
description='Testimonial for candidates to show up their code-foo',
author='Dmitry Shulyak',
author_email='dmitri.shulyak@gmail.com',
url='https://github.com/shudmi/ngx-task',
classifiers=[
'License :: Apache License 2.0',
'Programming Language :: Python',
'Programming Language :: Python 3',
'Programming Language :: Python 3.4',
],
packages=find_packages(exclude=['tests', 'tests.*']),
install_requires=[],
tests_require=[
"pytest==3.0.7",
],
entry_points="""
[console_scripts]
ngx_generate=ngx_task.cli:generate_data
ngx_process=ngx_task.cli:process_data
"""
)
|
from setuptools import find_packages, setup
setup(
name='ngx-task',
version='0.1',
description='Testimonial for candidates to show up their code-foo',
author='Dmitry Shulyak',
author_email='dmitri.shulyak@gmail.com',
url='https://github.com/shudmi/ngx-task',
classifiers=[
'License :: Apache License 2.0',
'Programming Language :: Python',
'Programming Language :: Python 3',
'Programming Language :: Python 3.4',
],
packages=find_packages(exclude=['tests', 'tests.*']),
install_requires=[],
entry_points="""
[console_scripts]
ngx_generate=ngx_task.cli.generate_data
ngx_process=ngx_task.cli.process_data
"""
)
Add tests requirements, fix console scripts definitionsfrom setuptools import find_packages, setup
setup(
name='ngx-task',
version='0.2',
description='Testimonial for candidates to show up their code-foo',
author='Dmitry Shulyak',
author_email='dmitri.shulyak@gmail.com',
url='https://github.com/shudmi/ngx-task',
classifiers=[
'License :: Apache License 2.0',
'Programming Language :: Python',
'Programming Language :: Python 3',
'Programming Language :: Python 3.4',
],
packages=find_packages(exclude=['tests', 'tests.*']),
install_requires=[],
tests_require=[
"pytest==3.0.7",
],
entry_points="""
[console_scripts]
ngx_generate=ngx_task.cli:generate_data
ngx_process=ngx_task.cli:process_data
"""
)
|
<commit_before>from setuptools import find_packages, setup
setup(
name='ngx-task',
version='0.1',
description='Testimonial for candidates to show up their code-foo',
author='Dmitry Shulyak',
author_email='dmitri.shulyak@gmail.com',
url='https://github.com/shudmi/ngx-task',
classifiers=[
'License :: Apache License 2.0',
'Programming Language :: Python',
'Programming Language :: Python 3',
'Programming Language :: Python 3.4',
],
packages=find_packages(exclude=['tests', 'tests.*']),
install_requires=[],
entry_points="""
[console_scripts]
ngx_generate=ngx_task.cli.generate_data
ngx_process=ngx_task.cli.process_data
"""
)
<commit_msg>Add tests requirements, fix console scripts definitions<commit_after>from setuptools import find_packages, setup
setup(
name='ngx-task',
version='0.2',
description='Testimonial for candidates to show up their code-foo',
author='Dmitry Shulyak',
author_email='dmitri.shulyak@gmail.com',
url='https://github.com/shudmi/ngx-task',
classifiers=[
'License :: Apache License 2.0',
'Programming Language :: Python',
'Programming Language :: Python 3',
'Programming Language :: Python 3.4',
],
packages=find_packages(exclude=['tests', 'tests.*']),
install_requires=[],
tests_require=[
"pytest==3.0.7",
],
entry_points="""
[console_scripts]
ngx_generate=ngx_task.cli:generate_data
ngx_process=ngx_task.cli:process_data
"""
)
|
bebe2be26d21289bb43936d3895c4b29126d508c
|
setup.py
|
setup.py
|
from setuptools import setup
# "import" __version__
for line in open('nbsphinx.py'):
if line.startswith('__version__'):
exec(line)
break
setup(
name='nbsphinx',
version=__version__,
py_modules=['nbsphinx'],
install_requires=[
'docutils',
'jinja2',
'nbconvert',
'traitlets',
'nbformat',
'sphinx>=1.3.2,!=1.5.0',
],
author='Matthias Geier',
author_email='Matthias.Geier@gmail.com',
description='Jupyter Notebook Tools for Sphinx',
long_description=open('README.rst').read(),
license='MIT',
keywords='Sphinx Jupyter notebook'.split(),
url='http://nbsphinx.rtfd.io/',
platforms='any',
classifiers=[
'Framework :: Sphinx',
'Framework :: Sphinx :: Extension',
'Framework :: Jupyter',
'Intended Audience :: Education',
'Intended Audience :: Science/Research',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 3',
'Topic :: Documentation :: Sphinx',
],
zip_safe=True,
)
|
from setuptools import setup
# "import" __version__
for line in open('nbsphinx.py'):
if line.startswith('__version__'):
exec(line)
break
setup(
name='nbsphinx',
version=__version__,
py_modules=['nbsphinx'],
install_requires=[
'docutils',
'jinja2',
'nbconvert',
'traitlets',
'nbformat',
'sphinx>=1.3.2,!=1.5.0',
],
author='Matthias Geier',
author_email='Matthias.Geier@gmail.com',
description='Jupyter Notebook Tools for Sphinx',
long_description=open('README.rst').read(),
license='MIT',
keywords='Sphinx Jupyter notebook'.split(),
url='http://nbsphinx.readthedocs.io/',
platforms='any',
classifiers=[
'Framework :: Sphinx',
'Framework :: Sphinx :: Extension',
'Framework :: Jupyter',
'Intended Audience :: Education',
'Intended Audience :: Science/Research',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 3',
'Topic :: Documentation :: Sphinx',
],
zip_safe=True,
)
|
Change rtfd -> readthedocs in package description
|
DOC: Change rtfd -> readthedocs in package description
|
Python
|
mit
|
spatialaudio/nbsphinx,spatialaudio/nbsphinx,spatialaudio/nbsphinx
|
from setuptools import setup
# "import" __version__
for line in open('nbsphinx.py'):
if line.startswith('__version__'):
exec(line)
break
setup(
name='nbsphinx',
version=__version__,
py_modules=['nbsphinx'],
install_requires=[
'docutils',
'jinja2',
'nbconvert',
'traitlets',
'nbformat',
'sphinx>=1.3.2,!=1.5.0',
],
author='Matthias Geier',
author_email='Matthias.Geier@gmail.com',
description='Jupyter Notebook Tools for Sphinx',
long_description=open('README.rst').read(),
license='MIT',
keywords='Sphinx Jupyter notebook'.split(),
url='http://nbsphinx.rtfd.io/',
platforms='any',
classifiers=[
'Framework :: Sphinx',
'Framework :: Sphinx :: Extension',
'Framework :: Jupyter',
'Intended Audience :: Education',
'Intended Audience :: Science/Research',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 3',
'Topic :: Documentation :: Sphinx',
],
zip_safe=True,
)
DOC: Change rtfd -> readthedocs in package description
|
from setuptools import setup
# "import" __version__
for line in open('nbsphinx.py'):
if line.startswith('__version__'):
exec(line)
break
setup(
name='nbsphinx',
version=__version__,
py_modules=['nbsphinx'],
install_requires=[
'docutils',
'jinja2',
'nbconvert',
'traitlets',
'nbformat',
'sphinx>=1.3.2,!=1.5.0',
],
author='Matthias Geier',
author_email='Matthias.Geier@gmail.com',
description='Jupyter Notebook Tools for Sphinx',
long_description=open('README.rst').read(),
license='MIT',
keywords='Sphinx Jupyter notebook'.split(),
url='http://nbsphinx.readthedocs.io/',
platforms='any',
classifiers=[
'Framework :: Sphinx',
'Framework :: Sphinx :: Extension',
'Framework :: Jupyter',
'Intended Audience :: Education',
'Intended Audience :: Science/Research',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 3',
'Topic :: Documentation :: Sphinx',
],
zip_safe=True,
)
|
<commit_before>from setuptools import setup
# "import" __version__
for line in open('nbsphinx.py'):
if line.startswith('__version__'):
exec(line)
break
setup(
name='nbsphinx',
version=__version__,
py_modules=['nbsphinx'],
install_requires=[
'docutils',
'jinja2',
'nbconvert',
'traitlets',
'nbformat',
'sphinx>=1.3.2,!=1.5.0',
],
author='Matthias Geier',
author_email='Matthias.Geier@gmail.com',
description='Jupyter Notebook Tools for Sphinx',
long_description=open('README.rst').read(),
license='MIT',
keywords='Sphinx Jupyter notebook'.split(),
url='http://nbsphinx.rtfd.io/',
platforms='any',
classifiers=[
'Framework :: Sphinx',
'Framework :: Sphinx :: Extension',
'Framework :: Jupyter',
'Intended Audience :: Education',
'Intended Audience :: Science/Research',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 3',
'Topic :: Documentation :: Sphinx',
],
zip_safe=True,
)
<commit_msg>DOC: Change rtfd -> readthedocs in package description<commit_after>
|
from setuptools import setup
# "import" __version__
for line in open('nbsphinx.py'):
if line.startswith('__version__'):
exec(line)
break
setup(
name='nbsphinx',
version=__version__,
py_modules=['nbsphinx'],
install_requires=[
'docutils',
'jinja2',
'nbconvert',
'traitlets',
'nbformat',
'sphinx>=1.3.2,!=1.5.0',
],
author='Matthias Geier',
author_email='Matthias.Geier@gmail.com',
description='Jupyter Notebook Tools for Sphinx',
long_description=open('README.rst').read(),
license='MIT',
keywords='Sphinx Jupyter notebook'.split(),
url='http://nbsphinx.readthedocs.io/',
platforms='any',
classifiers=[
'Framework :: Sphinx',
'Framework :: Sphinx :: Extension',
'Framework :: Jupyter',
'Intended Audience :: Education',
'Intended Audience :: Science/Research',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 3',
'Topic :: Documentation :: Sphinx',
],
zip_safe=True,
)
|
from setuptools import setup
# "import" __version__
for line in open('nbsphinx.py'):
if line.startswith('__version__'):
exec(line)
break
setup(
name='nbsphinx',
version=__version__,
py_modules=['nbsphinx'],
install_requires=[
'docutils',
'jinja2',
'nbconvert',
'traitlets',
'nbformat',
'sphinx>=1.3.2,!=1.5.0',
],
author='Matthias Geier',
author_email='Matthias.Geier@gmail.com',
description='Jupyter Notebook Tools for Sphinx',
long_description=open('README.rst').read(),
license='MIT',
keywords='Sphinx Jupyter notebook'.split(),
url='http://nbsphinx.rtfd.io/',
platforms='any',
classifiers=[
'Framework :: Sphinx',
'Framework :: Sphinx :: Extension',
'Framework :: Jupyter',
'Intended Audience :: Education',
'Intended Audience :: Science/Research',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 3',
'Topic :: Documentation :: Sphinx',
],
zip_safe=True,
)
DOC: Change rtfd -> readthedocs in package descriptionfrom setuptools import setup
# "import" __version__
for line in open('nbsphinx.py'):
if line.startswith('__version__'):
exec(line)
break
setup(
name='nbsphinx',
version=__version__,
py_modules=['nbsphinx'],
install_requires=[
'docutils',
'jinja2',
'nbconvert',
'traitlets',
'nbformat',
'sphinx>=1.3.2,!=1.5.0',
],
author='Matthias Geier',
author_email='Matthias.Geier@gmail.com',
description='Jupyter Notebook Tools for Sphinx',
long_description=open('README.rst').read(),
license='MIT',
keywords='Sphinx Jupyter notebook'.split(),
url='http://nbsphinx.readthedocs.io/',
platforms='any',
classifiers=[
'Framework :: Sphinx',
'Framework :: Sphinx :: Extension',
'Framework :: Jupyter',
'Intended Audience :: Education',
'Intended Audience :: Science/Research',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 3',
'Topic :: Documentation :: Sphinx',
],
zip_safe=True,
)
|
<commit_before>from setuptools import setup
# "import" __version__
for line in open('nbsphinx.py'):
if line.startswith('__version__'):
exec(line)
break
setup(
name='nbsphinx',
version=__version__,
py_modules=['nbsphinx'],
install_requires=[
'docutils',
'jinja2',
'nbconvert',
'traitlets',
'nbformat',
'sphinx>=1.3.2,!=1.5.0',
],
author='Matthias Geier',
author_email='Matthias.Geier@gmail.com',
description='Jupyter Notebook Tools for Sphinx',
long_description=open('README.rst').read(),
license='MIT',
keywords='Sphinx Jupyter notebook'.split(),
url='http://nbsphinx.rtfd.io/',
platforms='any',
classifiers=[
'Framework :: Sphinx',
'Framework :: Sphinx :: Extension',
'Framework :: Jupyter',
'Intended Audience :: Education',
'Intended Audience :: Science/Research',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 3',
'Topic :: Documentation :: Sphinx',
],
zip_safe=True,
)
<commit_msg>DOC: Change rtfd -> readthedocs in package description<commit_after>from setuptools import setup
# "import" __version__
for line in open('nbsphinx.py'):
if line.startswith('__version__'):
exec(line)
break
setup(
name='nbsphinx',
version=__version__,
py_modules=['nbsphinx'],
install_requires=[
'docutils',
'jinja2',
'nbconvert',
'traitlets',
'nbformat',
'sphinx>=1.3.2,!=1.5.0',
],
author='Matthias Geier',
author_email='Matthias.Geier@gmail.com',
description='Jupyter Notebook Tools for Sphinx',
long_description=open('README.rst').read(),
license='MIT',
keywords='Sphinx Jupyter notebook'.split(),
url='http://nbsphinx.readthedocs.io/',
platforms='any',
classifiers=[
'Framework :: Sphinx',
'Framework :: Sphinx :: Extension',
'Framework :: Jupyter',
'Intended Audience :: Education',
'Intended Audience :: Science/Research',
'License :: OSI Approved :: MIT License',
'Operating System :: OS Independent',
'Programming Language :: Python',
'Programming Language :: Python :: 3',
'Topic :: Documentation :: Sphinx',
],
zip_safe=True,
)
|
9d19fb7ada5caaa2dc74736cd12635bed3d8516a
|
setup.py
|
setup.py
|
import os
from setuptools import setup, find_packages
setup(name='morepath',
version = '0.1dev',
description="A micro web-framework with superpowers",
author="Martijn Faassen",
author_email="faassen@startifact.com",
license="BSD",
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=[
'setuptools',
'venusian',
'reg',
'werkzeug',
],
extras_require = dict(
test=['pytest >= 2.0',
'pytest-cov'],
),
)
|
import os
from setuptools import setup, find_packages
setup(name='morepath',
version = '0.1dev',
description="A micro web-framework with superpowers",
author="Martijn Faassen",
author_email="faassen@startifact.com",
license="BSD",
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=[
'setuptools',
'venusian >= 1.0a8',
'reg',
'werkzeug >= 0.9.4',
],
extras_require = dict(
test=['pytest >= 2.0',
'pytest-cov'],
),
)
|
Put in some version requirements.
|
Put in some version requirements.
|
Python
|
bsd-3-clause
|
taschini/morepath,faassen/morepath,morepath/morepath
|
import os
from setuptools import setup, find_packages
setup(name='morepath',
version = '0.1dev',
description="A micro web-framework with superpowers",
author="Martijn Faassen",
author_email="faassen@startifact.com",
license="BSD",
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=[
'setuptools',
'venusian',
'reg',
'werkzeug',
],
extras_require = dict(
test=['pytest >= 2.0',
'pytest-cov'],
),
)
Put in some version requirements.
|
import os
from setuptools import setup, find_packages
setup(name='morepath',
version = '0.1dev',
description="A micro web-framework with superpowers",
author="Martijn Faassen",
author_email="faassen@startifact.com",
license="BSD",
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=[
'setuptools',
'venusian >= 1.0a8',
'reg',
'werkzeug >= 0.9.4',
],
extras_require = dict(
test=['pytest >= 2.0',
'pytest-cov'],
),
)
|
<commit_before>import os
from setuptools import setup, find_packages
setup(name='morepath',
version = '0.1dev',
description="A micro web-framework with superpowers",
author="Martijn Faassen",
author_email="faassen@startifact.com",
license="BSD",
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=[
'setuptools',
'venusian',
'reg',
'werkzeug',
],
extras_require = dict(
test=['pytest >= 2.0',
'pytest-cov'],
),
)
<commit_msg>Put in some version requirements.<commit_after>
|
import os
from setuptools import setup, find_packages
setup(name='morepath',
version = '0.1dev',
description="A micro web-framework with superpowers",
author="Martijn Faassen",
author_email="faassen@startifact.com",
license="BSD",
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=[
'setuptools',
'venusian >= 1.0a8',
'reg',
'werkzeug >= 0.9.4',
],
extras_require = dict(
test=['pytest >= 2.0',
'pytest-cov'],
),
)
|
import os
from setuptools import setup, find_packages
setup(name='morepath',
version = '0.1dev',
description="A micro web-framework with superpowers",
author="Martijn Faassen",
author_email="faassen@startifact.com",
license="BSD",
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=[
'setuptools',
'venusian',
'reg',
'werkzeug',
],
extras_require = dict(
test=['pytest >= 2.0',
'pytest-cov'],
),
)
Put in some version requirements.import os
from setuptools import setup, find_packages
setup(name='morepath',
version = '0.1dev',
description="A micro web-framework with superpowers",
author="Martijn Faassen",
author_email="faassen@startifact.com",
license="BSD",
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=[
'setuptools',
'venusian >= 1.0a8',
'reg',
'werkzeug >= 0.9.4',
],
extras_require = dict(
test=['pytest >= 2.0',
'pytest-cov'],
),
)
|
<commit_before>import os
from setuptools import setup, find_packages
setup(name='morepath',
version = '0.1dev',
description="A micro web-framework with superpowers",
author="Martijn Faassen",
author_email="faassen@startifact.com",
license="BSD",
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=[
'setuptools',
'venusian',
'reg',
'werkzeug',
],
extras_require = dict(
test=['pytest >= 2.0',
'pytest-cov'],
),
)
<commit_msg>Put in some version requirements.<commit_after>import os
from setuptools import setup, find_packages
setup(name='morepath',
version = '0.1dev',
description="A micro web-framework with superpowers",
author="Martijn Faassen",
author_email="faassen@startifact.com",
license="BSD",
packages=find_packages(),
include_package_data=True,
zip_safe=False,
install_requires=[
'setuptools',
'venusian >= 1.0a8',
'reg',
'werkzeug >= 0.9.4',
],
extras_require = dict(
test=['pytest >= 2.0',
'pytest-cov'],
),
)
|
cbbcbb5707a90929da8d47f6b3322cebec983279
|
setup.py
|
setup.py
|
from setuptools import setup
from distutils.core import Command
import os
import sys
class TestCommand(Command):
description = "Run tests"
user_options = []
def initialize_options(self):
pass
def finalize_options(self):
pass
def run(self):
import subprocess
errno = subprocess.call(['nosetests', '--debug=DEBUG', '-s'])
raise SystemExit(errno)
setup(
name='mafan',
version='0.2.7',
author='Herman Schaaf',
author_email='herman@ironzebra.com',
packages=['mafan', 'mafan.hanzidentifier'],
scripts=['bin/convert.py'],
url='https://github.com/hermanschaaf/mafan',
license='LICENSE.txt',
description='A toolbox for working with the Chinese language in Python',
long_description=open('docs/README.md').read(),
cmdclass={
'test': TestCommand,
},
install_requires=[
"jieba == 0.29",
"argparse == 1.1",
"chardet == 2.1.1",
"wsgiref == 0.1.2",
],
)
|
from setuptools import setup
from distutils.core import Command
import os
import sys
class TestCommand(Command):
description = "Run tests"
user_options = []
def initialize_options(self):
pass
def finalize_options(self):
pass
def run(self):
import subprocess
errno = subprocess.call(['nosetests', '--debug=DEBUG', '-s'])
raise SystemExit(errno)
setup(
name='mafan',
version='0.2.10',
author='Herman Schaaf',
author_email='herman@ironzebra.com',
packages=[
'mafan',
'mafan.hanzidentifier',
'mafan.third_party',
'mafan.third_party.jianfan'
],
scripts=['bin/convert.py'],
url='https://github.com/hermanschaaf/mafan',
license='LICENSE.txt',
description='A toolbox for working with the Chinese language in Python',
long_description=open('docs/README.md').read(),
cmdclass={
'test': TestCommand,
},
install_requires=[
"jieba == 0.29",
"argparse == 1.1",
"chardet == 2.1.1",
"wsgiref == 0.1.2",
],
)
|
Add jianfan to python package installation
|
Add jianfan to python package installation
|
Python
|
mit
|
hermanschaaf/mafan,cychiang/mafan
|
from setuptools import setup
from distutils.core import Command
import os
import sys
class TestCommand(Command):
description = "Run tests"
user_options = []
def initialize_options(self):
pass
def finalize_options(self):
pass
def run(self):
import subprocess
errno = subprocess.call(['nosetests', '--debug=DEBUG', '-s'])
raise SystemExit(errno)
setup(
name='mafan',
version='0.2.7',
author='Herman Schaaf',
author_email='herman@ironzebra.com',
packages=['mafan', 'mafan.hanzidentifier'],
scripts=['bin/convert.py'],
url='https://github.com/hermanschaaf/mafan',
license='LICENSE.txt',
description='A toolbox for working with the Chinese language in Python',
long_description=open('docs/README.md').read(),
cmdclass={
'test': TestCommand,
},
install_requires=[
"jieba == 0.29",
"argparse == 1.1",
"chardet == 2.1.1",
"wsgiref == 0.1.2",
],
)
Add jianfan to python package installation
|
from setuptools import setup
from distutils.core import Command
import os
import sys
class TestCommand(Command):
description = "Run tests"
user_options = []
def initialize_options(self):
pass
def finalize_options(self):
pass
def run(self):
import subprocess
errno = subprocess.call(['nosetests', '--debug=DEBUG', '-s'])
raise SystemExit(errno)
setup(
name='mafan',
version='0.2.10',
author='Herman Schaaf',
author_email='herman@ironzebra.com',
packages=[
'mafan',
'mafan.hanzidentifier',
'mafan.third_party',
'mafan.third_party.jianfan'
],
scripts=['bin/convert.py'],
url='https://github.com/hermanschaaf/mafan',
license='LICENSE.txt',
description='A toolbox for working with the Chinese language in Python',
long_description=open('docs/README.md').read(),
cmdclass={
'test': TestCommand,
},
install_requires=[
"jieba == 0.29",
"argparse == 1.1",
"chardet == 2.1.1",
"wsgiref == 0.1.2",
],
)
|
<commit_before>from setuptools import setup
from distutils.core import Command
import os
import sys
class TestCommand(Command):
description = "Run tests"
user_options = []
def initialize_options(self):
pass
def finalize_options(self):
pass
def run(self):
import subprocess
errno = subprocess.call(['nosetests', '--debug=DEBUG', '-s'])
raise SystemExit(errno)
setup(
name='mafan',
version='0.2.7',
author='Herman Schaaf',
author_email='herman@ironzebra.com',
packages=['mafan', 'mafan.hanzidentifier'],
scripts=['bin/convert.py'],
url='https://github.com/hermanschaaf/mafan',
license='LICENSE.txt',
description='A toolbox for working with the Chinese language in Python',
long_description=open('docs/README.md').read(),
cmdclass={
'test': TestCommand,
},
install_requires=[
"jieba == 0.29",
"argparse == 1.1",
"chardet == 2.1.1",
"wsgiref == 0.1.2",
],
)
<commit_msg>Add jianfan to python package installation<commit_after>
|
from setuptools import setup
from distutils.core import Command
import os
import sys
class TestCommand(Command):
description = "Run tests"
user_options = []
def initialize_options(self):
pass
def finalize_options(self):
pass
def run(self):
import subprocess
errno = subprocess.call(['nosetests', '--debug=DEBUG', '-s'])
raise SystemExit(errno)
setup(
name='mafan',
version='0.2.10',
author='Herman Schaaf',
author_email='herman@ironzebra.com',
packages=[
'mafan',
'mafan.hanzidentifier',
'mafan.third_party',
'mafan.third_party.jianfan'
],
scripts=['bin/convert.py'],
url='https://github.com/hermanschaaf/mafan',
license='LICENSE.txt',
description='A toolbox for working with the Chinese language in Python',
long_description=open('docs/README.md').read(),
cmdclass={
'test': TestCommand,
},
install_requires=[
"jieba == 0.29",
"argparse == 1.1",
"chardet == 2.1.1",
"wsgiref == 0.1.2",
],
)
|
from setuptools import setup
from distutils.core import Command
import os
import sys
class TestCommand(Command):
description = "Run tests"
user_options = []
def initialize_options(self):
pass
def finalize_options(self):
pass
def run(self):
import subprocess
errno = subprocess.call(['nosetests', '--debug=DEBUG', '-s'])
raise SystemExit(errno)
setup(
name='mafan',
version='0.2.7',
author='Herman Schaaf',
author_email='herman@ironzebra.com',
packages=['mafan', 'mafan.hanzidentifier'],
scripts=['bin/convert.py'],
url='https://github.com/hermanschaaf/mafan',
license='LICENSE.txt',
description='A toolbox for working with the Chinese language in Python',
long_description=open('docs/README.md').read(),
cmdclass={
'test': TestCommand,
},
install_requires=[
"jieba == 0.29",
"argparse == 1.1",
"chardet == 2.1.1",
"wsgiref == 0.1.2",
],
)
Add jianfan to python package installationfrom setuptools import setup
from distutils.core import Command
import os
import sys
class TestCommand(Command):
description = "Run tests"
user_options = []
def initialize_options(self):
pass
def finalize_options(self):
pass
def run(self):
import subprocess
errno = subprocess.call(['nosetests', '--debug=DEBUG', '-s'])
raise SystemExit(errno)
setup(
name='mafan',
version='0.2.10',
author='Herman Schaaf',
author_email='herman@ironzebra.com',
packages=[
'mafan',
'mafan.hanzidentifier',
'mafan.third_party',
'mafan.third_party.jianfan'
],
scripts=['bin/convert.py'],
url='https://github.com/hermanschaaf/mafan',
license='LICENSE.txt',
description='A toolbox for working with the Chinese language in Python',
long_description=open('docs/README.md').read(),
cmdclass={
'test': TestCommand,
},
install_requires=[
"jieba == 0.29",
"argparse == 1.1",
"chardet == 2.1.1",
"wsgiref == 0.1.2",
],
)
|
<commit_before>from setuptools import setup
from distutils.core import Command
import os
import sys
class TestCommand(Command):
description = "Run tests"
user_options = []
def initialize_options(self):
pass
def finalize_options(self):
pass
def run(self):
import subprocess
errno = subprocess.call(['nosetests', '--debug=DEBUG', '-s'])
raise SystemExit(errno)
setup(
name='mafan',
version='0.2.7',
author='Herman Schaaf',
author_email='herman@ironzebra.com',
packages=['mafan', 'mafan.hanzidentifier'],
scripts=['bin/convert.py'],
url='https://github.com/hermanschaaf/mafan',
license='LICENSE.txt',
description='A toolbox for working with the Chinese language in Python',
long_description=open('docs/README.md').read(),
cmdclass={
'test': TestCommand,
},
install_requires=[
"jieba == 0.29",
"argparse == 1.1",
"chardet == 2.1.1",
"wsgiref == 0.1.2",
],
)
<commit_msg>Add jianfan to python package installation<commit_after>from setuptools import setup
from distutils.core import Command
import os
import sys
class TestCommand(Command):
description = "Run tests"
user_options = []
def initialize_options(self):
pass
def finalize_options(self):
pass
def run(self):
import subprocess
errno = subprocess.call(['nosetests', '--debug=DEBUG', '-s'])
raise SystemExit(errno)
setup(
name='mafan',
version='0.2.10',
author='Herman Schaaf',
author_email='herman@ironzebra.com',
packages=[
'mafan',
'mafan.hanzidentifier',
'mafan.third_party',
'mafan.third_party.jianfan'
],
scripts=['bin/convert.py'],
url='https://github.com/hermanschaaf/mafan',
license='LICENSE.txt',
description='A toolbox for working with the Chinese language in Python',
long_description=open('docs/README.md').read(),
cmdclass={
'test': TestCommand,
},
install_requires=[
"jieba == 0.29",
"argparse == 1.1",
"chardet == 2.1.1",
"wsgiref == 0.1.2",
],
)
|
ba2e263e40e1324f8fd6c3ee012a8d9ada46bdd1
|
setup.py
|
setup.py
|
#!/usr/local/bin/python3
from distutils.core import setup, Extension
setup(
name = 'lazy',
version = '1.0',
description = 'Lazy',
ext_modules = [
Extension(
name = 'b._collections',
sources = [
'src/collections.c',
],
),
Extension(
name = 'b._types',
depends = [
'include/memoizer.h',
],
include_dirs = [
'include',
],
sources = [
'src/memoizer.c',
'src/types.c',
],
),
],
)
|
#!/usr/local/bin/python3
from distutils.core import setup, Extension
# Workaround -Werror=statement-after-declaration
# http://bugs.python.org/issue18211
import os
os.environ['CFLAGS'] = '-Wno-unused-result'
setup(
name = 'lazy',
version = '1.0',
description = 'Lazy',
ext_modules = [
Extension(
name = 'b._collections',
sources = [
'src/collections.c',
],
),
Extension(
name = 'b._types',
depends = [
'include/memoizer.h',
],
include_dirs = [
'include',
],
sources = [
'src/memoizer.c',
'src/types.c',
],
),
],
)
|
Disable strict C90 flag that appears to be accidentally being used by distutils
|
Disable strict C90 flag that appears to be accidentally being used by distutils
|
Python
|
apache-2.0
|
blake-sheridan/py,blake-sheridan/py
|
#!/usr/local/bin/python3
from distutils.core import setup, Extension
setup(
name = 'lazy',
version = '1.0',
description = 'Lazy',
ext_modules = [
Extension(
name = 'b._collections',
sources = [
'src/collections.c',
],
),
Extension(
name = 'b._types',
depends = [
'include/memoizer.h',
],
include_dirs = [
'include',
],
sources = [
'src/memoizer.c',
'src/types.c',
],
),
],
)
Disable strict C90 flag that appears to be accidentally being used by distutils
|
#!/usr/local/bin/python3
from distutils.core import setup, Extension
# Workaround -Werror=statement-after-declaration
# http://bugs.python.org/issue18211
import os
os.environ['CFLAGS'] = '-Wno-unused-result'
setup(
name = 'lazy',
version = '1.0',
description = 'Lazy',
ext_modules = [
Extension(
name = 'b._collections',
sources = [
'src/collections.c',
],
),
Extension(
name = 'b._types',
depends = [
'include/memoizer.h',
],
include_dirs = [
'include',
],
sources = [
'src/memoizer.c',
'src/types.c',
],
),
],
)
|
<commit_before>#!/usr/local/bin/python3
from distutils.core import setup, Extension
setup(
name = 'lazy',
version = '1.0',
description = 'Lazy',
ext_modules = [
Extension(
name = 'b._collections',
sources = [
'src/collections.c',
],
),
Extension(
name = 'b._types',
depends = [
'include/memoizer.h',
],
include_dirs = [
'include',
],
sources = [
'src/memoizer.c',
'src/types.c',
],
),
],
)
<commit_msg>Disable strict C90 flag that appears to be accidentally being used by distutils<commit_after>
|
#!/usr/local/bin/python3
from distutils.core import setup, Extension
# Workaround -Werror=statement-after-declaration
# http://bugs.python.org/issue18211
import os
os.environ['CFLAGS'] = '-Wno-unused-result'
setup(
name = 'lazy',
version = '1.0',
description = 'Lazy',
ext_modules = [
Extension(
name = 'b._collections',
sources = [
'src/collections.c',
],
),
Extension(
name = 'b._types',
depends = [
'include/memoizer.h',
],
include_dirs = [
'include',
],
sources = [
'src/memoizer.c',
'src/types.c',
],
),
],
)
|
#!/usr/local/bin/python3
from distutils.core import setup, Extension
setup(
name = 'lazy',
version = '1.0',
description = 'Lazy',
ext_modules = [
Extension(
name = 'b._collections',
sources = [
'src/collections.c',
],
),
Extension(
name = 'b._types',
depends = [
'include/memoizer.h',
],
include_dirs = [
'include',
],
sources = [
'src/memoizer.c',
'src/types.c',
],
),
],
)
Disable strict C90 flag that appears to be accidentally being used by distutils#!/usr/local/bin/python3
from distutils.core import setup, Extension
# Workaround -Werror=statement-after-declaration
# http://bugs.python.org/issue18211
import os
os.environ['CFLAGS'] = '-Wno-unused-result'
setup(
name = 'lazy',
version = '1.0',
description = 'Lazy',
ext_modules = [
Extension(
name = 'b._collections',
sources = [
'src/collections.c',
],
),
Extension(
name = 'b._types',
depends = [
'include/memoizer.h',
],
include_dirs = [
'include',
],
sources = [
'src/memoizer.c',
'src/types.c',
],
),
],
)
|
<commit_before>#!/usr/local/bin/python3
from distutils.core import setup, Extension
setup(
name = 'lazy',
version = '1.0',
description = 'Lazy',
ext_modules = [
Extension(
name = 'b._collections',
sources = [
'src/collections.c',
],
),
Extension(
name = 'b._types',
depends = [
'include/memoizer.h',
],
include_dirs = [
'include',
],
sources = [
'src/memoizer.c',
'src/types.c',
],
),
],
)
<commit_msg>Disable strict C90 flag that appears to be accidentally being used by distutils<commit_after>#!/usr/local/bin/python3
from distutils.core import setup, Extension
# Workaround -Werror=statement-after-declaration
# http://bugs.python.org/issue18211
import os
os.environ['CFLAGS'] = '-Wno-unused-result'
setup(
name = 'lazy',
version = '1.0',
description = 'Lazy',
ext_modules = [
Extension(
name = 'b._collections',
sources = [
'src/collections.c',
],
),
Extension(
name = 'b._types',
depends = [
'include/memoizer.h',
],
include_dirs = [
'include',
],
sources = [
'src/memoizer.c',
'src/types.c',
],
),
],
)
|
ce0f4a30cad570557ad67122333041806d411adc
|
tasks.py
|
tasks.py
|
from invoke import Collection
from invocations import docs
from invocations.checks import blacken
from invocations.packaging import release
from invocations.pytest import test, coverage
ns = Collection(test, coverage, release, blacken, docs)
ns.configure({"packaging": {"sign": True}})
|
from invoke import Collection
from invocations import docs
from invocations.checks import blacken
from invocations.packaging import release
from invocations.pytest import test, coverage
ns = Collection(test, coverage, release, blacken, docs)
ns.configure(
{"packaging": {"sign": True, "changelog_file": "docs/changelog.rst"}}
)
|
Set changelog_file for invocations release task, which now dry-runs ok
|
Set changelog_file for invocations release task, which now dry-runs ok
|
Python
|
bsd-2-clause
|
bitprophet/lexicon
|
from invoke import Collection
from invocations import docs
from invocations.checks import blacken
from invocations.packaging import release
from invocations.pytest import test, coverage
ns = Collection(test, coverage, release, blacken, docs)
ns.configure({"packaging": {"sign": True}})
Set changelog_file for invocations release task, which now dry-runs ok
|
from invoke import Collection
from invocations import docs
from invocations.checks import blacken
from invocations.packaging import release
from invocations.pytest import test, coverage
ns = Collection(test, coverage, release, blacken, docs)
ns.configure(
{"packaging": {"sign": True, "changelog_file": "docs/changelog.rst"}}
)
|
<commit_before>from invoke import Collection
from invocations import docs
from invocations.checks import blacken
from invocations.packaging import release
from invocations.pytest import test, coverage
ns = Collection(test, coverage, release, blacken, docs)
ns.configure({"packaging": {"sign": True}})
<commit_msg>Set changelog_file for invocations release task, which now dry-runs ok<commit_after>
|
from invoke import Collection
from invocations import docs
from invocations.checks import blacken
from invocations.packaging import release
from invocations.pytest import test, coverage
ns = Collection(test, coverage, release, blacken, docs)
ns.configure(
{"packaging": {"sign": True, "changelog_file": "docs/changelog.rst"}}
)
|
from invoke import Collection
from invocations import docs
from invocations.checks import blacken
from invocations.packaging import release
from invocations.pytest import test, coverage
ns = Collection(test, coverage, release, blacken, docs)
ns.configure({"packaging": {"sign": True}})
Set changelog_file for invocations release task, which now dry-runs okfrom invoke import Collection
from invocations import docs
from invocations.checks import blacken
from invocations.packaging import release
from invocations.pytest import test, coverage
ns = Collection(test, coverage, release, blacken, docs)
ns.configure(
{"packaging": {"sign": True, "changelog_file": "docs/changelog.rst"}}
)
|
<commit_before>from invoke import Collection
from invocations import docs
from invocations.checks import blacken
from invocations.packaging import release
from invocations.pytest import test, coverage
ns = Collection(test, coverage, release, blacken, docs)
ns.configure({"packaging": {"sign": True}})
<commit_msg>Set changelog_file for invocations release task, which now dry-runs ok<commit_after>from invoke import Collection
from invocations import docs
from invocations.checks import blacken
from invocations.packaging import release
from invocations.pytest import test, coverage
ns = Collection(test, coverage, release, blacken, docs)
ns.configure(
{"packaging": {"sign": True, "changelog_file": "docs/changelog.rst"}}
)
|
35fab0222543a2f32ef395bf6b622bad29533ceb
|
tests.py
|
tests.py
|
import unittest
from gtlaunch import Launcher
class MockOptions(object):
def __init__(self):
self.verbose = False
self.config = ''
self.project = ''
class LauncherTestCase(unittest.TestCase):
def setUp(self):
self.options = MockOptions()
def test_lazy_init(self):
launcher = Launcher(self.options, lazy=True)
self.assertIsNone(launcher.project)
def test_no_cwd(self):
project = {
'tabs': [],
}
launcher = Launcher(self.options, lazy=True)
args = launcher.build_args(project)
self.assertNotIn('--working-directory', args)
def test_cwd(self):
project = {
'cwd': '/home/test',
'tabs': [],
}
launcher = Launcher(self.options, lazy=True)
args = launcher.build_args(project)
idx = args.index('--working-directory')
self.assertEqual(args[idx + 1], project['cwd'])
def test_args_maximize(self):
project = {
'cwd': '~',
'tabs': [],
}
launcher = Launcher(self.options, lazy=True)
args = launcher.build_args(project)
self.assertIn('--maximize', args)
if __name__ == '__main__':
unittest.main()
|
import unittest
from gtlaunch import Launcher
class MockOptions(object):
def __init__(self):
self.verbose = False
self.config = ''
self.project = ''
class LauncherTestCase(unittest.TestCase):
def setUp(self):
self.options = MockOptions()
self.launcher = Launcher(self.options, lazy=True)
def test_lazy_init(self):
self.assertIsNone(self.launcher.project)
def test_no_cwd(self):
project = {
'tabs': [],
}
args = self.launcher.build_args(project)
self.assertNotIn('--working-directory', args)
def test_cwd(self):
project = {
'cwd': '/home/test',
'tabs': [],
}
args = self.launcher.build_args(project)
idx = args.index('--working-directory')
self.assertEqual(args[idx + 1], project['cwd'])
def test_args_maximize(self):
project = {
'cwd': '~',
'tabs': [],
}
args = self.launcher.build_args(project)
self.assertIn('--maximize', args)
if __name__ == '__main__':
unittest.main()
|
Create lazy launcher in setUp.
|
Create lazy launcher in setUp.
|
Python
|
mit
|
GoldenLine/gtlaunch
|
import unittest
from gtlaunch import Launcher
class MockOptions(object):
def __init__(self):
self.verbose = False
self.config = ''
self.project = ''
class LauncherTestCase(unittest.TestCase):
def setUp(self):
self.options = MockOptions()
def test_lazy_init(self):
launcher = Launcher(self.options, lazy=True)
self.assertIsNone(launcher.project)
def test_no_cwd(self):
project = {
'tabs': [],
}
launcher = Launcher(self.options, lazy=True)
args = launcher.build_args(project)
self.assertNotIn('--working-directory', args)
def test_cwd(self):
project = {
'cwd': '/home/test',
'tabs': [],
}
launcher = Launcher(self.options, lazy=True)
args = launcher.build_args(project)
idx = args.index('--working-directory')
self.assertEqual(args[idx + 1], project['cwd'])
def test_args_maximize(self):
project = {
'cwd': '~',
'tabs': [],
}
launcher = Launcher(self.options, lazy=True)
args = launcher.build_args(project)
self.assertIn('--maximize', args)
if __name__ == '__main__':
unittest.main()
Create lazy launcher in setUp.
|
import unittest
from gtlaunch import Launcher
class MockOptions(object):
def __init__(self):
self.verbose = False
self.config = ''
self.project = ''
class LauncherTestCase(unittest.TestCase):
def setUp(self):
self.options = MockOptions()
self.launcher = Launcher(self.options, lazy=True)
def test_lazy_init(self):
self.assertIsNone(self.launcher.project)
def test_no_cwd(self):
project = {
'tabs': [],
}
args = self.launcher.build_args(project)
self.assertNotIn('--working-directory', args)
def test_cwd(self):
project = {
'cwd': '/home/test',
'tabs': [],
}
args = self.launcher.build_args(project)
idx = args.index('--working-directory')
self.assertEqual(args[idx + 1], project['cwd'])
def test_args_maximize(self):
project = {
'cwd': '~',
'tabs': [],
}
args = self.launcher.build_args(project)
self.assertIn('--maximize', args)
if __name__ == '__main__':
unittest.main()
|
<commit_before>import unittest
from gtlaunch import Launcher
class MockOptions(object):
def __init__(self):
self.verbose = False
self.config = ''
self.project = ''
class LauncherTestCase(unittest.TestCase):
def setUp(self):
self.options = MockOptions()
def test_lazy_init(self):
launcher = Launcher(self.options, lazy=True)
self.assertIsNone(launcher.project)
def test_no_cwd(self):
project = {
'tabs': [],
}
launcher = Launcher(self.options, lazy=True)
args = launcher.build_args(project)
self.assertNotIn('--working-directory', args)
def test_cwd(self):
project = {
'cwd': '/home/test',
'tabs': [],
}
launcher = Launcher(self.options, lazy=True)
args = launcher.build_args(project)
idx = args.index('--working-directory')
self.assertEqual(args[idx + 1], project['cwd'])
def test_args_maximize(self):
project = {
'cwd': '~',
'tabs': [],
}
launcher = Launcher(self.options, lazy=True)
args = launcher.build_args(project)
self.assertIn('--maximize', args)
if __name__ == '__main__':
unittest.main()
<commit_msg>Create lazy launcher in setUp.<commit_after>
|
import unittest
from gtlaunch import Launcher
class MockOptions(object):
def __init__(self):
self.verbose = False
self.config = ''
self.project = ''
class LauncherTestCase(unittest.TestCase):
def setUp(self):
self.options = MockOptions()
self.launcher = Launcher(self.options, lazy=True)
def test_lazy_init(self):
self.assertIsNone(self.launcher.project)
def test_no_cwd(self):
project = {
'tabs': [],
}
args = self.launcher.build_args(project)
self.assertNotIn('--working-directory', args)
def test_cwd(self):
project = {
'cwd': '/home/test',
'tabs': [],
}
args = self.launcher.build_args(project)
idx = args.index('--working-directory')
self.assertEqual(args[idx + 1], project['cwd'])
def test_args_maximize(self):
project = {
'cwd': '~',
'tabs': [],
}
args = self.launcher.build_args(project)
self.assertIn('--maximize', args)
if __name__ == '__main__':
unittest.main()
|
import unittest
from gtlaunch import Launcher
class MockOptions(object):
def __init__(self):
self.verbose = False
self.config = ''
self.project = ''
class LauncherTestCase(unittest.TestCase):
def setUp(self):
self.options = MockOptions()
def test_lazy_init(self):
launcher = Launcher(self.options, lazy=True)
self.assertIsNone(launcher.project)
def test_no_cwd(self):
project = {
'tabs': [],
}
launcher = Launcher(self.options, lazy=True)
args = launcher.build_args(project)
self.assertNotIn('--working-directory', args)
def test_cwd(self):
project = {
'cwd': '/home/test',
'tabs': [],
}
launcher = Launcher(self.options, lazy=True)
args = launcher.build_args(project)
idx = args.index('--working-directory')
self.assertEqual(args[idx + 1], project['cwd'])
def test_args_maximize(self):
project = {
'cwd': '~',
'tabs': [],
}
launcher = Launcher(self.options, lazy=True)
args = launcher.build_args(project)
self.assertIn('--maximize', args)
if __name__ == '__main__':
unittest.main()
Create lazy launcher in setUp.import unittest
from gtlaunch import Launcher
class MockOptions(object):
def __init__(self):
self.verbose = False
self.config = ''
self.project = ''
class LauncherTestCase(unittest.TestCase):
def setUp(self):
self.options = MockOptions()
self.launcher = Launcher(self.options, lazy=True)
def test_lazy_init(self):
self.assertIsNone(self.launcher.project)
def test_no_cwd(self):
project = {
'tabs': [],
}
args = self.launcher.build_args(project)
self.assertNotIn('--working-directory', args)
def test_cwd(self):
project = {
'cwd': '/home/test',
'tabs': [],
}
args = self.launcher.build_args(project)
idx = args.index('--working-directory')
self.assertEqual(args[idx + 1], project['cwd'])
def test_args_maximize(self):
project = {
'cwd': '~',
'tabs': [],
}
args = self.launcher.build_args(project)
self.assertIn('--maximize', args)
if __name__ == '__main__':
unittest.main()
|
<commit_before>import unittest
from gtlaunch import Launcher
class MockOptions(object):
def __init__(self):
self.verbose = False
self.config = ''
self.project = ''
class LauncherTestCase(unittest.TestCase):
def setUp(self):
self.options = MockOptions()
def test_lazy_init(self):
launcher = Launcher(self.options, lazy=True)
self.assertIsNone(launcher.project)
def test_no_cwd(self):
project = {
'tabs': [],
}
launcher = Launcher(self.options, lazy=True)
args = launcher.build_args(project)
self.assertNotIn('--working-directory', args)
def test_cwd(self):
project = {
'cwd': '/home/test',
'tabs': [],
}
launcher = Launcher(self.options, lazy=True)
args = launcher.build_args(project)
idx = args.index('--working-directory')
self.assertEqual(args[idx + 1], project['cwd'])
def test_args_maximize(self):
project = {
'cwd': '~',
'tabs': [],
}
launcher = Launcher(self.options, lazy=True)
args = launcher.build_args(project)
self.assertIn('--maximize', args)
if __name__ == '__main__':
unittest.main()
<commit_msg>Create lazy launcher in setUp.<commit_after>import unittest
from gtlaunch import Launcher
class MockOptions(object):
def __init__(self):
self.verbose = False
self.config = ''
self.project = ''
class LauncherTestCase(unittest.TestCase):
def setUp(self):
self.options = MockOptions()
self.launcher = Launcher(self.options, lazy=True)
def test_lazy_init(self):
self.assertIsNone(self.launcher.project)
def test_no_cwd(self):
project = {
'tabs': [],
}
args = self.launcher.build_args(project)
self.assertNotIn('--working-directory', args)
def test_cwd(self):
project = {
'cwd': '/home/test',
'tabs': [],
}
args = self.launcher.build_args(project)
idx = args.index('--working-directory')
self.assertEqual(args[idx + 1], project['cwd'])
def test_args_maximize(self):
project = {
'cwd': '~',
'tabs': [],
}
args = self.launcher.build_args(project)
self.assertIn('--maximize', args)
if __name__ == '__main__':
unittest.main()
|
f7852806c3198d58162b66e18bfd9998ef33b63c
|
lexos/receivers/stats_receiver.py
|
lexos/receivers/stats_receiver.py
|
from lexos.receivers.base_receiver import BaseReceiver
class StatsReceiver(BaseReceiver):
def __init__(self):
"""So far there is no frontend option for statistics analysis"""
super().__init__()
def options_from_front_end(self):
"""So far there is no frontend option for statistics analysis"""
pass
|
from lexos.receivers.base_receiver import BaseReceiver
class StatsReceiver(BaseReceiver):
def __init__(self):
"""So far there is no frontend option for statistics analysis"""
super().__init__()
def options_from_front_end(self):
"""So far there is no frontend option for statistics analysis"""
raise NotImplementedError
|
Modify receiver to prevent using in future
|
Modify receiver to prevent using in future
|
Python
|
mit
|
WheatonCS/Lexos,WheatonCS/Lexos,WheatonCS/Lexos
|
from lexos.receivers.base_receiver import BaseReceiver
class StatsReceiver(BaseReceiver):
def __init__(self):
"""So far there is no frontend option for statistics analysis"""
super().__init__()
def options_from_front_end(self):
"""So far there is no frontend option for statistics analysis"""
pass
Modify receiver to prevent using in future
|
from lexos.receivers.base_receiver import BaseReceiver
class StatsReceiver(BaseReceiver):
def __init__(self):
"""So far there is no frontend option for statistics analysis"""
super().__init__()
def options_from_front_end(self):
"""So far there is no frontend option for statistics analysis"""
raise NotImplementedError
|
<commit_before>from lexos.receivers.base_receiver import BaseReceiver
class StatsReceiver(BaseReceiver):
def __init__(self):
"""So far there is no frontend option for statistics analysis"""
super().__init__()
def options_from_front_end(self):
"""So far there is no frontend option for statistics analysis"""
pass
<commit_msg>Modify receiver to prevent using in future<commit_after>
|
from lexos.receivers.base_receiver import BaseReceiver
class StatsReceiver(BaseReceiver):
def __init__(self):
"""So far there is no frontend option for statistics analysis"""
super().__init__()
def options_from_front_end(self):
"""So far there is no frontend option for statistics analysis"""
raise NotImplementedError
|
from lexos.receivers.base_receiver import BaseReceiver
class StatsReceiver(BaseReceiver):
def __init__(self):
"""So far there is no frontend option for statistics analysis"""
super().__init__()
def options_from_front_end(self):
"""So far there is no frontend option for statistics analysis"""
pass
Modify receiver to prevent using in futurefrom lexos.receivers.base_receiver import BaseReceiver
class StatsReceiver(BaseReceiver):
def __init__(self):
"""So far there is no frontend option for statistics analysis"""
super().__init__()
def options_from_front_end(self):
"""So far there is no frontend option for statistics analysis"""
raise NotImplementedError
|
<commit_before>from lexos.receivers.base_receiver import BaseReceiver
class StatsReceiver(BaseReceiver):
def __init__(self):
"""So far there is no frontend option for statistics analysis"""
super().__init__()
def options_from_front_end(self):
"""So far there is no frontend option for statistics analysis"""
pass
<commit_msg>Modify receiver to prevent using in future<commit_after>from lexos.receivers.base_receiver import BaseReceiver
class StatsReceiver(BaseReceiver):
def __init__(self):
"""So far there is no frontend option for statistics analysis"""
super().__init__()
def options_from_front_end(self):
"""So far there is no frontend option for statistics analysis"""
raise NotImplementedError
|
addee67fbf46a795c9de4669c9951c84b6590d98
|
cartoframes/context/base_context.py
|
cartoframes/context/base_context.py
|
from abc import ABCMeta, abstractmethod
class BaseContext():
__metaclass__ = ABCMeta
@abstractmethod
def download(self):
pass
@abstractmethod
def upload(self):
pass
@abstractmethod
def execute_query(self):
pass
@abstractmethod
def execute_long_running_query(self):
pass
|
from abc import ABCMeta, abstractmethod
class BaseContext():
__metaclass__ = ABCMeta
@abstractmethod
def download(self, query, retry_times=0):
pass
@abstractmethod
def upload(self, query, data):
pass
@abstractmethod
def execute_query(self, query, parse_json=True, do_post=True, format=None, **request_args):
pass
@abstractmethod
def execute_long_running_query(self, query):
pass
|
Add params in BaseContext abtract methods
|
Add params in BaseContext abtract methods
|
Python
|
bsd-3-clause
|
CartoDB/cartoframes,CartoDB/cartoframes
|
from abc import ABCMeta, abstractmethod
class BaseContext():
__metaclass__ = ABCMeta
@abstractmethod
def download(self):
pass
@abstractmethod
def upload(self):
pass
@abstractmethod
def execute_query(self):
pass
@abstractmethod
def execute_long_running_query(self):
pass
Add params in BaseContext abtract methods
|
from abc import ABCMeta, abstractmethod
class BaseContext():
__metaclass__ = ABCMeta
@abstractmethod
def download(self, query, retry_times=0):
pass
@abstractmethod
def upload(self, query, data):
pass
@abstractmethod
def execute_query(self, query, parse_json=True, do_post=True, format=None, **request_args):
pass
@abstractmethod
def execute_long_running_query(self, query):
pass
|
<commit_before>from abc import ABCMeta, abstractmethod
class BaseContext():
__metaclass__ = ABCMeta
@abstractmethod
def download(self):
pass
@abstractmethod
def upload(self):
pass
@abstractmethod
def execute_query(self):
pass
@abstractmethod
def execute_long_running_query(self):
pass
<commit_msg>Add params in BaseContext abtract methods<commit_after>
|
from abc import ABCMeta, abstractmethod
class BaseContext():
__metaclass__ = ABCMeta
@abstractmethod
def download(self, query, retry_times=0):
pass
@abstractmethod
def upload(self, query, data):
pass
@abstractmethod
def execute_query(self, query, parse_json=True, do_post=True, format=None, **request_args):
pass
@abstractmethod
def execute_long_running_query(self, query):
pass
|
from abc import ABCMeta, abstractmethod
class BaseContext():
__metaclass__ = ABCMeta
@abstractmethod
def download(self):
pass
@abstractmethod
def upload(self):
pass
@abstractmethod
def execute_query(self):
pass
@abstractmethod
def execute_long_running_query(self):
pass
Add params in BaseContext abtract methodsfrom abc import ABCMeta, abstractmethod
class BaseContext():
__metaclass__ = ABCMeta
@abstractmethod
def download(self, query, retry_times=0):
pass
@abstractmethod
def upload(self, query, data):
pass
@abstractmethod
def execute_query(self, query, parse_json=True, do_post=True, format=None, **request_args):
pass
@abstractmethod
def execute_long_running_query(self, query):
pass
|
<commit_before>from abc import ABCMeta, abstractmethod
class BaseContext():
__metaclass__ = ABCMeta
@abstractmethod
def download(self):
pass
@abstractmethod
def upload(self):
pass
@abstractmethod
def execute_query(self):
pass
@abstractmethod
def execute_long_running_query(self):
pass
<commit_msg>Add params in BaseContext abtract methods<commit_after>from abc import ABCMeta, abstractmethod
class BaseContext():
__metaclass__ = ABCMeta
@abstractmethod
def download(self, query, retry_times=0):
pass
@abstractmethod
def upload(self, query, data):
pass
@abstractmethod
def execute_query(self, query, parse_json=True, do_post=True, format=None, **request_args):
pass
@abstractmethod
def execute_long_running_query(self, query):
pass
|
b084e02dd2cf7b492c69090b6acd548066c7c34f
|
pos_picking_state_fix/models/pos_picking.py
|
pos_picking_state_fix/models/pos_picking.py
|
# -*- coding: utf-8 -*-
# See README file for full copyright and licensing details.
from openerp import models, api
class PosPicking(models.Model):
_inherit = 'pos.order'
@api.multi
def create_picking(self):
try:
super(PosPicking, self).create_picking()
except:
# Cancel move lines
if self.picking_id.state != 'done':
for move in self.picking_id.move_lines:
move.do_unreserve()
return True
|
# -*- coding: utf-8 -*-
# See README file for full copyright and licensing details.
import time
from openerp import models, api
from openerp.tools import DEFAULT_SERVER_DATETIME_FORMAT
class PosPicking(models.Model):
_inherit = 'pos.order'
@api.multi
def create_picking(self):
try:
super(PosPicking, self).create_picking()
except:
if self.picking_id.state != 'done':
for move in self.picking_id.move_lines:
if move.quant_ids:
# We pass this move to done because the quants were already moved
move.write({'state': 'done', 'date': time.strftime(DEFAULT_SERVER_DATETIME_FORMAT)})
else:
# If there are no moved quants we pass the move to Waiting Availability
move.do_unreserve()
return True
|
Check if quants are moved and pass moves to done to avoid duplication
|
[FIX] Check if quants are moved and pass moves to done to avoid duplication
|
Python
|
agpl-3.0
|
rgbconsulting/rgb-addons,rgbconsulting/rgb-pos,rgbconsulting/rgb-pos,rgbconsulting/rgb-addons
|
# -*- coding: utf-8 -*-
# See README file for full copyright and licensing details.
from openerp import models, api
class PosPicking(models.Model):
_inherit = 'pos.order'
@api.multi
def create_picking(self):
try:
super(PosPicking, self).create_picking()
except:
# Cancel move lines
if self.picking_id.state != 'done':
for move in self.picking_id.move_lines:
move.do_unreserve()
return True
[FIX] Check if quants are moved and pass moves to done to avoid duplication
|
# -*- coding: utf-8 -*-
# See README file for full copyright and licensing details.
import time
from openerp import models, api
from openerp.tools import DEFAULT_SERVER_DATETIME_FORMAT
class PosPicking(models.Model):
_inherit = 'pos.order'
@api.multi
def create_picking(self):
try:
super(PosPicking, self).create_picking()
except:
if self.picking_id.state != 'done':
for move in self.picking_id.move_lines:
if move.quant_ids:
# We pass this move to done because the quants were already moved
move.write({'state': 'done', 'date': time.strftime(DEFAULT_SERVER_DATETIME_FORMAT)})
else:
# If there are no moved quants we pass the move to Waiting Availability
move.do_unreserve()
return True
|
<commit_before># -*- coding: utf-8 -*-
# See README file for full copyright and licensing details.
from openerp import models, api
class PosPicking(models.Model):
_inherit = 'pos.order'
@api.multi
def create_picking(self):
try:
super(PosPicking, self).create_picking()
except:
# Cancel move lines
if self.picking_id.state != 'done':
for move in self.picking_id.move_lines:
move.do_unreserve()
return True
<commit_msg>[FIX] Check if quants are moved and pass moves to done to avoid duplication<commit_after>
|
# -*- coding: utf-8 -*-
# See README file for full copyright and licensing details.
import time
from openerp import models, api
from openerp.tools import DEFAULT_SERVER_DATETIME_FORMAT
class PosPicking(models.Model):
_inherit = 'pos.order'
@api.multi
def create_picking(self):
try:
super(PosPicking, self).create_picking()
except:
if self.picking_id.state != 'done':
for move in self.picking_id.move_lines:
if move.quant_ids:
# We pass this move to done because the quants were already moved
move.write({'state': 'done', 'date': time.strftime(DEFAULT_SERVER_DATETIME_FORMAT)})
else:
# If there are no moved quants we pass the move to Waiting Availability
move.do_unreserve()
return True
|
# -*- coding: utf-8 -*-
# See README file for full copyright and licensing details.
from openerp import models, api
class PosPicking(models.Model):
_inherit = 'pos.order'
@api.multi
def create_picking(self):
try:
super(PosPicking, self).create_picking()
except:
# Cancel move lines
if self.picking_id.state != 'done':
for move in self.picking_id.move_lines:
move.do_unreserve()
return True
[FIX] Check if quants are moved and pass moves to done to avoid duplication# -*- coding: utf-8 -*-
# See README file for full copyright and licensing details.
import time
from openerp import models, api
from openerp.tools import DEFAULT_SERVER_DATETIME_FORMAT
class PosPicking(models.Model):
_inherit = 'pos.order'
@api.multi
def create_picking(self):
try:
super(PosPicking, self).create_picking()
except:
if self.picking_id.state != 'done':
for move in self.picking_id.move_lines:
if move.quant_ids:
# We pass this move to done because the quants were already moved
move.write({'state': 'done', 'date': time.strftime(DEFAULT_SERVER_DATETIME_FORMAT)})
else:
# If there are no moved quants we pass the move to Waiting Availability
move.do_unreserve()
return True
|
<commit_before># -*- coding: utf-8 -*-
# See README file for full copyright and licensing details.
from openerp import models, api
class PosPicking(models.Model):
_inherit = 'pos.order'
@api.multi
def create_picking(self):
try:
super(PosPicking, self).create_picking()
except:
# Cancel move lines
if self.picking_id.state != 'done':
for move in self.picking_id.move_lines:
move.do_unreserve()
return True
<commit_msg>[FIX] Check if quants are moved and pass moves to done to avoid duplication<commit_after># -*- coding: utf-8 -*-
# See README file for full copyright and licensing details.
import time
from openerp import models, api
from openerp.tools import DEFAULT_SERVER_DATETIME_FORMAT
class PosPicking(models.Model):
_inherit = 'pos.order'
@api.multi
def create_picking(self):
try:
super(PosPicking, self).create_picking()
except:
if self.picking_id.state != 'done':
for move in self.picking_id.move_lines:
if move.quant_ids:
# We pass this move to done because the quants were already moved
move.write({'state': 'done', 'date': time.strftime(DEFAULT_SERVER_DATETIME_FORMAT)})
else:
# If there are no moved quants we pass the move to Waiting Availability
move.do_unreserve()
return True
|
5b011488b5fcfd17f2029e833b757d24d437908e
|
document_page_project/__manifest__.py
|
document_page_project/__manifest__.py
|
# Copyright 2019 ForgeFlow S.L. (https://www.forgeflow.com) - Lois Rilo
# License AGPL-3.0 or later (https://www.gnu.org/licenses/agpl.html).
{
"name": "Document Page Project",
"summary": "This module links document pages to projects",
"version": "13.0.1.0.1",
"development_status": "Production/Stable",
"category": "Project",
"author": "ForgeFlow, Odoo Community Association (OCA)",
"website": "https://github.com/OCA/knowledge",
"license": "AGPL-3",
"depends": ["project", "document_page"],
"data": ["views/document_page_views.xml", "views/project_project_views.xml"],
"installable": True,
}
|
# Copyright 2019 ForgeFlow S.L. (https://www.forgeflow.com) - Lois Rilo
# License AGPL-3.0 or later (https://www.gnu.org/licenses/agpl.html).
{
"name": "Document Page Project",
"summary": "This module links document pages to projects",
"version": "13.0.1.0.1",
"development_status": "Beta",
"category": "Project",
"author": "ForgeFlow, Odoo Community Association (OCA)",
"website": "https://github.com/OCA/knowledge",
"license": "AGPL-3",
"depends": ["project", "document_page"],
"data": ["views/document_page_views.xml", "views/project_project_views.xml"],
"installable": True,
}
|
Revert to Beta as document_page is Beta
|
[REV] document_page_project: Revert to Beta as document_page is Beta
|
Python
|
agpl-3.0
|
OCA/knowledge,OCA/knowledge,OCA/knowledge
|
# Copyright 2019 ForgeFlow S.L. (https://www.forgeflow.com) - Lois Rilo
# License AGPL-3.0 or later (https://www.gnu.org/licenses/agpl.html).
{
"name": "Document Page Project",
"summary": "This module links document pages to projects",
"version": "13.0.1.0.1",
"development_status": "Production/Stable",
"category": "Project",
"author": "ForgeFlow, Odoo Community Association (OCA)",
"website": "https://github.com/OCA/knowledge",
"license": "AGPL-3",
"depends": ["project", "document_page"],
"data": ["views/document_page_views.xml", "views/project_project_views.xml"],
"installable": True,
}
[REV] document_page_project: Revert to Beta as document_page is Beta
|
# Copyright 2019 ForgeFlow S.L. (https://www.forgeflow.com) - Lois Rilo
# License AGPL-3.0 or later (https://www.gnu.org/licenses/agpl.html).
{
"name": "Document Page Project",
"summary": "This module links document pages to projects",
"version": "13.0.1.0.1",
"development_status": "Beta",
"category": "Project",
"author": "ForgeFlow, Odoo Community Association (OCA)",
"website": "https://github.com/OCA/knowledge",
"license": "AGPL-3",
"depends": ["project", "document_page"],
"data": ["views/document_page_views.xml", "views/project_project_views.xml"],
"installable": True,
}
|
<commit_before># Copyright 2019 ForgeFlow S.L. (https://www.forgeflow.com) - Lois Rilo
# License AGPL-3.0 or later (https://www.gnu.org/licenses/agpl.html).
{
"name": "Document Page Project",
"summary": "This module links document pages to projects",
"version": "13.0.1.0.1",
"development_status": "Production/Stable",
"category": "Project",
"author": "ForgeFlow, Odoo Community Association (OCA)",
"website": "https://github.com/OCA/knowledge",
"license": "AGPL-3",
"depends": ["project", "document_page"],
"data": ["views/document_page_views.xml", "views/project_project_views.xml"],
"installable": True,
}
<commit_msg>[REV] document_page_project: Revert to Beta as document_page is Beta<commit_after>
|
# Copyright 2019 ForgeFlow S.L. (https://www.forgeflow.com) - Lois Rilo
# License AGPL-3.0 or later (https://www.gnu.org/licenses/agpl.html).
{
"name": "Document Page Project",
"summary": "This module links document pages to projects",
"version": "13.0.1.0.1",
"development_status": "Beta",
"category": "Project",
"author": "ForgeFlow, Odoo Community Association (OCA)",
"website": "https://github.com/OCA/knowledge",
"license": "AGPL-3",
"depends": ["project", "document_page"],
"data": ["views/document_page_views.xml", "views/project_project_views.xml"],
"installable": True,
}
|
# Copyright 2019 ForgeFlow S.L. (https://www.forgeflow.com) - Lois Rilo
# License AGPL-3.0 or later (https://www.gnu.org/licenses/agpl.html).
{
"name": "Document Page Project",
"summary": "This module links document pages to projects",
"version": "13.0.1.0.1",
"development_status": "Production/Stable",
"category": "Project",
"author": "ForgeFlow, Odoo Community Association (OCA)",
"website": "https://github.com/OCA/knowledge",
"license": "AGPL-3",
"depends": ["project", "document_page"],
"data": ["views/document_page_views.xml", "views/project_project_views.xml"],
"installable": True,
}
[REV] document_page_project: Revert to Beta as document_page is Beta# Copyright 2019 ForgeFlow S.L. (https://www.forgeflow.com) - Lois Rilo
# License AGPL-3.0 or later (https://www.gnu.org/licenses/agpl.html).
{
"name": "Document Page Project",
"summary": "This module links document pages to projects",
"version": "13.0.1.0.1",
"development_status": "Beta",
"category": "Project",
"author": "ForgeFlow, Odoo Community Association (OCA)",
"website": "https://github.com/OCA/knowledge",
"license": "AGPL-3",
"depends": ["project", "document_page"],
"data": ["views/document_page_views.xml", "views/project_project_views.xml"],
"installable": True,
}
|
<commit_before># Copyright 2019 ForgeFlow S.L. (https://www.forgeflow.com) - Lois Rilo
# License AGPL-3.0 or later (https://www.gnu.org/licenses/agpl.html).
{
"name": "Document Page Project",
"summary": "This module links document pages to projects",
"version": "13.0.1.0.1",
"development_status": "Production/Stable",
"category": "Project",
"author": "ForgeFlow, Odoo Community Association (OCA)",
"website": "https://github.com/OCA/knowledge",
"license": "AGPL-3",
"depends": ["project", "document_page"],
"data": ["views/document_page_views.xml", "views/project_project_views.xml"],
"installable": True,
}
<commit_msg>[REV] document_page_project: Revert to Beta as document_page is Beta<commit_after># Copyright 2019 ForgeFlow S.L. (https://www.forgeflow.com) - Lois Rilo
# License AGPL-3.0 or later (https://www.gnu.org/licenses/agpl.html).
{
"name": "Document Page Project",
"summary": "This module links document pages to projects",
"version": "13.0.1.0.1",
"development_status": "Beta",
"category": "Project",
"author": "ForgeFlow, Odoo Community Association (OCA)",
"website": "https://github.com/OCA/knowledge",
"license": "AGPL-3",
"depends": ["project", "document_page"],
"data": ["views/document_page_views.xml", "views/project_project_views.xml"],
"installable": True,
}
|
f40bf1441121c138877e27bd23bcef73cf5c2cef
|
cisco_olt_http/tests/test_operations.py
|
cisco_olt_http/tests/test_operations.py
|
import os
import pytest
import requests
from cisco_olt_http import operations
from cisco_olt_http.client import Client
@pytest.fixture
def data_dir():
return os.path.abspath(
os.path.join(os.path.dirname(__file__), 'data'))
def test_get_data():
client = Client('http://base-url')
show_equipment_op = operations.ShowEquipmentOp(client)
op_data = show_equipment_op.get_data()
assert op_data
class TestOperationResult:
def test_ok_response(self, data_dir, mocker):
response = mocker.Mock(autospec=requests.Response)
with open(os.path.join(data_dir, 'ok_response.xml')) as of:
response.content = of.read()
operation_result = operations.OperationResult(response)
assert not operation_result.error
assert operation_result.error_str == 'OK'
|
import os
import pytest
import requests
from cisco_olt_http import operations
from cisco_olt_http.client import Client
@pytest.fixture
def data_dir():
return os.path.abspath(
os.path.join(os.path.dirname(__file__), 'data'))
@pytest.fixture
def ok_response(data_dir, mocker):
response = mocker.Mock(autospec=requests.Response)
with open(os.path.join(data_dir, 'ok_response.xml')) as of:
response.content = of.read()
return response
def test_get_data():
client = Client('http://base-url')
show_equipment_op = operations.ShowEquipmentOp(client)
op_data = show_equipment_op.get_data()
assert op_data
class TestOperationResult:
def test_ok_response(self, ok_response):
operation_result = operations.OperationResult(ok_response)
assert not operation_result.error
assert operation_result.error_str == 'OK'
|
Move ok response creation to pytest fixture
|
Move ok response creation to pytest fixture
|
Python
|
mit
|
Vnet-as/cisco-olt-http-client,beezz/cisco-olt-http-client
|
import os
import pytest
import requests
from cisco_olt_http import operations
from cisco_olt_http.client import Client
@pytest.fixture
def data_dir():
return os.path.abspath(
os.path.join(os.path.dirname(__file__), 'data'))
def test_get_data():
client = Client('http://base-url')
show_equipment_op = operations.ShowEquipmentOp(client)
op_data = show_equipment_op.get_data()
assert op_data
class TestOperationResult:
def test_ok_response(self, data_dir, mocker):
response = mocker.Mock(autospec=requests.Response)
with open(os.path.join(data_dir, 'ok_response.xml')) as of:
response.content = of.read()
operation_result = operations.OperationResult(response)
assert not operation_result.error
assert operation_result.error_str == 'OK'
Move ok response creation to pytest fixture
|
import os
import pytest
import requests
from cisco_olt_http import operations
from cisco_olt_http.client import Client
@pytest.fixture
def data_dir():
return os.path.abspath(
os.path.join(os.path.dirname(__file__), 'data'))
@pytest.fixture
def ok_response(data_dir, mocker):
response = mocker.Mock(autospec=requests.Response)
with open(os.path.join(data_dir, 'ok_response.xml')) as of:
response.content = of.read()
return response
def test_get_data():
client = Client('http://base-url')
show_equipment_op = operations.ShowEquipmentOp(client)
op_data = show_equipment_op.get_data()
assert op_data
class TestOperationResult:
def test_ok_response(self, ok_response):
operation_result = operations.OperationResult(ok_response)
assert not operation_result.error
assert operation_result.error_str == 'OK'
|
<commit_before>
import os
import pytest
import requests
from cisco_olt_http import operations
from cisco_olt_http.client import Client
@pytest.fixture
def data_dir():
return os.path.abspath(
os.path.join(os.path.dirname(__file__), 'data'))
def test_get_data():
client = Client('http://base-url')
show_equipment_op = operations.ShowEquipmentOp(client)
op_data = show_equipment_op.get_data()
assert op_data
class TestOperationResult:
def test_ok_response(self, data_dir, mocker):
response = mocker.Mock(autospec=requests.Response)
with open(os.path.join(data_dir, 'ok_response.xml')) as of:
response.content = of.read()
operation_result = operations.OperationResult(response)
assert not operation_result.error
assert operation_result.error_str == 'OK'
<commit_msg>Move ok response creation to pytest fixture<commit_after>
|
import os
import pytest
import requests
from cisco_olt_http import operations
from cisco_olt_http.client import Client
@pytest.fixture
def data_dir():
return os.path.abspath(
os.path.join(os.path.dirname(__file__), 'data'))
@pytest.fixture
def ok_response(data_dir, mocker):
response = mocker.Mock(autospec=requests.Response)
with open(os.path.join(data_dir, 'ok_response.xml')) as of:
response.content = of.read()
return response
def test_get_data():
client = Client('http://base-url')
show_equipment_op = operations.ShowEquipmentOp(client)
op_data = show_equipment_op.get_data()
assert op_data
class TestOperationResult:
def test_ok_response(self, ok_response):
operation_result = operations.OperationResult(ok_response)
assert not operation_result.error
assert operation_result.error_str == 'OK'
|
import os
import pytest
import requests
from cisco_olt_http import operations
from cisco_olt_http.client import Client
@pytest.fixture
def data_dir():
return os.path.abspath(
os.path.join(os.path.dirname(__file__), 'data'))
def test_get_data():
client = Client('http://base-url')
show_equipment_op = operations.ShowEquipmentOp(client)
op_data = show_equipment_op.get_data()
assert op_data
class TestOperationResult:
def test_ok_response(self, data_dir, mocker):
response = mocker.Mock(autospec=requests.Response)
with open(os.path.join(data_dir, 'ok_response.xml')) as of:
response.content = of.read()
operation_result = operations.OperationResult(response)
assert not operation_result.error
assert operation_result.error_str == 'OK'
Move ok response creation to pytest fixture
import os
import pytest
import requests
from cisco_olt_http import operations
from cisco_olt_http.client import Client
@pytest.fixture
def data_dir():
return os.path.abspath(
os.path.join(os.path.dirname(__file__), 'data'))
@pytest.fixture
def ok_response(data_dir, mocker):
response = mocker.Mock(autospec=requests.Response)
with open(os.path.join(data_dir, 'ok_response.xml')) as of:
response.content = of.read()
return response
def test_get_data():
client = Client('http://base-url')
show_equipment_op = operations.ShowEquipmentOp(client)
op_data = show_equipment_op.get_data()
assert op_data
class TestOperationResult:
def test_ok_response(self, ok_response):
operation_result = operations.OperationResult(ok_response)
assert not operation_result.error
assert operation_result.error_str == 'OK'
|
<commit_before>
import os
import pytest
import requests
from cisco_olt_http import operations
from cisco_olt_http.client import Client
@pytest.fixture
def data_dir():
return os.path.abspath(
os.path.join(os.path.dirname(__file__), 'data'))
def test_get_data():
client = Client('http://base-url')
show_equipment_op = operations.ShowEquipmentOp(client)
op_data = show_equipment_op.get_data()
assert op_data
class TestOperationResult:
def test_ok_response(self, data_dir, mocker):
response = mocker.Mock(autospec=requests.Response)
with open(os.path.join(data_dir, 'ok_response.xml')) as of:
response.content = of.read()
operation_result = operations.OperationResult(response)
assert not operation_result.error
assert operation_result.error_str == 'OK'
<commit_msg>Move ok response creation to pytest fixture<commit_after>
import os
import pytest
import requests
from cisco_olt_http import operations
from cisco_olt_http.client import Client
@pytest.fixture
def data_dir():
return os.path.abspath(
os.path.join(os.path.dirname(__file__), 'data'))
@pytest.fixture
def ok_response(data_dir, mocker):
response = mocker.Mock(autospec=requests.Response)
with open(os.path.join(data_dir, 'ok_response.xml')) as of:
response.content = of.read()
return response
def test_get_data():
client = Client('http://base-url')
show_equipment_op = operations.ShowEquipmentOp(client)
op_data = show_equipment_op.get_data()
assert op_data
class TestOperationResult:
def test_ok_response(self, ok_response):
operation_result = operations.OperationResult(ok_response)
assert not operation_result.error
assert operation_result.error_str == 'OK'
|
19af4b5c8c849750dd0885ea4fcfb651545b7985
|
migrations/002_add_month_start.py
|
migrations/002_add_month_start.py
|
"""
Add _week_start_at field to all documents in all collections
"""
from backdrop.core.bucket import utc
from backdrop.core.records import Record
import logging
log = logging.getLogger(__name__)
def up(db):
for name in db.collection_names():
log.info("Migrating collection: {0}".format(name))
collection = db[name]
query = {
"_timestamp": {"$exists": True},
"_month_start_at": {"$exists": False}
}
for document in collection.find(query):
document['_timestamp'] = utc(document['_timestamp'])
if '_week_start_at' in document:
document.pop('_week_start_at')
record = Record(document)
collection.save(record.to_mongo())
|
"""
Add _week_start_at field to all documents in all collections
"""
from backdrop.core.bucket import utc
from backdrop.core.records import Record
import logging
log = logging.getLogger(__name__)
def up(db):
for name in db.collection_names():
log.info("Migrating collection: {0}".format(name))
collection = db[name]
query = {
"_timestamp": {"$exists": True},
"_month_start_at": {"$exists": False}
}
for document in collection.find(query):
document['_timestamp'] = utc(document['_timestamp'])
if '_week_start_at' in document:
document.pop('_week_start_at')
if '_updated_at' in document:
document.pop('_updated_at')
record = Record(document)
collection.save(record.to_mongo())
|
Remove disallowed fields before resaving on migrations.
|
Remove disallowed fields before resaving on migrations.
- TODO: fix this properly.
|
Python
|
mit
|
alphagov/backdrop,alphagov/backdrop,alphagov/backdrop
|
"""
Add _week_start_at field to all documents in all collections
"""
from backdrop.core.bucket import utc
from backdrop.core.records import Record
import logging
log = logging.getLogger(__name__)
def up(db):
for name in db.collection_names():
log.info("Migrating collection: {0}".format(name))
collection = db[name]
query = {
"_timestamp": {"$exists": True},
"_month_start_at": {"$exists": False}
}
for document in collection.find(query):
document['_timestamp'] = utc(document['_timestamp'])
if '_week_start_at' in document:
document.pop('_week_start_at')
record = Record(document)
collection.save(record.to_mongo())
Remove disallowed fields before resaving on migrations.
- TODO: fix this properly.
|
"""
Add _week_start_at field to all documents in all collections
"""
from backdrop.core.bucket import utc
from backdrop.core.records import Record
import logging
log = logging.getLogger(__name__)
def up(db):
for name in db.collection_names():
log.info("Migrating collection: {0}".format(name))
collection = db[name]
query = {
"_timestamp": {"$exists": True},
"_month_start_at": {"$exists": False}
}
for document in collection.find(query):
document['_timestamp'] = utc(document['_timestamp'])
if '_week_start_at' in document:
document.pop('_week_start_at')
if '_updated_at' in document:
document.pop('_updated_at')
record = Record(document)
collection.save(record.to_mongo())
|
<commit_before>"""
Add _week_start_at field to all documents in all collections
"""
from backdrop.core.bucket import utc
from backdrop.core.records import Record
import logging
log = logging.getLogger(__name__)
def up(db):
for name in db.collection_names():
log.info("Migrating collection: {0}".format(name))
collection = db[name]
query = {
"_timestamp": {"$exists": True},
"_month_start_at": {"$exists": False}
}
for document in collection.find(query):
document['_timestamp'] = utc(document['_timestamp'])
if '_week_start_at' in document:
document.pop('_week_start_at')
record = Record(document)
collection.save(record.to_mongo())
<commit_msg>Remove disallowed fields before resaving on migrations.
- TODO: fix this properly.<commit_after>
|
"""
Add _week_start_at field to all documents in all collections
"""
from backdrop.core.bucket import utc
from backdrop.core.records import Record
import logging
log = logging.getLogger(__name__)
def up(db):
for name in db.collection_names():
log.info("Migrating collection: {0}".format(name))
collection = db[name]
query = {
"_timestamp": {"$exists": True},
"_month_start_at": {"$exists": False}
}
for document in collection.find(query):
document['_timestamp'] = utc(document['_timestamp'])
if '_week_start_at' in document:
document.pop('_week_start_at')
if '_updated_at' in document:
document.pop('_updated_at')
record = Record(document)
collection.save(record.to_mongo())
|
"""
Add _week_start_at field to all documents in all collections
"""
from backdrop.core.bucket import utc
from backdrop.core.records import Record
import logging
log = logging.getLogger(__name__)
def up(db):
for name in db.collection_names():
log.info("Migrating collection: {0}".format(name))
collection = db[name]
query = {
"_timestamp": {"$exists": True},
"_month_start_at": {"$exists": False}
}
for document in collection.find(query):
document['_timestamp'] = utc(document['_timestamp'])
if '_week_start_at' in document:
document.pop('_week_start_at')
record = Record(document)
collection.save(record.to_mongo())
Remove disallowed fields before resaving on migrations.
- TODO: fix this properly."""
Add _week_start_at field to all documents in all collections
"""
from backdrop.core.bucket import utc
from backdrop.core.records import Record
import logging
log = logging.getLogger(__name__)
def up(db):
for name in db.collection_names():
log.info("Migrating collection: {0}".format(name))
collection = db[name]
query = {
"_timestamp": {"$exists": True},
"_month_start_at": {"$exists": False}
}
for document in collection.find(query):
document['_timestamp'] = utc(document['_timestamp'])
if '_week_start_at' in document:
document.pop('_week_start_at')
if '_updated_at' in document:
document.pop('_updated_at')
record = Record(document)
collection.save(record.to_mongo())
|
<commit_before>"""
Add _week_start_at field to all documents in all collections
"""
from backdrop.core.bucket import utc
from backdrop.core.records import Record
import logging
log = logging.getLogger(__name__)
def up(db):
for name in db.collection_names():
log.info("Migrating collection: {0}".format(name))
collection = db[name]
query = {
"_timestamp": {"$exists": True},
"_month_start_at": {"$exists": False}
}
for document in collection.find(query):
document['_timestamp'] = utc(document['_timestamp'])
if '_week_start_at' in document:
document.pop('_week_start_at')
record = Record(document)
collection.save(record.to_mongo())
<commit_msg>Remove disallowed fields before resaving on migrations.
- TODO: fix this properly.<commit_after>"""
Add _week_start_at field to all documents in all collections
"""
from backdrop.core.bucket import utc
from backdrop.core.records import Record
import logging
log = logging.getLogger(__name__)
def up(db):
for name in db.collection_names():
log.info("Migrating collection: {0}".format(name))
collection = db[name]
query = {
"_timestamp": {"$exists": True},
"_month_start_at": {"$exists": False}
}
for document in collection.find(query):
document['_timestamp'] = utc(document['_timestamp'])
if '_week_start_at' in document:
document.pop('_week_start_at')
if '_updated_at' in document:
document.pop('_updated_at')
record = Record(document)
collection.save(record.to_mongo())
|
de5c0c9107156a073670d68fcb04e575e08f9b80
|
sympy/__init__.py
|
sympy/__init__.py
|
__version__ = "0.5.0"
from sympy.core import *
from series import *
from simplify import *
from solvers import *
from matrices import *
from geometry import *
from polynomials import *
from utilities import *
#from specfun import *
from integrals import *
try:
from plotting import Plot
except ImportError, e:
print str(e)
|
__version__ = "0.5.0"
from sympy.core import *
from series import *
from simplify import *
from solvers import *
from matrices import *
from geometry import *
from polynomials import *
from utilities import *
#from specfun import *
from integrals import *
try:
from plotting import Plot
except ImportError, e:
class Plot(object):
def __init__(*args, **kwargs):
raise e
|
Hide ctypes import error until Plot() is called.
|
Hide ctypes import error until Plot() is called.
|
Python
|
bsd-3-clause
|
kmacinnis/sympy,Curious72/sympy,meghana1995/sympy,MechCoder/sympy,saurabhjn76/sympy,VaibhavAgarwalVA/sympy,Designist/sympy,lidavidm/sympy,skidzo/sympy,beni55/sympy,Davidjohnwilson/sympy,jaimahajan1997/sympy,kmacinnis/sympy,MridulS/sympy,mcdaniel67/sympy,jbbskinny/sympy,pandeyadarsh/sympy,jaimahajan1997/sympy,mattpap/sympy-polys,lindsayad/sympy,souravsingh/sympy,ryanGT/sympy,Sumith1896/sympy,tovrstra/sympy,moble/sympy,emon10005/sympy,kaushik94/sympy,mafiya69/sympy,emon10005/sympy,Vishluck/sympy,Arafatk/sympy,MridulS/sympy,debugger22/sympy,madan96/sympy,Shaswat27/sympy,Titan-C/sympy,shikil/sympy,rahuldan/sympy,vipulroxx/sympy,toolforger/sympy,shipci/sympy,KevinGoodsell/sympy,cswiercz/sympy,abhiii5459/sympy,Gadal/sympy,jbaayen/sympy,ChristinaZografou/sympy,iamutkarshtiwari/sympy,rahuldan/sympy,kmacinnis/sympy,sahilshekhawat/sympy,asm666/sympy,chaffra/sympy,kevalds51/sympy,maniteja123/sympy,sunny94/temp,vipulroxx/sympy,drufat/sympy,shikil/sympy,saurabhjn76/sympy,mafiya69/sympy,atreyv/sympy,Curious72/sympy,hrashk/sympy,Mitchkoens/sympy,dqnykamp/sympy,ahhda/sympy,liangjiaxing/sympy,madan96/sympy,lindsayad/sympy,iamutkarshtiwari/sympy,VaibhavAgarwalVA/sympy,bukzor/sympy,wyom/sympy,flacjacket/sympy,sahmed95/sympy,Titan-C/sympy,bukzor/sympy,jamesblunt/sympy,kaushik94/sympy,cswiercz/sympy,sahilshekhawat/sympy,skirpichev/omg,dqnykamp/sympy,wyom/sympy,AunShiLord/sympy,cccfran/sympy,mafiya69/sympy,AkademieOlympia/sympy,Sumith1896/sympy,jbbskinny/sympy,kumarkrishna/sympy,atreyv/sympy,ga7g08/sympy,postvakje/sympy,farhaanbukhsh/sympy,pbrady/sympy,moble/sympy,ahhda/sympy,minrk/sympy,garvitr/sympy,skidzo/sympy,MridulS/sympy,meghana1995/sympy,ahhda/sympy,yashsharan/sympy,drufat/sympy,shipci/sympy,lindsayad/sympy,aktech/sympy,kaushik94/sympy,pernici/sympy,abloomston/sympy,atsao72/sympy,hrashk/sympy,minrk/sympy,sahmed95/sympy,sampadsaha5/sympy,wanglongqi/sympy,chaffra/sympy,hargup/sympy,AkademieOlympia/sympy,hargup/sympy,moble/sympy,yashsharan/sympy,garvitr/sympy,aktech/sympy,kevalds51/sympy,liangjiaxing/sympy,emon10005/sympy,atreyv/sympy,shipci/sympy,iamutkarshtiwari/sympy,hazelnusse/sympy-old,asm666/sympy,Davidjohnwilson/sympy,fperez/sympy,oliverlee/sympy,postvakje/sympy,yukoba/sympy,amitjamadagni/sympy,Vishluck/sympy,AunShiLord/sympy,kaichogami/sympy,AunShiLord/sympy,sunny94/temp,madan96/sympy,srjoglekar246/sympy,ChristinaZografou/sympy,Arafatk/sympy,cccfran/sympy,VaibhavAgarwalVA/sympy,abloomston/sympy,atsao72/sympy,ga7g08/sympy,Mitchkoens/sympy,Curious72/sympy,Designist/sympy,ga7g08/sympy,Shaswat27/sympy,ChristinaZografou/sympy,grevutiu-gabriel/sympy,mcdaniel67/sympy,hazelnusse/sympy-old,grevutiu-gabriel/sympy,kumarkrishna/sympy,toolforger/sympy,abhiii5459/sympy,yukoba/sympy,beni55/sympy,Gadal/sympy,jamesblunt/sympy,hargup/sympy,Gadal/sympy,jaimahajan1997/sympy,aktech/sympy,diofant/diofant,Mitchkoens/sympy,kaichogami/sympy,bukzor/sympy,souravsingh/sympy,sahmed95/sympy,toolforger/sympy,cswiercz/sympy,abloomston/sympy,MechCoder/sympy,Shaswat27/sympy,pbrady/sympy,wanglongqi/sympy,sampadsaha5/sympy,pandeyadarsh/sympy,drufat/sympy,yukoba/sympy,kaichogami/sympy,MechCoder/sympy,wanglongqi/sympy,Vishluck/sympy,abhiii5459/sympy,vipulroxx/sympy,AkademieOlympia/sympy,Arafatk/sympy,mcdaniel67/sympy,kumarkrishna/sympy,lidavidm/sympy,jerli/sympy,sampadsaha5/sympy,sahilshekhawat/sympy,debugger22/sympy,souravsingh/sympy,pandeyadarsh/sympy,grevutiu-gabriel/sympy,wyom/sympy,shikil/sympy,maniteja123/sympy,skidzo/sympy,jamesblunt/sympy,amitjamadagni/sympy,kevalds51/sympy,rahuldan/sympy,beni55/sympy,atsao72/sympy,oliverlee/sympy,jerli/sympy,Davidjohnwilson/sympy,saurabhjn76/sympy,postvakje/sympy,liangjiaxing/sympy,chaffra/sympy,dqnykamp/sympy,garvitr/sympy,maniteja123/sympy,yashsharan/sympy,oliverlee/sympy,Designist/sympy,asm666/sympy,farhaanbukhsh/sympy,jbbskinny/sympy,sunny94/temp,Sumith1896/sympy,farhaanbukhsh/sympy,cccfran/sympy,hrashk/sympy,Titan-C/sympy,debugger22/sympy,pbrady/sympy,lidavidm/sympy,meghana1995/sympy,jerli/sympy
|
__version__ = "0.5.0"
from sympy.core import *
from series import *
from simplify import *
from solvers import *
from matrices import *
from geometry import *
from polynomials import *
from utilities import *
#from specfun import *
from integrals import *
try:
from plotting import Plot
except ImportError, e:
print str(e)
Hide ctypes import error until Plot() is called.
|
__version__ = "0.5.0"
from sympy.core import *
from series import *
from simplify import *
from solvers import *
from matrices import *
from geometry import *
from polynomials import *
from utilities import *
#from specfun import *
from integrals import *
try:
from plotting import Plot
except ImportError, e:
class Plot(object):
def __init__(*args, **kwargs):
raise e
|
<commit_before>
__version__ = "0.5.0"
from sympy.core import *
from series import *
from simplify import *
from solvers import *
from matrices import *
from geometry import *
from polynomials import *
from utilities import *
#from specfun import *
from integrals import *
try:
from plotting import Plot
except ImportError, e:
print str(e)
<commit_msg>Hide ctypes import error until Plot() is called.<commit_after>
|
__version__ = "0.5.0"
from sympy.core import *
from series import *
from simplify import *
from solvers import *
from matrices import *
from geometry import *
from polynomials import *
from utilities import *
#from specfun import *
from integrals import *
try:
from plotting import Plot
except ImportError, e:
class Plot(object):
def __init__(*args, **kwargs):
raise e
|
__version__ = "0.5.0"
from sympy.core import *
from series import *
from simplify import *
from solvers import *
from matrices import *
from geometry import *
from polynomials import *
from utilities import *
#from specfun import *
from integrals import *
try:
from plotting import Plot
except ImportError, e:
print str(e)
Hide ctypes import error until Plot() is called.
__version__ = "0.5.0"
from sympy.core import *
from series import *
from simplify import *
from solvers import *
from matrices import *
from geometry import *
from polynomials import *
from utilities import *
#from specfun import *
from integrals import *
try:
from plotting import Plot
except ImportError, e:
class Plot(object):
def __init__(*args, **kwargs):
raise e
|
<commit_before>
__version__ = "0.5.0"
from sympy.core import *
from series import *
from simplify import *
from solvers import *
from matrices import *
from geometry import *
from polynomials import *
from utilities import *
#from specfun import *
from integrals import *
try:
from plotting import Plot
except ImportError, e:
print str(e)
<commit_msg>Hide ctypes import error until Plot() is called.<commit_after>
__version__ = "0.5.0"
from sympy.core import *
from series import *
from simplify import *
from solvers import *
from matrices import *
from geometry import *
from polynomials import *
from utilities import *
#from specfun import *
from integrals import *
try:
from plotting import Plot
except ImportError, e:
class Plot(object):
def __init__(*args, **kwargs):
raise e
|
b82dbd63aedf8a6a6af494b6d6be697a9f4230d5
|
tests/test_utils.py
|
tests/test_utils.py
|
import pickle
from six.moves import range
from fuel.utils import do_not_pickle_attributes
@do_not_pickle_attributes("non_pickable", "bulky_attr")
class TestClass(object):
def __init__(self):
self.load()
def load(self):
self.bulky_attr = list(range(100))
self.non_pickable = lambda x: x
def test_do_not_pickle_attributes():
cl = TestClass()
dump = pickle.dumps(cl)
loaded = pickle.loads(dump)
assert loaded.bulky_attr == list(range(100))
assert loaded.non_pickable is not None
|
import pickle
from six.moves import range
from fuel.utils import do_not_pickle_attributes, expand_axis_label
@do_not_pickle_attributes("non_pickable", "bulky_attr")
class TestClass(object):
def __init__(self):
self.load()
def load(self):
self.bulky_attr = list(range(100))
self.non_pickable = lambda x: x
def test_do_not_pickle_attributes():
cl = TestClass()
dump = pickle.dumps(cl)
loaded = pickle.loads(dump)
assert loaded.bulky_attr == list(range(100))
assert loaded.non_pickable is not None
def test_expand_axis_label():
assert expand_axis_label('b') == 'batch'
assert expand_axis_label('c') == 'channel'
assert expand_axis_label('t') == 'time'
assert expand_axis_label('0') == 'axis_0'
assert expand_axis_label('1') == 'axis_1'
assert expand_axis_label('0b') == '0b'
assert expand_axis_label('') == ''
|
Add unit test for expand_axis_label
|
Add unit test for expand_axis_label
|
Python
|
mit
|
dwf/fuel,ejls/fuel,udibr/fuel,rizar/fuel,capybaralet/fuel,rizar/fuel,EderSantana/fuel,EderSantana/fuel,orhanf/fuel,aalmah/fuel,mila-udem/fuel,mjwillson/fuel,glewis17/fuel,orhanf/fuel,dhruvparamhans/fuel,hantek/fuel,lamblin/fuel,jbornschein/fuel,dribnet/fuel,markusnagel/fuel,udibr/fuel,harmdevries89/fuel,dribnet/fuel,glewis17/fuel,janchorowski/fuel,harmdevries89/fuel,chrishokamp/fuel,jbornschein/fuel,vdumoulin/fuel,codeaudit/fuel,aalmah/fuel,markusnagel/fuel,dmitriy-serdyuk/fuel,rodrigob/fuel,dwf/fuel,dmitriy-serdyuk/fuel,bouthilx/fuel,bouthilx/fuel,capybaralet/fuel,janchorowski/fuel,laurent-dinh/fuel,dhruvparamhans/fuel,mjwillson/fuel,chrishokamp/fuel,mila-udem/fuel,lamblin/fuel,ejls/fuel,vdumoulin/fuel,laurent-dinh/fuel,codeaudit/fuel,hantek/fuel,rodrigob/fuel
|
import pickle
from six.moves import range
from fuel.utils import do_not_pickle_attributes
@do_not_pickle_attributes("non_pickable", "bulky_attr")
class TestClass(object):
def __init__(self):
self.load()
def load(self):
self.bulky_attr = list(range(100))
self.non_pickable = lambda x: x
def test_do_not_pickle_attributes():
cl = TestClass()
dump = pickle.dumps(cl)
loaded = pickle.loads(dump)
assert loaded.bulky_attr == list(range(100))
assert loaded.non_pickable is not None
Add unit test for expand_axis_label
|
import pickle
from six.moves import range
from fuel.utils import do_not_pickle_attributes, expand_axis_label
@do_not_pickle_attributes("non_pickable", "bulky_attr")
class TestClass(object):
def __init__(self):
self.load()
def load(self):
self.bulky_attr = list(range(100))
self.non_pickable = lambda x: x
def test_do_not_pickle_attributes():
cl = TestClass()
dump = pickle.dumps(cl)
loaded = pickle.loads(dump)
assert loaded.bulky_attr == list(range(100))
assert loaded.non_pickable is not None
def test_expand_axis_label():
assert expand_axis_label('b') == 'batch'
assert expand_axis_label('c') == 'channel'
assert expand_axis_label('t') == 'time'
assert expand_axis_label('0') == 'axis_0'
assert expand_axis_label('1') == 'axis_1'
assert expand_axis_label('0b') == '0b'
assert expand_axis_label('') == ''
|
<commit_before>import pickle
from six.moves import range
from fuel.utils import do_not_pickle_attributes
@do_not_pickle_attributes("non_pickable", "bulky_attr")
class TestClass(object):
def __init__(self):
self.load()
def load(self):
self.bulky_attr = list(range(100))
self.non_pickable = lambda x: x
def test_do_not_pickle_attributes():
cl = TestClass()
dump = pickle.dumps(cl)
loaded = pickle.loads(dump)
assert loaded.bulky_attr == list(range(100))
assert loaded.non_pickable is not None
<commit_msg>Add unit test for expand_axis_label<commit_after>
|
import pickle
from six.moves import range
from fuel.utils import do_not_pickle_attributes, expand_axis_label
@do_not_pickle_attributes("non_pickable", "bulky_attr")
class TestClass(object):
def __init__(self):
self.load()
def load(self):
self.bulky_attr = list(range(100))
self.non_pickable = lambda x: x
def test_do_not_pickle_attributes():
cl = TestClass()
dump = pickle.dumps(cl)
loaded = pickle.loads(dump)
assert loaded.bulky_attr == list(range(100))
assert loaded.non_pickable is not None
def test_expand_axis_label():
assert expand_axis_label('b') == 'batch'
assert expand_axis_label('c') == 'channel'
assert expand_axis_label('t') == 'time'
assert expand_axis_label('0') == 'axis_0'
assert expand_axis_label('1') == 'axis_1'
assert expand_axis_label('0b') == '0b'
assert expand_axis_label('') == ''
|
import pickle
from six.moves import range
from fuel.utils import do_not_pickle_attributes
@do_not_pickle_attributes("non_pickable", "bulky_attr")
class TestClass(object):
def __init__(self):
self.load()
def load(self):
self.bulky_attr = list(range(100))
self.non_pickable = lambda x: x
def test_do_not_pickle_attributes():
cl = TestClass()
dump = pickle.dumps(cl)
loaded = pickle.loads(dump)
assert loaded.bulky_attr == list(range(100))
assert loaded.non_pickable is not None
Add unit test for expand_axis_labelimport pickle
from six.moves import range
from fuel.utils import do_not_pickle_attributes, expand_axis_label
@do_not_pickle_attributes("non_pickable", "bulky_attr")
class TestClass(object):
def __init__(self):
self.load()
def load(self):
self.bulky_attr = list(range(100))
self.non_pickable = lambda x: x
def test_do_not_pickle_attributes():
cl = TestClass()
dump = pickle.dumps(cl)
loaded = pickle.loads(dump)
assert loaded.bulky_attr == list(range(100))
assert loaded.non_pickable is not None
def test_expand_axis_label():
assert expand_axis_label('b') == 'batch'
assert expand_axis_label('c') == 'channel'
assert expand_axis_label('t') == 'time'
assert expand_axis_label('0') == 'axis_0'
assert expand_axis_label('1') == 'axis_1'
assert expand_axis_label('0b') == '0b'
assert expand_axis_label('') == ''
|
<commit_before>import pickle
from six.moves import range
from fuel.utils import do_not_pickle_attributes
@do_not_pickle_attributes("non_pickable", "bulky_attr")
class TestClass(object):
def __init__(self):
self.load()
def load(self):
self.bulky_attr = list(range(100))
self.non_pickable = lambda x: x
def test_do_not_pickle_attributes():
cl = TestClass()
dump = pickle.dumps(cl)
loaded = pickle.loads(dump)
assert loaded.bulky_attr == list(range(100))
assert loaded.non_pickable is not None
<commit_msg>Add unit test for expand_axis_label<commit_after>import pickle
from six.moves import range
from fuel.utils import do_not_pickle_attributes, expand_axis_label
@do_not_pickle_attributes("non_pickable", "bulky_attr")
class TestClass(object):
def __init__(self):
self.load()
def load(self):
self.bulky_attr = list(range(100))
self.non_pickable = lambda x: x
def test_do_not_pickle_attributes():
cl = TestClass()
dump = pickle.dumps(cl)
loaded = pickle.loads(dump)
assert loaded.bulky_attr == list(range(100))
assert loaded.non_pickable is not None
def test_expand_axis_label():
assert expand_axis_label('b') == 'batch'
assert expand_axis_label('c') == 'channel'
assert expand_axis_label('t') == 'time'
assert expand_axis_label('0') == 'axis_0'
assert expand_axis_label('1') == 'axis_1'
assert expand_axis_label('0b') == '0b'
assert expand_axis_label('') == ''
|
3e6f835a88183182b6ebba25c61666735a69fc81
|
tests/vaultshell.py
|
tests/vaultshell.py
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
import unittest
class VaultShellTests(unittest.TestCase):
def test_basic(self):
print "test basic. Pass"
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
import unittest
import vault_shell.vault_commandhelper as VaultHelper
class VaultShellTests(unittest.TestCase):
def test_basic(self):
print "test basic. Pass"
vaulthelper = VaultHelper.VaultCommandHelper()
self.failUnless(vaulthelper is not None)
def test_execute_vault_commands(self):
vaulthelper = VaultHelper.VaultCommandHelper()
output = vaulthelper.execute_vault_commands(['vault'])
self.failUnless(output is not None)
def test_get_commandkey_from_cmdlist(self):
vaulthelper = VaultHelper.VaultCommandHelper()
cmdkey = vaulthelper.get_commandkey_from_cmdlist(["token-create"])
self.assertEqual(cmdkey,
"vault_token-create",
msg="cmdkey did not match")
|
Add more tests for the vault commandhelper
|
Add more tests for the vault commandhelper
|
Python
|
apache-2.0
|
bdastur/vault-shell
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
import unittest
class VaultShellTests(unittest.TestCase):
def test_basic(self):
print "test basic. Pass"
Add more tests for the vault commandhelper
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
import unittest
import vault_shell.vault_commandhelper as VaultHelper
class VaultShellTests(unittest.TestCase):
def test_basic(self):
print "test basic. Pass"
vaulthelper = VaultHelper.VaultCommandHelper()
self.failUnless(vaulthelper is not None)
def test_execute_vault_commands(self):
vaulthelper = VaultHelper.VaultCommandHelper()
output = vaulthelper.execute_vault_commands(['vault'])
self.failUnless(output is not None)
def test_get_commandkey_from_cmdlist(self):
vaulthelper = VaultHelper.VaultCommandHelper()
cmdkey = vaulthelper.get_commandkey_from_cmdlist(["token-create"])
self.assertEqual(cmdkey,
"vault_token-create",
msg="cmdkey did not match")
|
<commit_before>#!/usr/bin/env python
# -*- coding: utf-8 -*-
import unittest
class VaultShellTests(unittest.TestCase):
def test_basic(self):
print "test basic. Pass"
<commit_msg>Add more tests for the vault commandhelper<commit_after>
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
import unittest
import vault_shell.vault_commandhelper as VaultHelper
class VaultShellTests(unittest.TestCase):
def test_basic(self):
print "test basic. Pass"
vaulthelper = VaultHelper.VaultCommandHelper()
self.failUnless(vaulthelper is not None)
def test_execute_vault_commands(self):
vaulthelper = VaultHelper.VaultCommandHelper()
output = vaulthelper.execute_vault_commands(['vault'])
self.failUnless(output is not None)
def test_get_commandkey_from_cmdlist(self):
vaulthelper = VaultHelper.VaultCommandHelper()
cmdkey = vaulthelper.get_commandkey_from_cmdlist(["token-create"])
self.assertEqual(cmdkey,
"vault_token-create",
msg="cmdkey did not match")
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
import unittest
class VaultShellTests(unittest.TestCase):
def test_basic(self):
print "test basic. Pass"
Add more tests for the vault commandhelper#!/usr/bin/env python
# -*- coding: utf-8 -*-
import unittest
import vault_shell.vault_commandhelper as VaultHelper
class VaultShellTests(unittest.TestCase):
def test_basic(self):
print "test basic. Pass"
vaulthelper = VaultHelper.VaultCommandHelper()
self.failUnless(vaulthelper is not None)
def test_execute_vault_commands(self):
vaulthelper = VaultHelper.VaultCommandHelper()
output = vaulthelper.execute_vault_commands(['vault'])
self.failUnless(output is not None)
def test_get_commandkey_from_cmdlist(self):
vaulthelper = VaultHelper.VaultCommandHelper()
cmdkey = vaulthelper.get_commandkey_from_cmdlist(["token-create"])
self.assertEqual(cmdkey,
"vault_token-create",
msg="cmdkey did not match")
|
<commit_before>#!/usr/bin/env python
# -*- coding: utf-8 -*-
import unittest
class VaultShellTests(unittest.TestCase):
def test_basic(self):
print "test basic. Pass"
<commit_msg>Add more tests for the vault commandhelper<commit_after>#!/usr/bin/env python
# -*- coding: utf-8 -*-
import unittest
import vault_shell.vault_commandhelper as VaultHelper
class VaultShellTests(unittest.TestCase):
def test_basic(self):
print "test basic. Pass"
vaulthelper = VaultHelper.VaultCommandHelper()
self.failUnless(vaulthelper is not None)
def test_execute_vault_commands(self):
vaulthelper = VaultHelper.VaultCommandHelper()
output = vaulthelper.execute_vault_commands(['vault'])
self.failUnless(output is not None)
def test_get_commandkey_from_cmdlist(self):
vaulthelper = VaultHelper.VaultCommandHelper()
cmdkey = vaulthelper.get_commandkey_from_cmdlist(["token-create"])
self.assertEqual(cmdkey,
"vault_token-create",
msg="cmdkey did not match")
|
7b50c9290a8c8d3481d9147ebb66d3b7868ad7fc
|
bouncer-plumbing/mlab-to-bouncer/makeconfig.py
|
bouncer-plumbing/mlab-to-bouncer/makeconfig.py
|
#!/usr/bin/env python
import sys
import yaml
def read_parts_from_stdin():
data = sys.stdin.read()
parts_string = data.split("----")
parts_parsed = []
for part in parts_string:
part_parsed = yaml.safe_load(part)
parts_parsed.append(part_parsed)
return parts_parsed
def assemble_bouncer_config(parts):
merged_parts = { }
for part in parts:
merged_parts.update(part)
bouncer_config = { 'collector': merged_parts }
return yaml.dump(bouncer_config)
def write_bouncer_config(bouncer_config, path):
print bouncer_config
parts = read_parts_from_stdin()
bouncer_config = assemble_bouncer_config(parts)
write_bouncer_config(bouncer_config, '/home/mlab/data/bouncer.yaml')
|
#!/usr/bin/env python
import sys
import yaml
def read_parts_from_stdin():
data = sys.stdin.read()
parts_string = data.split("----")
parts_parsed = []
for part in parts_string:
part_parsed = yaml.safe_load(part)
parts_parsed.append(part_parsed)
return parts_parsed
def assemble_bouncer_config(parts):
merged_parts = { }
for part in parts:
merged_parts.update(part)
bouncer_config = { 'collector': merged_parts }
return yaml.dump(bouncer_config)
def write_bouncer_config(path, bouncer_config_contents):
try:
f = open(path, 'w')
f.write(bouncer_config_contents)
f.close()
except IOError:
print "Couldn't write to bouncer config file."
exit(1)
bouncer_config_path = '/home/mlab/data/bouncer.yaml'
if len(sys.argv) >= 2:
bouncer_config_path = sys.argv[1]
# FIXME: Read from the mlab-ns simulator.
parts = read_parts_from_stdin()
bouncer_config = assemble_bouncer_config(parts)
write_bouncer_config(bouncer_config_path, bouncer_config)
|
Write to bouncer config file
|
Write to bouncer config file
|
Python
|
apache-2.0
|
m-lab/ooni-support,hellais/ooni-support,m-lab/ooni-support,hellais/ooni-support
|
#!/usr/bin/env python
import sys
import yaml
def read_parts_from_stdin():
data = sys.stdin.read()
parts_string = data.split("----")
parts_parsed = []
for part in parts_string:
part_parsed = yaml.safe_load(part)
parts_parsed.append(part_parsed)
return parts_parsed
def assemble_bouncer_config(parts):
merged_parts = { }
for part in parts:
merged_parts.update(part)
bouncer_config = { 'collector': merged_parts }
return yaml.dump(bouncer_config)
def write_bouncer_config(bouncer_config, path):
print bouncer_config
parts = read_parts_from_stdin()
bouncer_config = assemble_bouncer_config(parts)
write_bouncer_config(bouncer_config, '/home/mlab/data/bouncer.yaml')
Write to bouncer config file
|
#!/usr/bin/env python
import sys
import yaml
def read_parts_from_stdin():
data = sys.stdin.read()
parts_string = data.split("----")
parts_parsed = []
for part in parts_string:
part_parsed = yaml.safe_load(part)
parts_parsed.append(part_parsed)
return parts_parsed
def assemble_bouncer_config(parts):
merged_parts = { }
for part in parts:
merged_parts.update(part)
bouncer_config = { 'collector': merged_parts }
return yaml.dump(bouncer_config)
def write_bouncer_config(path, bouncer_config_contents):
try:
f = open(path, 'w')
f.write(bouncer_config_contents)
f.close()
except IOError:
print "Couldn't write to bouncer config file."
exit(1)
bouncer_config_path = '/home/mlab/data/bouncer.yaml'
if len(sys.argv) >= 2:
bouncer_config_path = sys.argv[1]
# FIXME: Read from the mlab-ns simulator.
parts = read_parts_from_stdin()
bouncer_config = assemble_bouncer_config(parts)
write_bouncer_config(bouncer_config_path, bouncer_config)
|
<commit_before>#!/usr/bin/env python
import sys
import yaml
def read_parts_from_stdin():
data = sys.stdin.read()
parts_string = data.split("----")
parts_parsed = []
for part in parts_string:
part_parsed = yaml.safe_load(part)
parts_parsed.append(part_parsed)
return parts_parsed
def assemble_bouncer_config(parts):
merged_parts = { }
for part in parts:
merged_parts.update(part)
bouncer_config = { 'collector': merged_parts }
return yaml.dump(bouncer_config)
def write_bouncer_config(bouncer_config, path):
print bouncer_config
parts = read_parts_from_stdin()
bouncer_config = assemble_bouncer_config(parts)
write_bouncer_config(bouncer_config, '/home/mlab/data/bouncer.yaml')
<commit_msg>Write to bouncer config file<commit_after>
|
#!/usr/bin/env python
import sys
import yaml
def read_parts_from_stdin():
data = sys.stdin.read()
parts_string = data.split("----")
parts_parsed = []
for part in parts_string:
part_parsed = yaml.safe_load(part)
parts_parsed.append(part_parsed)
return parts_parsed
def assemble_bouncer_config(parts):
merged_parts = { }
for part in parts:
merged_parts.update(part)
bouncer_config = { 'collector': merged_parts }
return yaml.dump(bouncer_config)
def write_bouncer_config(path, bouncer_config_contents):
try:
f = open(path, 'w')
f.write(bouncer_config_contents)
f.close()
except IOError:
print "Couldn't write to bouncer config file."
exit(1)
bouncer_config_path = '/home/mlab/data/bouncer.yaml'
if len(sys.argv) >= 2:
bouncer_config_path = sys.argv[1]
# FIXME: Read from the mlab-ns simulator.
parts = read_parts_from_stdin()
bouncer_config = assemble_bouncer_config(parts)
write_bouncer_config(bouncer_config_path, bouncer_config)
|
#!/usr/bin/env python
import sys
import yaml
def read_parts_from_stdin():
data = sys.stdin.read()
parts_string = data.split("----")
parts_parsed = []
for part in parts_string:
part_parsed = yaml.safe_load(part)
parts_parsed.append(part_parsed)
return parts_parsed
def assemble_bouncer_config(parts):
merged_parts = { }
for part in parts:
merged_parts.update(part)
bouncer_config = { 'collector': merged_parts }
return yaml.dump(bouncer_config)
def write_bouncer_config(bouncer_config, path):
print bouncer_config
parts = read_parts_from_stdin()
bouncer_config = assemble_bouncer_config(parts)
write_bouncer_config(bouncer_config, '/home/mlab/data/bouncer.yaml')
Write to bouncer config file#!/usr/bin/env python
import sys
import yaml
def read_parts_from_stdin():
data = sys.stdin.read()
parts_string = data.split("----")
parts_parsed = []
for part in parts_string:
part_parsed = yaml.safe_load(part)
parts_parsed.append(part_parsed)
return parts_parsed
def assemble_bouncer_config(parts):
merged_parts = { }
for part in parts:
merged_parts.update(part)
bouncer_config = { 'collector': merged_parts }
return yaml.dump(bouncer_config)
def write_bouncer_config(path, bouncer_config_contents):
try:
f = open(path, 'w')
f.write(bouncer_config_contents)
f.close()
except IOError:
print "Couldn't write to bouncer config file."
exit(1)
bouncer_config_path = '/home/mlab/data/bouncer.yaml'
if len(sys.argv) >= 2:
bouncer_config_path = sys.argv[1]
# FIXME: Read from the mlab-ns simulator.
parts = read_parts_from_stdin()
bouncer_config = assemble_bouncer_config(parts)
write_bouncer_config(bouncer_config_path, bouncer_config)
|
<commit_before>#!/usr/bin/env python
import sys
import yaml
def read_parts_from_stdin():
data = sys.stdin.read()
parts_string = data.split("----")
parts_parsed = []
for part in parts_string:
part_parsed = yaml.safe_load(part)
parts_parsed.append(part_parsed)
return parts_parsed
def assemble_bouncer_config(parts):
merged_parts = { }
for part in parts:
merged_parts.update(part)
bouncer_config = { 'collector': merged_parts }
return yaml.dump(bouncer_config)
def write_bouncer_config(bouncer_config, path):
print bouncer_config
parts = read_parts_from_stdin()
bouncer_config = assemble_bouncer_config(parts)
write_bouncer_config(bouncer_config, '/home/mlab/data/bouncer.yaml')
<commit_msg>Write to bouncer config file<commit_after>#!/usr/bin/env python
import sys
import yaml
def read_parts_from_stdin():
data = sys.stdin.read()
parts_string = data.split("----")
parts_parsed = []
for part in parts_string:
part_parsed = yaml.safe_load(part)
parts_parsed.append(part_parsed)
return parts_parsed
def assemble_bouncer_config(parts):
merged_parts = { }
for part in parts:
merged_parts.update(part)
bouncer_config = { 'collector': merged_parts }
return yaml.dump(bouncer_config)
def write_bouncer_config(path, bouncer_config_contents):
try:
f = open(path, 'w')
f.write(bouncer_config_contents)
f.close()
except IOError:
print "Couldn't write to bouncer config file."
exit(1)
bouncer_config_path = '/home/mlab/data/bouncer.yaml'
if len(sys.argv) >= 2:
bouncer_config_path = sys.argv[1]
# FIXME: Read from the mlab-ns simulator.
parts = read_parts_from_stdin()
bouncer_config = assemble_bouncer_config(parts)
write_bouncer_config(bouncer_config_path, bouncer_config)
|
8170ad6cdfd2346bc24a3d743663b4866416ca83
|
Engine.py
|
Engine.py
|
#Imports
import pygame
from pygame.locals import *
from OpenGL.GL import *
from OpenGL.GLU import *
from shapes import Shape, Cube
#Create a game class
class Game(object):
#Constructor
def __init__(self, title, width, height, bgcolour):
#Initialise pygame
pygame.init()
#Set the size of the window
self.size = self.width, self.height = width, height
#Set the default perspective and clipping distances
self.fov = 45.0
self.aspectratio = width / height
self.minrender = 0.1
self.maxrender = 80
#Set the pygame mode to use double buffering and open gl
pygame.set_mode(self.size, DOUBLEBUF|OPENGL)
#Set the perspective
self.setperspective()
#Create an empty list of shapes to render
self.shapes = []
#Create a function to update the perspective
def setperspective(self):
#Set the perspective
gluPerspective(self.fov, self.aspectratio, self.minrender, self.maxrender)
|
#Imports
import pygame
from pygame.locals import *
from OpenGL.GL import *
from OpenGL.GLU import *
from shapes import Shape, Cube
#Create a game class
class Game(object):
#Constructor
def __init__(self, title, width, height, bgcolour):
#Initialise pygame
pygame.init()
#Set the size of the window
self.size = self.width, self.height = width, height
#Set the default perspective and clipping distances
self.fov = 45.0
self.aspectratio = width / height
self.minrender = 0.1
self.maxrender = 80
#Set the pygame mode to use double buffering and open gl
pygame.set_mode(self.size, DOUBLEBUF|OPENGL)
#Set the perspective
self.setperspective()
#Create an empty list of shapes to render
self.shapes = []
#Create a function to update the perspective
def setperspective(self):
#Set the perspective
gluPerspective(self.fov, self.aspectratio, self.minrender, self.maxrender)
#Create a function to add a shape
def addshape(self, s):
self.shapes.append(s)
#Create a function to render the shapes
def render(self):
#For each of the shapes, check the type and render it
for s in shapes:
#If the shape is a cube, call the rendercube method
if s.type == Shape.CUBE:
rendercube(s)
|
Add functions to add shapes and iterate over each shape to render.
|
Add functions to add shapes and iterate over each shape to render.
|
Python
|
mit
|
thebillington/pyPhys3D
|
#Imports
import pygame
from pygame.locals import *
from OpenGL.GL import *
from OpenGL.GLU import *
from shapes import Shape, Cube
#Create a game class
class Game(object):
#Constructor
def __init__(self, title, width, height, bgcolour):
#Initialise pygame
pygame.init()
#Set the size of the window
self.size = self.width, self.height = width, height
#Set the default perspective and clipping distances
self.fov = 45.0
self.aspectratio = width / height
self.minrender = 0.1
self.maxrender = 80
#Set the pygame mode to use double buffering and open gl
pygame.set_mode(self.size, DOUBLEBUF|OPENGL)
#Set the perspective
self.setperspective()
#Create an empty list of shapes to render
self.shapes = []
#Create a function to update the perspective
def setperspective(self):
#Set the perspective
gluPerspective(self.fov, self.aspectratio, self.minrender, self.maxrender)
Add functions to add shapes and iterate over each shape to render.
|
#Imports
import pygame
from pygame.locals import *
from OpenGL.GL import *
from OpenGL.GLU import *
from shapes import Shape, Cube
#Create a game class
class Game(object):
#Constructor
def __init__(self, title, width, height, bgcolour):
#Initialise pygame
pygame.init()
#Set the size of the window
self.size = self.width, self.height = width, height
#Set the default perspective and clipping distances
self.fov = 45.0
self.aspectratio = width / height
self.minrender = 0.1
self.maxrender = 80
#Set the pygame mode to use double buffering and open gl
pygame.set_mode(self.size, DOUBLEBUF|OPENGL)
#Set the perspective
self.setperspective()
#Create an empty list of shapes to render
self.shapes = []
#Create a function to update the perspective
def setperspective(self):
#Set the perspective
gluPerspective(self.fov, self.aspectratio, self.minrender, self.maxrender)
#Create a function to add a shape
def addshape(self, s):
self.shapes.append(s)
#Create a function to render the shapes
def render(self):
#For each of the shapes, check the type and render it
for s in shapes:
#If the shape is a cube, call the rendercube method
if s.type == Shape.CUBE:
rendercube(s)
|
<commit_before>#Imports
import pygame
from pygame.locals import *
from OpenGL.GL import *
from OpenGL.GLU import *
from shapes import Shape, Cube
#Create a game class
class Game(object):
#Constructor
def __init__(self, title, width, height, bgcolour):
#Initialise pygame
pygame.init()
#Set the size of the window
self.size = self.width, self.height = width, height
#Set the default perspective and clipping distances
self.fov = 45.0
self.aspectratio = width / height
self.minrender = 0.1
self.maxrender = 80
#Set the pygame mode to use double buffering and open gl
pygame.set_mode(self.size, DOUBLEBUF|OPENGL)
#Set the perspective
self.setperspective()
#Create an empty list of shapes to render
self.shapes = []
#Create a function to update the perspective
def setperspective(self):
#Set the perspective
gluPerspective(self.fov, self.aspectratio, self.minrender, self.maxrender)
<commit_msg>Add functions to add shapes and iterate over each shape to render.<commit_after>
|
#Imports
import pygame
from pygame.locals import *
from OpenGL.GL import *
from OpenGL.GLU import *
from shapes import Shape, Cube
#Create a game class
class Game(object):
#Constructor
def __init__(self, title, width, height, bgcolour):
#Initialise pygame
pygame.init()
#Set the size of the window
self.size = self.width, self.height = width, height
#Set the default perspective and clipping distances
self.fov = 45.0
self.aspectratio = width / height
self.minrender = 0.1
self.maxrender = 80
#Set the pygame mode to use double buffering and open gl
pygame.set_mode(self.size, DOUBLEBUF|OPENGL)
#Set the perspective
self.setperspective()
#Create an empty list of shapes to render
self.shapes = []
#Create a function to update the perspective
def setperspective(self):
#Set the perspective
gluPerspective(self.fov, self.aspectratio, self.minrender, self.maxrender)
#Create a function to add a shape
def addshape(self, s):
self.shapes.append(s)
#Create a function to render the shapes
def render(self):
#For each of the shapes, check the type and render it
for s in shapes:
#If the shape is a cube, call the rendercube method
if s.type == Shape.CUBE:
rendercube(s)
|
#Imports
import pygame
from pygame.locals import *
from OpenGL.GL import *
from OpenGL.GLU import *
from shapes import Shape, Cube
#Create a game class
class Game(object):
#Constructor
def __init__(self, title, width, height, bgcolour):
#Initialise pygame
pygame.init()
#Set the size of the window
self.size = self.width, self.height = width, height
#Set the default perspective and clipping distances
self.fov = 45.0
self.aspectratio = width / height
self.minrender = 0.1
self.maxrender = 80
#Set the pygame mode to use double buffering and open gl
pygame.set_mode(self.size, DOUBLEBUF|OPENGL)
#Set the perspective
self.setperspective()
#Create an empty list of shapes to render
self.shapes = []
#Create a function to update the perspective
def setperspective(self):
#Set the perspective
gluPerspective(self.fov, self.aspectratio, self.minrender, self.maxrender)
Add functions to add shapes and iterate over each shape to render.#Imports
import pygame
from pygame.locals import *
from OpenGL.GL import *
from OpenGL.GLU import *
from shapes import Shape, Cube
#Create a game class
class Game(object):
#Constructor
def __init__(self, title, width, height, bgcolour):
#Initialise pygame
pygame.init()
#Set the size of the window
self.size = self.width, self.height = width, height
#Set the default perspective and clipping distances
self.fov = 45.0
self.aspectratio = width / height
self.minrender = 0.1
self.maxrender = 80
#Set the pygame mode to use double buffering and open gl
pygame.set_mode(self.size, DOUBLEBUF|OPENGL)
#Set the perspective
self.setperspective()
#Create an empty list of shapes to render
self.shapes = []
#Create a function to update the perspective
def setperspective(self):
#Set the perspective
gluPerspective(self.fov, self.aspectratio, self.minrender, self.maxrender)
#Create a function to add a shape
def addshape(self, s):
self.shapes.append(s)
#Create a function to render the shapes
def render(self):
#For each of the shapes, check the type and render it
for s in shapes:
#If the shape is a cube, call the rendercube method
if s.type == Shape.CUBE:
rendercube(s)
|
<commit_before>#Imports
import pygame
from pygame.locals import *
from OpenGL.GL import *
from OpenGL.GLU import *
from shapes import Shape, Cube
#Create a game class
class Game(object):
#Constructor
def __init__(self, title, width, height, bgcolour):
#Initialise pygame
pygame.init()
#Set the size of the window
self.size = self.width, self.height = width, height
#Set the default perspective and clipping distances
self.fov = 45.0
self.aspectratio = width / height
self.minrender = 0.1
self.maxrender = 80
#Set the pygame mode to use double buffering and open gl
pygame.set_mode(self.size, DOUBLEBUF|OPENGL)
#Set the perspective
self.setperspective()
#Create an empty list of shapes to render
self.shapes = []
#Create a function to update the perspective
def setperspective(self):
#Set the perspective
gluPerspective(self.fov, self.aspectratio, self.minrender, self.maxrender)
<commit_msg>Add functions to add shapes and iterate over each shape to render.<commit_after>#Imports
import pygame
from pygame.locals import *
from OpenGL.GL import *
from OpenGL.GLU import *
from shapes import Shape, Cube
#Create a game class
class Game(object):
#Constructor
def __init__(self, title, width, height, bgcolour):
#Initialise pygame
pygame.init()
#Set the size of the window
self.size = self.width, self.height = width, height
#Set the default perspective and clipping distances
self.fov = 45.0
self.aspectratio = width / height
self.minrender = 0.1
self.maxrender = 80
#Set the pygame mode to use double buffering and open gl
pygame.set_mode(self.size, DOUBLEBUF|OPENGL)
#Set the perspective
self.setperspective()
#Create an empty list of shapes to render
self.shapes = []
#Create a function to update the perspective
def setperspective(self):
#Set the perspective
gluPerspective(self.fov, self.aspectratio, self.minrender, self.maxrender)
#Create a function to add a shape
def addshape(self, s):
self.shapes.append(s)
#Create a function to render the shapes
def render(self):
#For each of the shapes, check the type and render it
for s in shapes:
#If the shape is a cube, call the rendercube method
if s.type == Shape.CUBE:
rendercube(s)
|
53e1ff21bb219495f1b99f84dbb31624fdd35231
|
lpthw/ex33.py
|
lpthw/ex33.py
|
#i = 0
#numbers = []
#while i < 6:
# print "At the top i is %d" % i
# numbers.append(i)
#
# i += 1
# print "Numbers now: ", numbers
# print "At the bottom i is %d" % i
#
#
#print "The numbers: "
#
#for num in numbers:
# print num
#Study Drills
print "What's the limit of the list?"
a = raw_input("> ")
def list_numbers(a):
"""This function might add numbers to a list?"""
i = 0
numbers = []
while i < a:
print "At the top i is %d" % i
numbers.append(i)
i += 1
print "Numbers now: ", numbers
print "At the bottom i is %d" % i
print "The numbers: "
for num in numbers:
print num
return
list_numbers(a)
|
#i = 0
#numbers = []
#while i < 6:
# print "At the top i is %d" % i
# numbers.append(i)
#
# i += 1
# print "Numbers now: ", numbers
# print "At the bottom i is %d" % i
#
#
#print "The numbers: "
#
#for num in numbers:
# print num
#Study Drills
print "What's the limit of the list?"
a = int(raw_input("> "))
def list_numbers(a):
"""This function might add numbers to a list?"""
i = 0
numbers = []
while i < a:
print "At the top i is %d" % i
numbers.append(i)
i += 1
print "Numbers now: ", numbers
print "At the bottom i is %d" % i
print "The numbers: "
for num in numbers:
print num
return
list_numbers(a)
|
Fix that crazy error that would cause enless looping...
|
Fix that crazy error that would cause enless looping...
|
Python
|
mit
|
jaredmanning/learning,jaredmanning/learning
|
#i = 0
#numbers = []
#while i < 6:
# print "At the top i is %d" % i
# numbers.append(i)
#
# i += 1
# print "Numbers now: ", numbers
# print "At the bottom i is %d" % i
#
#
#print "The numbers: "
#
#for num in numbers:
# print num
#Study Drills
print "What's the limit of the list?"
a = raw_input("> ")
def list_numbers(a):
"""This function might add numbers to a list?"""
i = 0
numbers = []
while i < a:
print "At the top i is %d" % i
numbers.append(i)
i += 1
print "Numbers now: ", numbers
print "At the bottom i is %d" % i
print "The numbers: "
for num in numbers:
print num
return
list_numbers(a)
Fix that crazy error that would cause enless looping...
|
#i = 0
#numbers = []
#while i < 6:
# print "At the top i is %d" % i
# numbers.append(i)
#
# i += 1
# print "Numbers now: ", numbers
# print "At the bottom i is %d" % i
#
#
#print "The numbers: "
#
#for num in numbers:
# print num
#Study Drills
print "What's the limit of the list?"
a = int(raw_input("> "))
def list_numbers(a):
"""This function might add numbers to a list?"""
i = 0
numbers = []
while i < a:
print "At the top i is %d" % i
numbers.append(i)
i += 1
print "Numbers now: ", numbers
print "At the bottom i is %d" % i
print "The numbers: "
for num in numbers:
print num
return
list_numbers(a)
|
<commit_before>#i = 0
#numbers = []
#while i < 6:
# print "At the top i is %d" % i
# numbers.append(i)
#
# i += 1
# print "Numbers now: ", numbers
# print "At the bottom i is %d" % i
#
#
#print "The numbers: "
#
#for num in numbers:
# print num
#Study Drills
print "What's the limit of the list?"
a = raw_input("> ")
def list_numbers(a):
"""This function might add numbers to a list?"""
i = 0
numbers = []
while i < a:
print "At the top i is %d" % i
numbers.append(i)
i += 1
print "Numbers now: ", numbers
print "At the bottom i is %d" % i
print "The numbers: "
for num in numbers:
print num
return
list_numbers(a)
<commit_msg>Fix that crazy error that would cause enless looping...<commit_after>
|
#i = 0
#numbers = []
#while i < 6:
# print "At the top i is %d" % i
# numbers.append(i)
#
# i += 1
# print "Numbers now: ", numbers
# print "At the bottom i is %d" % i
#
#
#print "The numbers: "
#
#for num in numbers:
# print num
#Study Drills
print "What's the limit of the list?"
a = int(raw_input("> "))
def list_numbers(a):
"""This function might add numbers to a list?"""
i = 0
numbers = []
while i < a:
print "At the top i is %d" % i
numbers.append(i)
i += 1
print "Numbers now: ", numbers
print "At the bottom i is %d" % i
print "The numbers: "
for num in numbers:
print num
return
list_numbers(a)
|
#i = 0
#numbers = []
#while i < 6:
# print "At the top i is %d" % i
# numbers.append(i)
#
# i += 1
# print "Numbers now: ", numbers
# print "At the bottom i is %d" % i
#
#
#print "The numbers: "
#
#for num in numbers:
# print num
#Study Drills
print "What's the limit of the list?"
a = raw_input("> ")
def list_numbers(a):
"""This function might add numbers to a list?"""
i = 0
numbers = []
while i < a:
print "At the top i is %d" % i
numbers.append(i)
i += 1
print "Numbers now: ", numbers
print "At the bottom i is %d" % i
print "The numbers: "
for num in numbers:
print num
return
list_numbers(a)
Fix that crazy error that would cause enless looping...#i = 0
#numbers = []
#while i < 6:
# print "At the top i is %d" % i
# numbers.append(i)
#
# i += 1
# print "Numbers now: ", numbers
# print "At the bottom i is %d" % i
#
#
#print "The numbers: "
#
#for num in numbers:
# print num
#Study Drills
print "What's the limit of the list?"
a = int(raw_input("> "))
def list_numbers(a):
"""This function might add numbers to a list?"""
i = 0
numbers = []
while i < a:
print "At the top i is %d" % i
numbers.append(i)
i += 1
print "Numbers now: ", numbers
print "At the bottom i is %d" % i
print "The numbers: "
for num in numbers:
print num
return
list_numbers(a)
|
<commit_before>#i = 0
#numbers = []
#while i < 6:
# print "At the top i is %d" % i
# numbers.append(i)
#
# i += 1
# print "Numbers now: ", numbers
# print "At the bottom i is %d" % i
#
#
#print "The numbers: "
#
#for num in numbers:
# print num
#Study Drills
print "What's the limit of the list?"
a = raw_input("> ")
def list_numbers(a):
"""This function might add numbers to a list?"""
i = 0
numbers = []
while i < a:
print "At the top i is %d" % i
numbers.append(i)
i += 1
print "Numbers now: ", numbers
print "At the bottom i is %d" % i
print "The numbers: "
for num in numbers:
print num
return
list_numbers(a)
<commit_msg>Fix that crazy error that would cause enless looping...<commit_after>#i = 0
#numbers = []
#while i < 6:
# print "At the top i is %d" % i
# numbers.append(i)
#
# i += 1
# print "Numbers now: ", numbers
# print "At the bottom i is %d" % i
#
#
#print "The numbers: "
#
#for num in numbers:
# print num
#Study Drills
print "What's the limit of the list?"
a = int(raw_input("> "))
def list_numbers(a):
"""This function might add numbers to a list?"""
i = 0
numbers = []
while i < a:
print "At the top i is %d" % i
numbers.append(i)
i += 1
print "Numbers now: ", numbers
print "At the bottom i is %d" % i
print "The numbers: "
for num in numbers:
print num
return
list_numbers(a)
|
92adc02daae13f6ef24ae1ec2eafac77ce528a74
|
setup/timvideos/streaming/list_aws_hosts.py
|
setup/timvideos/streaming/list_aws_hosts.py
|
# list_aws_hosts.py
# list all active ec2 hosts
from boto import ec2
import pw
creds = pw.stream['aws']
ec2conn = ec2.connection.EC2Connection(creds['id'], creds['key'] )
reservations = ec2conn.get_all_instances()
instances = [i for r in reservations for i in r.instances]
for i in instances:
if not i.dns_name:
continue
print i.tags['Name'], i.id, i.dns_name
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
# list_aws_hosts.py
# list all active ec2 hosts
"""
Start / stop by name.
Start mission "list_aws_hosts.py start mission"
Stop mission "list_aws_hosts.py stop mission"
Status mission "list_aws_hosts.py status mission"
mission i-b59966c7 **OFF** stopped
"""
from boto import ec2
import pw
creds = pw.stream['aws']
def aws_start(instance, name):
if name == instance.tags['Name']:
instance.start()
def aws_stop(instance, name):
if name == instance.tags['Name']:
instance.stop()
def aws_status(instance, name=None):
if name and not name == instance.tags['Name']:
return
if not instance.dns_name:
print instance.tags['Name'], instance.id, '**OFF**', instance.state
else:
print instance.tags['Name'], instance.id, instance.dns_name, instance.state
def do_command(command, name):
ec2conn = ec2.connection.EC2Connection(creds['id'], creds['key'])
reservations = ec2conn.get_all_instances()
instances = [i for r in reservations for i in r.instances]
for i in instances:
command(i, name)
def do_status():
ec2conn = ec2.connection.EC2Connection(creds['id'], creds['key'])
reservations = ec2conn.get_all_instances()
instances = [i for r in reservations for i in r.instances]
for i in instances:
aws_status(i)
if __name__ == '__main__':
import sys
if len(sys.argv) == 3:
command, name = sys.argv[1:]
if command == 'start':
do_command(aws_start, name)
elif command == 'stop':
do_command(aws_stop, name)
else:
do_command(aws_status, name)
else:
do_status()
|
Update script to start, stop and status by name.
|
Update script to start, stop and status by name.
|
Python
|
mit
|
EricSchles/veyepar,CarlFK/veyepar,yoe/veyepar,yoe/veyepar,xfxf/veyepar,EricSchles/veyepar,yoe/veyepar,xfxf/veyepar,CarlFK/veyepar,xfxf/veyepar,CarlFK/veyepar,EricSchles/veyepar,CarlFK/veyepar,xfxf/veyepar,xfxf/veyepar,CarlFK/veyepar,yoe/veyepar,yoe/veyepar,EricSchles/veyepar,EricSchles/veyepar
|
# list_aws_hosts.py
# list all active ec2 hosts
from boto import ec2
import pw
creds = pw.stream['aws']
ec2conn = ec2.connection.EC2Connection(creds['id'], creds['key'] )
reservations = ec2conn.get_all_instances()
instances = [i for r in reservations for i in r.instances]
for i in instances:
if not i.dns_name:
continue
print i.tags['Name'], i.id, i.dns_name
Update script to start, stop and status by name.
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
# list_aws_hosts.py
# list all active ec2 hosts
"""
Start / stop by name.
Start mission "list_aws_hosts.py start mission"
Stop mission "list_aws_hosts.py stop mission"
Status mission "list_aws_hosts.py status mission"
mission i-b59966c7 **OFF** stopped
"""
from boto import ec2
import pw
creds = pw.stream['aws']
def aws_start(instance, name):
if name == instance.tags['Name']:
instance.start()
def aws_stop(instance, name):
if name == instance.tags['Name']:
instance.stop()
def aws_status(instance, name=None):
if name and not name == instance.tags['Name']:
return
if not instance.dns_name:
print instance.tags['Name'], instance.id, '**OFF**', instance.state
else:
print instance.tags['Name'], instance.id, instance.dns_name, instance.state
def do_command(command, name):
ec2conn = ec2.connection.EC2Connection(creds['id'], creds['key'])
reservations = ec2conn.get_all_instances()
instances = [i for r in reservations for i in r.instances]
for i in instances:
command(i, name)
def do_status():
ec2conn = ec2.connection.EC2Connection(creds['id'], creds['key'])
reservations = ec2conn.get_all_instances()
instances = [i for r in reservations for i in r.instances]
for i in instances:
aws_status(i)
if __name__ == '__main__':
import sys
if len(sys.argv) == 3:
command, name = sys.argv[1:]
if command == 'start':
do_command(aws_start, name)
elif command == 'stop':
do_command(aws_stop, name)
else:
do_command(aws_status, name)
else:
do_status()
|
<commit_before># list_aws_hosts.py
# list all active ec2 hosts
from boto import ec2
import pw
creds = pw.stream['aws']
ec2conn = ec2.connection.EC2Connection(creds['id'], creds['key'] )
reservations = ec2conn.get_all_instances()
instances = [i for r in reservations for i in r.instances]
for i in instances:
if not i.dns_name:
continue
print i.tags['Name'], i.id, i.dns_name
<commit_msg>Update script to start, stop and status by name.<commit_after>
|
#!/usr/bin/env python
# -*- coding: utf-8 -*-
# list_aws_hosts.py
# list all active ec2 hosts
"""
Start / stop by name.
Start mission "list_aws_hosts.py start mission"
Stop mission "list_aws_hosts.py stop mission"
Status mission "list_aws_hosts.py status mission"
mission i-b59966c7 **OFF** stopped
"""
from boto import ec2
import pw
creds = pw.stream['aws']
def aws_start(instance, name):
if name == instance.tags['Name']:
instance.start()
def aws_stop(instance, name):
if name == instance.tags['Name']:
instance.stop()
def aws_status(instance, name=None):
if name and not name == instance.tags['Name']:
return
if not instance.dns_name:
print instance.tags['Name'], instance.id, '**OFF**', instance.state
else:
print instance.tags['Name'], instance.id, instance.dns_name, instance.state
def do_command(command, name):
ec2conn = ec2.connection.EC2Connection(creds['id'], creds['key'])
reservations = ec2conn.get_all_instances()
instances = [i for r in reservations for i in r.instances]
for i in instances:
command(i, name)
def do_status():
ec2conn = ec2.connection.EC2Connection(creds['id'], creds['key'])
reservations = ec2conn.get_all_instances()
instances = [i for r in reservations for i in r.instances]
for i in instances:
aws_status(i)
if __name__ == '__main__':
import sys
if len(sys.argv) == 3:
command, name = sys.argv[1:]
if command == 'start':
do_command(aws_start, name)
elif command == 'stop':
do_command(aws_stop, name)
else:
do_command(aws_status, name)
else:
do_status()
|
# list_aws_hosts.py
# list all active ec2 hosts
from boto import ec2
import pw
creds = pw.stream['aws']
ec2conn = ec2.connection.EC2Connection(creds['id'], creds['key'] )
reservations = ec2conn.get_all_instances()
instances = [i for r in reservations for i in r.instances]
for i in instances:
if not i.dns_name:
continue
print i.tags['Name'], i.id, i.dns_name
Update script to start, stop and status by name.#!/usr/bin/env python
# -*- coding: utf-8 -*-
# list_aws_hosts.py
# list all active ec2 hosts
"""
Start / stop by name.
Start mission "list_aws_hosts.py start mission"
Stop mission "list_aws_hosts.py stop mission"
Status mission "list_aws_hosts.py status mission"
mission i-b59966c7 **OFF** stopped
"""
from boto import ec2
import pw
creds = pw.stream['aws']
def aws_start(instance, name):
if name == instance.tags['Name']:
instance.start()
def aws_stop(instance, name):
if name == instance.tags['Name']:
instance.stop()
def aws_status(instance, name=None):
if name and not name == instance.tags['Name']:
return
if not instance.dns_name:
print instance.tags['Name'], instance.id, '**OFF**', instance.state
else:
print instance.tags['Name'], instance.id, instance.dns_name, instance.state
def do_command(command, name):
ec2conn = ec2.connection.EC2Connection(creds['id'], creds['key'])
reservations = ec2conn.get_all_instances()
instances = [i for r in reservations for i in r.instances]
for i in instances:
command(i, name)
def do_status():
ec2conn = ec2.connection.EC2Connection(creds['id'], creds['key'])
reservations = ec2conn.get_all_instances()
instances = [i for r in reservations for i in r.instances]
for i in instances:
aws_status(i)
if __name__ == '__main__':
import sys
if len(sys.argv) == 3:
command, name = sys.argv[1:]
if command == 'start':
do_command(aws_start, name)
elif command == 'stop':
do_command(aws_stop, name)
else:
do_command(aws_status, name)
else:
do_status()
|
<commit_before># list_aws_hosts.py
# list all active ec2 hosts
from boto import ec2
import pw
creds = pw.stream['aws']
ec2conn = ec2.connection.EC2Connection(creds['id'], creds['key'] )
reservations = ec2conn.get_all_instances()
instances = [i for r in reservations for i in r.instances]
for i in instances:
if not i.dns_name:
continue
print i.tags['Name'], i.id, i.dns_name
<commit_msg>Update script to start, stop and status by name.<commit_after>#!/usr/bin/env python
# -*- coding: utf-8 -*-
# list_aws_hosts.py
# list all active ec2 hosts
"""
Start / stop by name.
Start mission "list_aws_hosts.py start mission"
Stop mission "list_aws_hosts.py stop mission"
Status mission "list_aws_hosts.py status mission"
mission i-b59966c7 **OFF** stopped
"""
from boto import ec2
import pw
creds = pw.stream['aws']
def aws_start(instance, name):
if name == instance.tags['Name']:
instance.start()
def aws_stop(instance, name):
if name == instance.tags['Name']:
instance.stop()
def aws_status(instance, name=None):
if name and not name == instance.tags['Name']:
return
if not instance.dns_name:
print instance.tags['Name'], instance.id, '**OFF**', instance.state
else:
print instance.tags['Name'], instance.id, instance.dns_name, instance.state
def do_command(command, name):
ec2conn = ec2.connection.EC2Connection(creds['id'], creds['key'])
reservations = ec2conn.get_all_instances()
instances = [i for r in reservations for i in r.instances]
for i in instances:
command(i, name)
def do_status():
ec2conn = ec2.connection.EC2Connection(creds['id'], creds['key'])
reservations = ec2conn.get_all_instances()
instances = [i for r in reservations for i in r.instances]
for i in instances:
aws_status(i)
if __name__ == '__main__':
import sys
if len(sys.argv) == 3:
command, name = sys.argv[1:]
if command == 'start':
do_command(aws_start, name)
elif command == 'stop':
do_command(aws_stop, name)
else:
do_command(aws_status, name)
else:
do_status()
|
6da69eb8f13dc56cc19d06a09d74005395de8989
|
fedmsg_meta_umb/tps.py
|
fedmsg_meta_umb/tps.py
|
# Copyright (C) 2017 Red Hat, Inc.
#
# fedmsg_meta_umb is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# fedmsg_meta_umb is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with fedmsg; if not, write to the Free Software
# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
#
# Authors: Gowrishankar Rajaiyan <grajaiya@redhat.com>
from fedmsg.meta.base import BaseProcessor
class TPSProcessor(BaseProcessor):
topic_prefix_re = r'/topic/VirtualTopic\.eng'
__name__ = 'tps'
def title(self, msg, **config):
return msg['topic'].split('.', 2)[-1]
def packages(self, msg, **config):
return set([msg['headers']['component'].rsplit('-', 2)[0]])
|
# Copyright (C) 2017 Red Hat, Inc.
#
# fedmsg_meta_umb is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# fedmsg_meta_umb is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with fedmsg; if not, write to the Free Software
# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
#
# Authors: Gowrishankar Rajaiyan <grajaiya@redhat.com>
from fedmsg.meta.base import BaseProcessor
class TPSProcessor(BaseProcessor):
topic_prefix_re = r'/topic/VirtualTopic\.eng'
__name__ = 'tps'
__description__ = 'package sanity testing of brew builds'
__obj__ = 'Test Package Sanity'
__docs__ = 'https://mojo.redhat.com/docs/DOC-0000000'
__link__ = 'https://sometpslink.engineering.redhat.com'
def title(self, msg, **config):
return msg['topic'].split('.', 2)[-1]
def packages(self, msg, **config):
return set([msg['headers']['component'].rsplit('-', 2)[0]])
|
Add missing attributes in TPSProcessor.
|
Add missing attributes in TPSProcessor.
Signed-off-by: shanks <4508bc7244505cbe28c7ec6ff0c99c8246ca3de6@redhat.com>
|
Python
|
lgpl-2.1
|
release-engineering/fedmsg_meta_umb
|
# Copyright (C) 2017 Red Hat, Inc.
#
# fedmsg_meta_umb is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# fedmsg_meta_umb is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with fedmsg; if not, write to the Free Software
# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
#
# Authors: Gowrishankar Rajaiyan <grajaiya@redhat.com>
from fedmsg.meta.base import BaseProcessor
class TPSProcessor(BaseProcessor):
topic_prefix_re = r'/topic/VirtualTopic\.eng'
__name__ = 'tps'
def title(self, msg, **config):
return msg['topic'].split('.', 2)[-1]
def packages(self, msg, **config):
return set([msg['headers']['component'].rsplit('-', 2)[0]])
Add missing attributes in TPSProcessor.
Signed-off-by: shanks <4508bc7244505cbe28c7ec6ff0c99c8246ca3de6@redhat.com>
|
# Copyright (C) 2017 Red Hat, Inc.
#
# fedmsg_meta_umb is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# fedmsg_meta_umb is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with fedmsg; if not, write to the Free Software
# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
#
# Authors: Gowrishankar Rajaiyan <grajaiya@redhat.com>
from fedmsg.meta.base import BaseProcessor
class TPSProcessor(BaseProcessor):
topic_prefix_re = r'/topic/VirtualTopic\.eng'
__name__ = 'tps'
__description__ = 'package sanity testing of brew builds'
__obj__ = 'Test Package Sanity'
__docs__ = 'https://mojo.redhat.com/docs/DOC-0000000'
__link__ = 'https://sometpslink.engineering.redhat.com'
def title(self, msg, **config):
return msg['topic'].split('.', 2)[-1]
def packages(self, msg, **config):
return set([msg['headers']['component'].rsplit('-', 2)[0]])
|
<commit_before># Copyright (C) 2017 Red Hat, Inc.
#
# fedmsg_meta_umb is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# fedmsg_meta_umb is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with fedmsg; if not, write to the Free Software
# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
#
# Authors: Gowrishankar Rajaiyan <grajaiya@redhat.com>
from fedmsg.meta.base import BaseProcessor
class TPSProcessor(BaseProcessor):
topic_prefix_re = r'/topic/VirtualTopic\.eng'
__name__ = 'tps'
def title(self, msg, **config):
return msg['topic'].split('.', 2)[-1]
def packages(self, msg, **config):
return set([msg['headers']['component'].rsplit('-', 2)[0]])
<commit_msg>Add missing attributes in TPSProcessor.
Signed-off-by: shanks <4508bc7244505cbe28c7ec6ff0c99c8246ca3de6@redhat.com><commit_after>
|
# Copyright (C) 2017 Red Hat, Inc.
#
# fedmsg_meta_umb is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# fedmsg_meta_umb is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with fedmsg; if not, write to the Free Software
# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
#
# Authors: Gowrishankar Rajaiyan <grajaiya@redhat.com>
from fedmsg.meta.base import BaseProcessor
class TPSProcessor(BaseProcessor):
topic_prefix_re = r'/topic/VirtualTopic\.eng'
__name__ = 'tps'
__description__ = 'package sanity testing of brew builds'
__obj__ = 'Test Package Sanity'
__docs__ = 'https://mojo.redhat.com/docs/DOC-0000000'
__link__ = 'https://sometpslink.engineering.redhat.com'
def title(self, msg, **config):
return msg['topic'].split('.', 2)[-1]
def packages(self, msg, **config):
return set([msg['headers']['component'].rsplit('-', 2)[0]])
|
# Copyright (C) 2017 Red Hat, Inc.
#
# fedmsg_meta_umb is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# fedmsg_meta_umb is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with fedmsg; if not, write to the Free Software
# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
#
# Authors: Gowrishankar Rajaiyan <grajaiya@redhat.com>
from fedmsg.meta.base import BaseProcessor
class TPSProcessor(BaseProcessor):
topic_prefix_re = r'/topic/VirtualTopic\.eng'
__name__ = 'tps'
def title(self, msg, **config):
return msg['topic'].split('.', 2)[-1]
def packages(self, msg, **config):
return set([msg['headers']['component'].rsplit('-', 2)[0]])
Add missing attributes in TPSProcessor.
Signed-off-by: shanks <4508bc7244505cbe28c7ec6ff0c99c8246ca3de6@redhat.com># Copyright (C) 2017 Red Hat, Inc.
#
# fedmsg_meta_umb is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# fedmsg_meta_umb is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with fedmsg; if not, write to the Free Software
# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
#
# Authors: Gowrishankar Rajaiyan <grajaiya@redhat.com>
from fedmsg.meta.base import BaseProcessor
class TPSProcessor(BaseProcessor):
topic_prefix_re = r'/topic/VirtualTopic\.eng'
__name__ = 'tps'
__description__ = 'package sanity testing of brew builds'
__obj__ = 'Test Package Sanity'
__docs__ = 'https://mojo.redhat.com/docs/DOC-0000000'
__link__ = 'https://sometpslink.engineering.redhat.com'
def title(self, msg, **config):
return msg['topic'].split('.', 2)[-1]
def packages(self, msg, **config):
return set([msg['headers']['component'].rsplit('-', 2)[0]])
|
<commit_before># Copyright (C) 2017 Red Hat, Inc.
#
# fedmsg_meta_umb is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# fedmsg_meta_umb is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with fedmsg; if not, write to the Free Software
# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
#
# Authors: Gowrishankar Rajaiyan <grajaiya@redhat.com>
from fedmsg.meta.base import BaseProcessor
class TPSProcessor(BaseProcessor):
topic_prefix_re = r'/topic/VirtualTopic\.eng'
__name__ = 'tps'
def title(self, msg, **config):
return msg['topic'].split('.', 2)[-1]
def packages(self, msg, **config):
return set([msg['headers']['component'].rsplit('-', 2)[0]])
<commit_msg>Add missing attributes in TPSProcessor.
Signed-off-by: shanks <4508bc7244505cbe28c7ec6ff0c99c8246ca3de6@redhat.com><commit_after># Copyright (C) 2017 Red Hat, Inc.
#
# fedmsg_meta_umb is free software; you can redistribute it and/or
# modify it under the terms of the GNU Lesser General Public
# License as published by the Free Software Foundation; either
# version 2.1 of the License, or (at your option) any later version.
#
# fedmsg_meta_umb is distributed in the hope that it will be useful,
# but WITHOUT ANY WARRANTY; without even the implied warranty of
# MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU
# Lesser General Public License for more details.
#
# You should have received a copy of the GNU Lesser General Public
# License along with fedmsg; if not, write to the Free Software
# Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301 USA
#
# Authors: Gowrishankar Rajaiyan <grajaiya@redhat.com>
from fedmsg.meta.base import BaseProcessor
class TPSProcessor(BaseProcessor):
topic_prefix_re = r'/topic/VirtualTopic\.eng'
__name__ = 'tps'
__description__ = 'package sanity testing of brew builds'
__obj__ = 'Test Package Sanity'
__docs__ = 'https://mojo.redhat.com/docs/DOC-0000000'
__link__ = 'https://sometpslink.engineering.redhat.com'
def title(self, msg, **config):
return msg['topic'].split('.', 2)[-1]
def packages(self, msg, **config):
return set([msg['headers']['component'].rsplit('-', 2)[0]])
|
153ed6a519d6836adb02b934cff44974a7132b6d
|
flake8/parseDocTest.py
|
flake8/parseDocTest.py
|
def parseFailDetails(failDetails):
""" Parse the line number of the doctest failure"""
import re
failDetails = failDetails.split(',')
lineNo = -1
if len(failDetails) == 3:
match = re.search("line.*?(\d+)", failDetails[1])
if match is None:
return lineNo
lineNo = int(match.group(1))
return lineNo
def parseDocTestResult(docTestResStr):
""" Extract the line number and filename of the doctest failure"""
lines = docTestResStr.split("\n")
for lineNo, line in enumerate(lines):
failure = line.find("Failed example:")
if failure != -1:
failDetails = lines[lineNo - 1]
yield parseFailDetails(failDetails)
|
def parseFailDetails(failDetails):
""" Parse the line number of the doctest failure
>>> parseFailDetails("blah")
-1
"""
import re
failDetails = failDetails.split(',')
lineNo = -1
if len(failDetails) == 3:
match = re.search("line.*?(\d+)", failDetails[1])
if match is None:
return lineNo
lineNo = int(match.group(1))
return lineNo
def parseDocTestResult(docTestResStr):
""" Extract the line number and filename of the doctest failure"""
lines = docTestResStr.split("\n")
for lineNo, line in enumerate(lines):
failure = line.find("Failed example:")
if failure != -1:
failDetails = lines[lineNo - 1]
yield parseFailDetails(failDetails)
if __name__ == "__main__":
from doctest import testmod
testmod()
|
Fix doc test failure parsing
|
Fix doc test failure parsing
|
Python
|
mit
|
softwaredoug/flake8_doctest
|
def parseFailDetails(failDetails):
""" Parse the line number of the doctest failure"""
import re
failDetails = failDetails.split(',')
lineNo = -1
if len(failDetails) == 3:
match = re.search("line.*?(\d+)", failDetails[1])
if match is None:
return lineNo
lineNo = int(match.group(1))
return lineNo
def parseDocTestResult(docTestResStr):
""" Extract the line number and filename of the doctest failure"""
lines = docTestResStr.split("\n")
for lineNo, line in enumerate(lines):
failure = line.find("Failed example:")
if failure != -1:
failDetails = lines[lineNo - 1]
yield parseFailDetails(failDetails)
Fix doc test failure parsing
|
def parseFailDetails(failDetails):
""" Parse the line number of the doctest failure
>>> parseFailDetails("blah")
-1
"""
import re
failDetails = failDetails.split(',')
lineNo = -1
if len(failDetails) == 3:
match = re.search("line.*?(\d+)", failDetails[1])
if match is None:
return lineNo
lineNo = int(match.group(1))
return lineNo
def parseDocTestResult(docTestResStr):
""" Extract the line number and filename of the doctest failure"""
lines = docTestResStr.split("\n")
for lineNo, line in enumerate(lines):
failure = line.find("Failed example:")
if failure != -1:
failDetails = lines[lineNo - 1]
yield parseFailDetails(failDetails)
if __name__ == "__main__":
from doctest import testmod
testmod()
|
<commit_before>def parseFailDetails(failDetails):
""" Parse the line number of the doctest failure"""
import re
failDetails = failDetails.split(',')
lineNo = -1
if len(failDetails) == 3:
match = re.search("line.*?(\d+)", failDetails[1])
if match is None:
return lineNo
lineNo = int(match.group(1))
return lineNo
def parseDocTestResult(docTestResStr):
""" Extract the line number and filename of the doctest failure"""
lines = docTestResStr.split("\n")
for lineNo, line in enumerate(lines):
failure = line.find("Failed example:")
if failure != -1:
failDetails = lines[lineNo - 1]
yield parseFailDetails(failDetails)
<commit_msg>Fix doc test failure parsing<commit_after>
|
def parseFailDetails(failDetails):
""" Parse the line number of the doctest failure
>>> parseFailDetails("blah")
-1
"""
import re
failDetails = failDetails.split(',')
lineNo = -1
if len(failDetails) == 3:
match = re.search("line.*?(\d+)", failDetails[1])
if match is None:
return lineNo
lineNo = int(match.group(1))
return lineNo
def parseDocTestResult(docTestResStr):
""" Extract the line number and filename of the doctest failure"""
lines = docTestResStr.split("\n")
for lineNo, line in enumerate(lines):
failure = line.find("Failed example:")
if failure != -1:
failDetails = lines[lineNo - 1]
yield parseFailDetails(failDetails)
if __name__ == "__main__":
from doctest import testmod
testmod()
|
def parseFailDetails(failDetails):
""" Parse the line number of the doctest failure"""
import re
failDetails = failDetails.split(',')
lineNo = -1
if len(failDetails) == 3:
match = re.search("line.*?(\d+)", failDetails[1])
if match is None:
return lineNo
lineNo = int(match.group(1))
return lineNo
def parseDocTestResult(docTestResStr):
""" Extract the line number and filename of the doctest failure"""
lines = docTestResStr.split("\n")
for lineNo, line in enumerate(lines):
failure = line.find("Failed example:")
if failure != -1:
failDetails = lines[lineNo - 1]
yield parseFailDetails(failDetails)
Fix doc test failure parsingdef parseFailDetails(failDetails):
""" Parse the line number of the doctest failure
>>> parseFailDetails("blah")
-1
"""
import re
failDetails = failDetails.split(',')
lineNo = -1
if len(failDetails) == 3:
match = re.search("line.*?(\d+)", failDetails[1])
if match is None:
return lineNo
lineNo = int(match.group(1))
return lineNo
def parseDocTestResult(docTestResStr):
""" Extract the line number and filename of the doctest failure"""
lines = docTestResStr.split("\n")
for lineNo, line in enumerate(lines):
failure = line.find("Failed example:")
if failure != -1:
failDetails = lines[lineNo - 1]
yield parseFailDetails(failDetails)
if __name__ == "__main__":
from doctest import testmod
testmod()
|
<commit_before>def parseFailDetails(failDetails):
""" Parse the line number of the doctest failure"""
import re
failDetails = failDetails.split(',')
lineNo = -1
if len(failDetails) == 3:
match = re.search("line.*?(\d+)", failDetails[1])
if match is None:
return lineNo
lineNo = int(match.group(1))
return lineNo
def parseDocTestResult(docTestResStr):
""" Extract the line number and filename of the doctest failure"""
lines = docTestResStr.split("\n")
for lineNo, line in enumerate(lines):
failure = line.find("Failed example:")
if failure != -1:
failDetails = lines[lineNo - 1]
yield parseFailDetails(failDetails)
<commit_msg>Fix doc test failure parsing<commit_after>def parseFailDetails(failDetails):
""" Parse the line number of the doctest failure
>>> parseFailDetails("blah")
-1
"""
import re
failDetails = failDetails.split(',')
lineNo = -1
if len(failDetails) == 3:
match = re.search("line.*?(\d+)", failDetails[1])
if match is None:
return lineNo
lineNo = int(match.group(1))
return lineNo
def parseDocTestResult(docTestResStr):
""" Extract the line number and filename of the doctest failure"""
lines = docTestResStr.split("\n")
for lineNo, line in enumerate(lines):
failure = line.find("Failed example:")
if failure != -1:
failDetails = lines[lineNo - 1]
yield parseFailDetails(failDetails)
if __name__ == "__main__":
from doctest import testmod
testmod()
|
808d089b2b93671ef3d4331007fc1c3da2dea0b5
|
example/urls.py
|
example/urls.py
|
from django.conf.urls import patterns
# Uncomment the next two lines to enable the admin:
# from django.contrib import admin
# admin.autodiscover()
urlpatterns = patterns('',
# Example:
# (r'^example/', include('example.foo.urls')),
# Uncomment the admin/doc line below and add 'django.contrib.admindocs'
# to INSTALLED_APPS to enable admin documentation:
# (r'^admin/doc/', include('django.contrib.admindocs.urls')),
# Uncomment the next line to enable the admin:
# (r'^admin/(.*)', admin.site.root),
('^$', 'rpc4django.views.serve_rpc_request'),
('^RPC2$', 'rpc4django.views.serve_rpc_request'),
)
|
from django.conf.urls import patterns
from rpc4django.views import serve_rpc_request
# Uncomment the next two lines to enable the admin:
# from django.contrib import admin
# admin.autodiscover()
urlpatterns = patterns('',
# Example:
# (r'^example/', include('example.foo.urls')),
# Uncomment the admin/doc line below and add 'django.contrib.admindocs'
# to INSTALLED_APPS to enable admin documentation:
# (r'^admin/doc/', include('django.contrib.admindocs.urls')),
# Uncomment the next line to enable the admin:
# (r'^admin/(.*)', admin.site.root),
('^$', 'rpc4django.views.serve_rpc_request'),
('^RPC2$', serve_rpc_request),
)
|
Use django 1.10 patterns style
|
Use django 1.10 patterns style
|
Python
|
bsd-3-clause
|
davidfischer/rpc4django,davidfischer/rpc4django,davidfischer/rpc4django
|
from django.conf.urls import patterns
# Uncomment the next two lines to enable the admin:
# from django.contrib import admin
# admin.autodiscover()
urlpatterns = patterns('',
# Example:
# (r'^example/', include('example.foo.urls')),
# Uncomment the admin/doc line below and add 'django.contrib.admindocs'
# to INSTALLED_APPS to enable admin documentation:
# (r'^admin/doc/', include('django.contrib.admindocs.urls')),
# Uncomment the next line to enable the admin:
# (r'^admin/(.*)', admin.site.root),
('^$', 'rpc4django.views.serve_rpc_request'),
('^RPC2$', 'rpc4django.views.serve_rpc_request'),
)
Use django 1.10 patterns style
|
from django.conf.urls import patterns
from rpc4django.views import serve_rpc_request
# Uncomment the next two lines to enable the admin:
# from django.contrib import admin
# admin.autodiscover()
urlpatterns = patterns('',
# Example:
# (r'^example/', include('example.foo.urls')),
# Uncomment the admin/doc line below and add 'django.contrib.admindocs'
# to INSTALLED_APPS to enable admin documentation:
# (r'^admin/doc/', include('django.contrib.admindocs.urls')),
# Uncomment the next line to enable the admin:
# (r'^admin/(.*)', admin.site.root),
('^$', 'rpc4django.views.serve_rpc_request'),
('^RPC2$', serve_rpc_request),
)
|
<commit_before>from django.conf.urls import patterns
# Uncomment the next two lines to enable the admin:
# from django.contrib import admin
# admin.autodiscover()
urlpatterns = patterns('',
# Example:
# (r'^example/', include('example.foo.urls')),
# Uncomment the admin/doc line below and add 'django.contrib.admindocs'
# to INSTALLED_APPS to enable admin documentation:
# (r'^admin/doc/', include('django.contrib.admindocs.urls')),
# Uncomment the next line to enable the admin:
# (r'^admin/(.*)', admin.site.root),
('^$', 'rpc4django.views.serve_rpc_request'),
('^RPC2$', 'rpc4django.views.serve_rpc_request'),
)
<commit_msg>Use django 1.10 patterns style<commit_after>
|
from django.conf.urls import patterns
from rpc4django.views import serve_rpc_request
# Uncomment the next two lines to enable the admin:
# from django.contrib import admin
# admin.autodiscover()
urlpatterns = patterns('',
# Example:
# (r'^example/', include('example.foo.urls')),
# Uncomment the admin/doc line below and add 'django.contrib.admindocs'
# to INSTALLED_APPS to enable admin documentation:
# (r'^admin/doc/', include('django.contrib.admindocs.urls')),
# Uncomment the next line to enable the admin:
# (r'^admin/(.*)', admin.site.root),
('^$', 'rpc4django.views.serve_rpc_request'),
('^RPC2$', serve_rpc_request),
)
|
from django.conf.urls import patterns
# Uncomment the next two lines to enable the admin:
# from django.contrib import admin
# admin.autodiscover()
urlpatterns = patterns('',
# Example:
# (r'^example/', include('example.foo.urls')),
# Uncomment the admin/doc line below and add 'django.contrib.admindocs'
# to INSTALLED_APPS to enable admin documentation:
# (r'^admin/doc/', include('django.contrib.admindocs.urls')),
# Uncomment the next line to enable the admin:
# (r'^admin/(.*)', admin.site.root),
('^$', 'rpc4django.views.serve_rpc_request'),
('^RPC2$', 'rpc4django.views.serve_rpc_request'),
)
Use django 1.10 patterns stylefrom django.conf.urls import patterns
from rpc4django.views import serve_rpc_request
# Uncomment the next two lines to enable the admin:
# from django.contrib import admin
# admin.autodiscover()
urlpatterns = patterns('',
# Example:
# (r'^example/', include('example.foo.urls')),
# Uncomment the admin/doc line below and add 'django.contrib.admindocs'
# to INSTALLED_APPS to enable admin documentation:
# (r'^admin/doc/', include('django.contrib.admindocs.urls')),
# Uncomment the next line to enable the admin:
# (r'^admin/(.*)', admin.site.root),
('^$', 'rpc4django.views.serve_rpc_request'),
('^RPC2$', serve_rpc_request),
)
|
<commit_before>from django.conf.urls import patterns
# Uncomment the next two lines to enable the admin:
# from django.contrib import admin
# admin.autodiscover()
urlpatterns = patterns('',
# Example:
# (r'^example/', include('example.foo.urls')),
# Uncomment the admin/doc line below and add 'django.contrib.admindocs'
# to INSTALLED_APPS to enable admin documentation:
# (r'^admin/doc/', include('django.contrib.admindocs.urls')),
# Uncomment the next line to enable the admin:
# (r'^admin/(.*)', admin.site.root),
('^$', 'rpc4django.views.serve_rpc_request'),
('^RPC2$', 'rpc4django.views.serve_rpc_request'),
)
<commit_msg>Use django 1.10 patterns style<commit_after>from django.conf.urls import patterns
from rpc4django.views import serve_rpc_request
# Uncomment the next two lines to enable the admin:
# from django.contrib import admin
# admin.autodiscover()
urlpatterns = patterns('',
# Example:
# (r'^example/', include('example.foo.urls')),
# Uncomment the admin/doc line below and add 'django.contrib.admindocs'
# to INSTALLED_APPS to enable admin documentation:
# (r'^admin/doc/', include('django.contrib.admindocs.urls')),
# Uncomment the next line to enable the admin:
# (r'^admin/(.*)', admin.site.root),
('^$', 'rpc4django.views.serve_rpc_request'),
('^RPC2$', serve_rpc_request),
)
|
8e225f890fd90112a125648cbd49507340cd3224
|
events/search_indexes.py
|
events/search_indexes.py
|
from haystack import indexes
from .models import Event, Place, PublicationStatus
from django.utils.html import strip_tags
class EventIndex(indexes.SearchIndex, indexes.Indexable):
text = indexes.CharField(document=True, use_template=True)
autosuggest = indexes.EdgeNgramField(model_attr='name')
start_time = indexes.DateField(model_attr='start_time')
end_time = indexes.DateField(model_attr='end_time')
def get_updated_field(self):
return 'last_modified_time'
def get_model(self):
return Event
def prepare(self, obj):
#obj.lang_keywords = obj.keywords.filter(language=get_language())
if obj.description:
obj.description = strip_tags(obj.description)
return super(EventIndex, self).prepare(obj)
def index_queryset(self, using=None):
return self.get_model().objects.filter(publication_status=PublicationStatus.PUBLIC)
class PlaceIndex(indexes.SearchIndex, indexes.Indexable):
text = indexes.CharField(document=True, use_template=True)
autosuggest = indexes.EdgeNgramField(model_attr='name')
def get_updated_field(self):
return 'last_modified_time'
def get_model(self):
return Place
|
from haystack import indexes
from .models import Event, Place, PublicationStatus
from django.utils.html import strip_tags
class EventIndex(indexes.SearchIndex, indexes.Indexable):
text = indexes.CharField(document=True, use_template=True)
autosuggest = indexes.EdgeNgramField(model_attr='name')
start_time = indexes.DateTimeField(model_attr='start_time')
end_time = indexes.DateTimeField(model_attr='end_time')
def get_updated_field(self):
return 'last_modified_time'
def get_model(self):
return Event
def prepare(self, obj):
#obj.lang_keywords = obj.keywords.filter(language=get_language())
if obj.description:
obj.description = strip_tags(obj.description)
return super(EventIndex, self).prepare(obj)
def index_queryset(self, using=None):
return self.get_model().objects.filter(publication_status=PublicationStatus.PUBLIC)
class PlaceIndex(indexes.SearchIndex, indexes.Indexable):
text = indexes.CharField(document=True, use_template=True)
autosuggest = indexes.EdgeNgramField(model_attr='name')
def get_updated_field(self):
return 'last_modified_time'
def get_model(self):
return Place
|
Fix type of EventIndex fields
|
Fix type of EventIndex fields
|
Python
|
mit
|
tuomas777/linkedevents,aapris/linkedevents,aapris/linkedevents,City-of-Helsinki/linkedevents,City-of-Helsinki/linkedevents,tuomas777/linkedevents,aapris/linkedevents,tuomas777/linkedevents,City-of-Helsinki/linkedevents
|
from haystack import indexes
from .models import Event, Place, PublicationStatus
from django.utils.html import strip_tags
class EventIndex(indexes.SearchIndex, indexes.Indexable):
text = indexes.CharField(document=True, use_template=True)
autosuggest = indexes.EdgeNgramField(model_attr='name')
start_time = indexes.DateField(model_attr='start_time')
end_time = indexes.DateField(model_attr='end_time')
def get_updated_field(self):
return 'last_modified_time'
def get_model(self):
return Event
def prepare(self, obj):
#obj.lang_keywords = obj.keywords.filter(language=get_language())
if obj.description:
obj.description = strip_tags(obj.description)
return super(EventIndex, self).prepare(obj)
def index_queryset(self, using=None):
return self.get_model().objects.filter(publication_status=PublicationStatus.PUBLIC)
class PlaceIndex(indexes.SearchIndex, indexes.Indexable):
text = indexes.CharField(document=True, use_template=True)
autosuggest = indexes.EdgeNgramField(model_attr='name')
def get_updated_field(self):
return 'last_modified_time'
def get_model(self):
return Place
Fix type of EventIndex fields
|
from haystack import indexes
from .models import Event, Place, PublicationStatus
from django.utils.html import strip_tags
class EventIndex(indexes.SearchIndex, indexes.Indexable):
text = indexes.CharField(document=True, use_template=True)
autosuggest = indexes.EdgeNgramField(model_attr='name')
start_time = indexes.DateTimeField(model_attr='start_time')
end_time = indexes.DateTimeField(model_attr='end_time')
def get_updated_field(self):
return 'last_modified_time'
def get_model(self):
return Event
def prepare(self, obj):
#obj.lang_keywords = obj.keywords.filter(language=get_language())
if obj.description:
obj.description = strip_tags(obj.description)
return super(EventIndex, self).prepare(obj)
def index_queryset(self, using=None):
return self.get_model().objects.filter(publication_status=PublicationStatus.PUBLIC)
class PlaceIndex(indexes.SearchIndex, indexes.Indexable):
text = indexes.CharField(document=True, use_template=True)
autosuggest = indexes.EdgeNgramField(model_attr='name')
def get_updated_field(self):
return 'last_modified_time'
def get_model(self):
return Place
|
<commit_before>from haystack import indexes
from .models import Event, Place, PublicationStatus
from django.utils.html import strip_tags
class EventIndex(indexes.SearchIndex, indexes.Indexable):
text = indexes.CharField(document=True, use_template=True)
autosuggest = indexes.EdgeNgramField(model_attr='name')
start_time = indexes.DateField(model_attr='start_time')
end_time = indexes.DateField(model_attr='end_time')
def get_updated_field(self):
return 'last_modified_time'
def get_model(self):
return Event
def prepare(self, obj):
#obj.lang_keywords = obj.keywords.filter(language=get_language())
if obj.description:
obj.description = strip_tags(obj.description)
return super(EventIndex, self).prepare(obj)
def index_queryset(self, using=None):
return self.get_model().objects.filter(publication_status=PublicationStatus.PUBLIC)
class PlaceIndex(indexes.SearchIndex, indexes.Indexable):
text = indexes.CharField(document=True, use_template=True)
autosuggest = indexes.EdgeNgramField(model_attr='name')
def get_updated_field(self):
return 'last_modified_time'
def get_model(self):
return Place
<commit_msg>Fix type of EventIndex fields<commit_after>
|
from haystack import indexes
from .models import Event, Place, PublicationStatus
from django.utils.html import strip_tags
class EventIndex(indexes.SearchIndex, indexes.Indexable):
text = indexes.CharField(document=True, use_template=True)
autosuggest = indexes.EdgeNgramField(model_attr='name')
start_time = indexes.DateTimeField(model_attr='start_time')
end_time = indexes.DateTimeField(model_attr='end_time')
def get_updated_field(self):
return 'last_modified_time'
def get_model(self):
return Event
def prepare(self, obj):
#obj.lang_keywords = obj.keywords.filter(language=get_language())
if obj.description:
obj.description = strip_tags(obj.description)
return super(EventIndex, self).prepare(obj)
def index_queryset(self, using=None):
return self.get_model().objects.filter(publication_status=PublicationStatus.PUBLIC)
class PlaceIndex(indexes.SearchIndex, indexes.Indexable):
text = indexes.CharField(document=True, use_template=True)
autosuggest = indexes.EdgeNgramField(model_attr='name')
def get_updated_field(self):
return 'last_modified_time'
def get_model(self):
return Place
|
from haystack import indexes
from .models import Event, Place, PublicationStatus
from django.utils.html import strip_tags
class EventIndex(indexes.SearchIndex, indexes.Indexable):
text = indexes.CharField(document=True, use_template=True)
autosuggest = indexes.EdgeNgramField(model_attr='name')
start_time = indexes.DateField(model_attr='start_time')
end_time = indexes.DateField(model_attr='end_time')
def get_updated_field(self):
return 'last_modified_time'
def get_model(self):
return Event
def prepare(self, obj):
#obj.lang_keywords = obj.keywords.filter(language=get_language())
if obj.description:
obj.description = strip_tags(obj.description)
return super(EventIndex, self).prepare(obj)
def index_queryset(self, using=None):
return self.get_model().objects.filter(publication_status=PublicationStatus.PUBLIC)
class PlaceIndex(indexes.SearchIndex, indexes.Indexable):
text = indexes.CharField(document=True, use_template=True)
autosuggest = indexes.EdgeNgramField(model_attr='name')
def get_updated_field(self):
return 'last_modified_time'
def get_model(self):
return Place
Fix type of EventIndex fieldsfrom haystack import indexes
from .models import Event, Place, PublicationStatus
from django.utils.html import strip_tags
class EventIndex(indexes.SearchIndex, indexes.Indexable):
text = indexes.CharField(document=True, use_template=True)
autosuggest = indexes.EdgeNgramField(model_attr='name')
start_time = indexes.DateTimeField(model_attr='start_time')
end_time = indexes.DateTimeField(model_attr='end_time')
def get_updated_field(self):
return 'last_modified_time'
def get_model(self):
return Event
def prepare(self, obj):
#obj.lang_keywords = obj.keywords.filter(language=get_language())
if obj.description:
obj.description = strip_tags(obj.description)
return super(EventIndex, self).prepare(obj)
def index_queryset(self, using=None):
return self.get_model().objects.filter(publication_status=PublicationStatus.PUBLIC)
class PlaceIndex(indexes.SearchIndex, indexes.Indexable):
text = indexes.CharField(document=True, use_template=True)
autosuggest = indexes.EdgeNgramField(model_attr='name')
def get_updated_field(self):
return 'last_modified_time'
def get_model(self):
return Place
|
<commit_before>from haystack import indexes
from .models import Event, Place, PublicationStatus
from django.utils.html import strip_tags
class EventIndex(indexes.SearchIndex, indexes.Indexable):
text = indexes.CharField(document=True, use_template=True)
autosuggest = indexes.EdgeNgramField(model_attr='name')
start_time = indexes.DateField(model_attr='start_time')
end_time = indexes.DateField(model_attr='end_time')
def get_updated_field(self):
return 'last_modified_time'
def get_model(self):
return Event
def prepare(self, obj):
#obj.lang_keywords = obj.keywords.filter(language=get_language())
if obj.description:
obj.description = strip_tags(obj.description)
return super(EventIndex, self).prepare(obj)
def index_queryset(self, using=None):
return self.get_model().objects.filter(publication_status=PublicationStatus.PUBLIC)
class PlaceIndex(indexes.SearchIndex, indexes.Indexable):
text = indexes.CharField(document=True, use_template=True)
autosuggest = indexes.EdgeNgramField(model_attr='name')
def get_updated_field(self):
return 'last_modified_time'
def get_model(self):
return Place
<commit_msg>Fix type of EventIndex fields<commit_after>from haystack import indexes
from .models import Event, Place, PublicationStatus
from django.utils.html import strip_tags
class EventIndex(indexes.SearchIndex, indexes.Indexable):
text = indexes.CharField(document=True, use_template=True)
autosuggest = indexes.EdgeNgramField(model_attr='name')
start_time = indexes.DateTimeField(model_attr='start_time')
end_time = indexes.DateTimeField(model_attr='end_time')
def get_updated_field(self):
return 'last_modified_time'
def get_model(self):
return Event
def prepare(self, obj):
#obj.lang_keywords = obj.keywords.filter(language=get_language())
if obj.description:
obj.description = strip_tags(obj.description)
return super(EventIndex, self).prepare(obj)
def index_queryset(self, using=None):
return self.get_model().objects.filter(publication_status=PublicationStatus.PUBLIC)
class PlaceIndex(indexes.SearchIndex, indexes.Indexable):
text = indexes.CharField(document=True, use_template=True)
autosuggest = indexes.EdgeNgramField(model_attr='name')
def get_updated_field(self):
return 'last_modified_time'
def get_model(self):
return Place
|
40905893c296e2c812539079925adfd25e39d44f
|
wger/wsgi.py
|
wger/wsgi.py
|
"""
WSGI config for workout_manager project.
This module contains the WSGI application used by Django's development server
and any production WSGI deployments. It should expose a module-level variable
named ``application``. Django's ``runserver`` and ``runfcgi`` commands discover
this application via the ``WSGI_APPLICATION`` setting.
Usually you will have the standard Django WSGI application here, but it also
might make sense to replace the whole Django WSGI application with a custom one
that later delegates to the Django one. For example, you could introduce WSGI
middleware here, or combine a Django application with an application of another
framework.
"""
import os
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "workout_manager.settings")
# This application object is used by any WSGI server configured to use this
# file. This includes Django's development server, if the WSGI_APPLICATION
# setting points here.
from django.core.wsgi import get_wsgi_application
application = get_wsgi_application()
# Apply WSGI middleware here.
# from helloworld.wsgi import HelloWorldApplication
# application = HelloWorldApplication(application)
|
"""
WSGI config for workout_manager project.
This module contains the WSGI application used by Django's development server
and any production WSGI deployments. It should expose a module-level variable
named ``application``. Django's ``runserver`` and ``runfcgi`` commands discover
this application via the ``WSGI_APPLICATION`` setting.
Usually you will have the standard Django WSGI application here, but it also
might make sense to replace the whole Django WSGI application with a custom one
that later delegates to the Django one. For example, you could introduce WSGI
middleware here, or combine a Django application with an application of another
framework.
"""
import os
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "settings")
# This application object is used by any WSGI server configured to use this
# file. This includes Django's development server, if the WSGI_APPLICATION
# setting points here.
from django.core.wsgi import get_wsgi_application
application = get_wsgi_application()
# Apply WSGI middleware here.
# from helloworld.wsgi import HelloWorldApplication
# application = HelloWorldApplication(application)
|
Change location of default settings in WSGI
|
Change location of default settings in WSGI
|
Python
|
agpl-3.0
|
petervanderdoes/wger,rolandgeider/wger,DeveloperMal/wger,rolandgeider/wger,wger-project/wger,kjagoo/wger_stark,DeveloperMal/wger,wger-project/wger,petervanderdoes/wger,rolandgeider/wger,wger-project/wger,kjagoo/wger_stark,DeveloperMal/wger,DeveloperMal/wger,kjagoo/wger_stark,wger-project/wger,petervanderdoes/wger,rolandgeider/wger,kjagoo/wger_stark,petervanderdoes/wger
|
"""
WSGI config for workout_manager project.
This module contains the WSGI application used by Django's development server
and any production WSGI deployments. It should expose a module-level variable
named ``application``. Django's ``runserver`` and ``runfcgi`` commands discover
this application via the ``WSGI_APPLICATION`` setting.
Usually you will have the standard Django WSGI application here, but it also
might make sense to replace the whole Django WSGI application with a custom one
that later delegates to the Django one. For example, you could introduce WSGI
middleware here, or combine a Django application with an application of another
framework.
"""
import os
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "workout_manager.settings")
# This application object is used by any WSGI server configured to use this
# file. This includes Django's development server, if the WSGI_APPLICATION
# setting points here.
from django.core.wsgi import get_wsgi_application
application = get_wsgi_application()
# Apply WSGI middleware here.
# from helloworld.wsgi import HelloWorldApplication
# application = HelloWorldApplication(application)
Change location of default settings in WSGI
|
"""
WSGI config for workout_manager project.
This module contains the WSGI application used by Django's development server
and any production WSGI deployments. It should expose a module-level variable
named ``application``. Django's ``runserver`` and ``runfcgi`` commands discover
this application via the ``WSGI_APPLICATION`` setting.
Usually you will have the standard Django WSGI application here, but it also
might make sense to replace the whole Django WSGI application with a custom one
that later delegates to the Django one. For example, you could introduce WSGI
middleware here, or combine a Django application with an application of another
framework.
"""
import os
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "settings")
# This application object is used by any WSGI server configured to use this
# file. This includes Django's development server, if the WSGI_APPLICATION
# setting points here.
from django.core.wsgi import get_wsgi_application
application = get_wsgi_application()
# Apply WSGI middleware here.
# from helloworld.wsgi import HelloWorldApplication
# application = HelloWorldApplication(application)
|
<commit_before>"""
WSGI config for workout_manager project.
This module contains the WSGI application used by Django's development server
and any production WSGI deployments. It should expose a module-level variable
named ``application``. Django's ``runserver`` and ``runfcgi`` commands discover
this application via the ``WSGI_APPLICATION`` setting.
Usually you will have the standard Django WSGI application here, but it also
might make sense to replace the whole Django WSGI application with a custom one
that later delegates to the Django one. For example, you could introduce WSGI
middleware here, or combine a Django application with an application of another
framework.
"""
import os
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "workout_manager.settings")
# This application object is used by any WSGI server configured to use this
# file. This includes Django's development server, if the WSGI_APPLICATION
# setting points here.
from django.core.wsgi import get_wsgi_application
application = get_wsgi_application()
# Apply WSGI middleware here.
# from helloworld.wsgi import HelloWorldApplication
# application = HelloWorldApplication(application)
<commit_msg>Change location of default settings in WSGI<commit_after>
|
"""
WSGI config for workout_manager project.
This module contains the WSGI application used by Django's development server
and any production WSGI deployments. It should expose a module-level variable
named ``application``. Django's ``runserver`` and ``runfcgi`` commands discover
this application via the ``WSGI_APPLICATION`` setting.
Usually you will have the standard Django WSGI application here, but it also
might make sense to replace the whole Django WSGI application with a custom one
that later delegates to the Django one. For example, you could introduce WSGI
middleware here, or combine a Django application with an application of another
framework.
"""
import os
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "settings")
# This application object is used by any WSGI server configured to use this
# file. This includes Django's development server, if the WSGI_APPLICATION
# setting points here.
from django.core.wsgi import get_wsgi_application
application = get_wsgi_application()
# Apply WSGI middleware here.
# from helloworld.wsgi import HelloWorldApplication
# application = HelloWorldApplication(application)
|
"""
WSGI config for workout_manager project.
This module contains the WSGI application used by Django's development server
and any production WSGI deployments. It should expose a module-level variable
named ``application``. Django's ``runserver`` and ``runfcgi`` commands discover
this application via the ``WSGI_APPLICATION`` setting.
Usually you will have the standard Django WSGI application here, but it also
might make sense to replace the whole Django WSGI application with a custom one
that later delegates to the Django one. For example, you could introduce WSGI
middleware here, or combine a Django application with an application of another
framework.
"""
import os
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "workout_manager.settings")
# This application object is used by any WSGI server configured to use this
# file. This includes Django's development server, if the WSGI_APPLICATION
# setting points here.
from django.core.wsgi import get_wsgi_application
application = get_wsgi_application()
# Apply WSGI middleware here.
# from helloworld.wsgi import HelloWorldApplication
# application = HelloWorldApplication(application)
Change location of default settings in WSGI"""
WSGI config for workout_manager project.
This module contains the WSGI application used by Django's development server
and any production WSGI deployments. It should expose a module-level variable
named ``application``. Django's ``runserver`` and ``runfcgi`` commands discover
this application via the ``WSGI_APPLICATION`` setting.
Usually you will have the standard Django WSGI application here, but it also
might make sense to replace the whole Django WSGI application with a custom one
that later delegates to the Django one. For example, you could introduce WSGI
middleware here, or combine a Django application with an application of another
framework.
"""
import os
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "settings")
# This application object is used by any WSGI server configured to use this
# file. This includes Django's development server, if the WSGI_APPLICATION
# setting points here.
from django.core.wsgi import get_wsgi_application
application = get_wsgi_application()
# Apply WSGI middleware here.
# from helloworld.wsgi import HelloWorldApplication
# application = HelloWorldApplication(application)
|
<commit_before>"""
WSGI config for workout_manager project.
This module contains the WSGI application used by Django's development server
and any production WSGI deployments. It should expose a module-level variable
named ``application``. Django's ``runserver`` and ``runfcgi`` commands discover
this application via the ``WSGI_APPLICATION`` setting.
Usually you will have the standard Django WSGI application here, but it also
might make sense to replace the whole Django WSGI application with a custom one
that later delegates to the Django one. For example, you could introduce WSGI
middleware here, or combine a Django application with an application of another
framework.
"""
import os
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "workout_manager.settings")
# This application object is used by any WSGI server configured to use this
# file. This includes Django's development server, if the WSGI_APPLICATION
# setting points here.
from django.core.wsgi import get_wsgi_application
application = get_wsgi_application()
# Apply WSGI middleware here.
# from helloworld.wsgi import HelloWorldApplication
# application = HelloWorldApplication(application)
<commit_msg>Change location of default settings in WSGI<commit_after>"""
WSGI config for workout_manager project.
This module contains the WSGI application used by Django's development server
and any production WSGI deployments. It should expose a module-level variable
named ``application``. Django's ``runserver`` and ``runfcgi`` commands discover
this application via the ``WSGI_APPLICATION`` setting.
Usually you will have the standard Django WSGI application here, but it also
might make sense to replace the whole Django WSGI application with a custom one
that later delegates to the Django one. For example, you could introduce WSGI
middleware here, or combine a Django application with an application of another
framework.
"""
import os
os.environ.setdefault("DJANGO_SETTINGS_MODULE", "settings")
# This application object is used by any WSGI server configured to use this
# file. This includes Django's development server, if the WSGI_APPLICATION
# setting points here.
from django.core.wsgi import get_wsgi_application
application = get_wsgi_application()
# Apply WSGI middleware here.
# from helloworld.wsgi import HelloWorldApplication
# application = HelloWorldApplication(application)
|
98fad1af84abe13eb64baad58c8a2faf3cd6cccb
|
tt_dailyemailblast/tasks.py
|
tt_dailyemailblast/tasks.py
|
from celery.task import task
from . import models
from . import send_backends
@task
def send_daily_email_blasts(blast_pk):
blast = models.DailyEmailBlast.objects.get(pk=blast_pk)
send_backends.sync_daily_email_blasts(blast)
@task
def send_recipients_list(recipients_list_pk, blast_pk):
blast = models.DailyEmailBlast.objects.get(pk=blast_pk)
recipients_list = models.RecipientList.objects.get(pk=recipients_list_pk)
send_backends.sync_recipients_list(recipients_list, blast)
@task
def send_recipients(recipient_pk, recipients_list_pk, blast_pk):
blast = models.DailyEmailBlast.objects.get(pk=blast_pk)
recipients_list = models.RecipientList.objects.get(pk=recipients_list_pk)
recipient = models.Recipient.objects.get(pk=recipient_pk)
send_backends.sync_recipient(recipient, recipients_list, blast)
|
from celery.task import task
from . import models
from .send_backends import sync
@task
def send_daily_email_blasts(blast_pk):
blast = models.DailyEmailBlast.objects.get(pk=blast_pk)
sync.sync_daily_email_blasts(blast)
@task
def send_recipients_list(recipients_list_pk, blast_pk):
blast = models.DailyEmailBlast.objects.get(pk=blast_pk)
recipients_list = models.RecipientList.objects.get(pk=recipients_list_pk)
sync.sync_recipients_list(recipients_list, blast)
@task
def send_recipients(recipient_pk, recipients_list_pk, blast_pk):
blast = models.DailyEmailBlast.objects.get(pk=blast_pk)
recipients_list = models.RecipientList.objects.get(pk=recipients_list_pk)
recipient = models.Recipient.objects.get(pk=recipient_pk)
sync.sync_recipient(recipient, recipients_list, blast)
|
Fix every single async task was broken
|
Fix every single async task was broken
|
Python
|
apache-2.0
|
texastribune/tt_dailyemailblast,texastribune/tt_dailyemailblast
|
from celery.task import task
from . import models
from . import send_backends
@task
def send_daily_email_blasts(blast_pk):
blast = models.DailyEmailBlast.objects.get(pk=blast_pk)
send_backends.sync_daily_email_blasts(blast)
@task
def send_recipients_list(recipients_list_pk, blast_pk):
blast = models.DailyEmailBlast.objects.get(pk=blast_pk)
recipients_list = models.RecipientList.objects.get(pk=recipients_list_pk)
send_backends.sync_recipients_list(recipients_list, blast)
@task
def send_recipients(recipient_pk, recipients_list_pk, blast_pk):
blast = models.DailyEmailBlast.objects.get(pk=blast_pk)
recipients_list = models.RecipientList.objects.get(pk=recipients_list_pk)
recipient = models.Recipient.objects.get(pk=recipient_pk)
send_backends.sync_recipient(recipient, recipients_list, blast)
Fix every single async task was broken
|
from celery.task import task
from . import models
from .send_backends import sync
@task
def send_daily_email_blasts(blast_pk):
blast = models.DailyEmailBlast.objects.get(pk=blast_pk)
sync.sync_daily_email_blasts(blast)
@task
def send_recipients_list(recipients_list_pk, blast_pk):
blast = models.DailyEmailBlast.objects.get(pk=blast_pk)
recipients_list = models.RecipientList.objects.get(pk=recipients_list_pk)
sync.sync_recipients_list(recipients_list, blast)
@task
def send_recipients(recipient_pk, recipients_list_pk, blast_pk):
blast = models.DailyEmailBlast.objects.get(pk=blast_pk)
recipients_list = models.RecipientList.objects.get(pk=recipients_list_pk)
recipient = models.Recipient.objects.get(pk=recipient_pk)
sync.sync_recipient(recipient, recipients_list, blast)
|
<commit_before>from celery.task import task
from . import models
from . import send_backends
@task
def send_daily_email_blasts(blast_pk):
blast = models.DailyEmailBlast.objects.get(pk=blast_pk)
send_backends.sync_daily_email_blasts(blast)
@task
def send_recipients_list(recipients_list_pk, blast_pk):
blast = models.DailyEmailBlast.objects.get(pk=blast_pk)
recipients_list = models.RecipientList.objects.get(pk=recipients_list_pk)
send_backends.sync_recipients_list(recipients_list, blast)
@task
def send_recipients(recipient_pk, recipients_list_pk, blast_pk):
blast = models.DailyEmailBlast.objects.get(pk=blast_pk)
recipients_list = models.RecipientList.objects.get(pk=recipients_list_pk)
recipient = models.Recipient.objects.get(pk=recipient_pk)
send_backends.sync_recipient(recipient, recipients_list, blast)
<commit_msg>Fix every single async task was broken<commit_after>
|
from celery.task import task
from . import models
from .send_backends import sync
@task
def send_daily_email_blasts(blast_pk):
blast = models.DailyEmailBlast.objects.get(pk=blast_pk)
sync.sync_daily_email_blasts(blast)
@task
def send_recipients_list(recipients_list_pk, blast_pk):
blast = models.DailyEmailBlast.objects.get(pk=blast_pk)
recipients_list = models.RecipientList.objects.get(pk=recipients_list_pk)
sync.sync_recipients_list(recipients_list, blast)
@task
def send_recipients(recipient_pk, recipients_list_pk, blast_pk):
blast = models.DailyEmailBlast.objects.get(pk=blast_pk)
recipients_list = models.RecipientList.objects.get(pk=recipients_list_pk)
recipient = models.Recipient.objects.get(pk=recipient_pk)
sync.sync_recipient(recipient, recipients_list, blast)
|
from celery.task import task
from . import models
from . import send_backends
@task
def send_daily_email_blasts(blast_pk):
blast = models.DailyEmailBlast.objects.get(pk=blast_pk)
send_backends.sync_daily_email_blasts(blast)
@task
def send_recipients_list(recipients_list_pk, blast_pk):
blast = models.DailyEmailBlast.objects.get(pk=blast_pk)
recipients_list = models.RecipientList.objects.get(pk=recipients_list_pk)
send_backends.sync_recipients_list(recipients_list, blast)
@task
def send_recipients(recipient_pk, recipients_list_pk, blast_pk):
blast = models.DailyEmailBlast.objects.get(pk=blast_pk)
recipients_list = models.RecipientList.objects.get(pk=recipients_list_pk)
recipient = models.Recipient.objects.get(pk=recipient_pk)
send_backends.sync_recipient(recipient, recipients_list, blast)
Fix every single async task was brokenfrom celery.task import task
from . import models
from .send_backends import sync
@task
def send_daily_email_blasts(blast_pk):
blast = models.DailyEmailBlast.objects.get(pk=blast_pk)
sync.sync_daily_email_blasts(blast)
@task
def send_recipients_list(recipients_list_pk, blast_pk):
blast = models.DailyEmailBlast.objects.get(pk=blast_pk)
recipients_list = models.RecipientList.objects.get(pk=recipients_list_pk)
sync.sync_recipients_list(recipients_list, blast)
@task
def send_recipients(recipient_pk, recipients_list_pk, blast_pk):
blast = models.DailyEmailBlast.objects.get(pk=blast_pk)
recipients_list = models.RecipientList.objects.get(pk=recipients_list_pk)
recipient = models.Recipient.objects.get(pk=recipient_pk)
sync.sync_recipient(recipient, recipients_list, blast)
|
<commit_before>from celery.task import task
from . import models
from . import send_backends
@task
def send_daily_email_blasts(blast_pk):
blast = models.DailyEmailBlast.objects.get(pk=blast_pk)
send_backends.sync_daily_email_blasts(blast)
@task
def send_recipients_list(recipients_list_pk, blast_pk):
blast = models.DailyEmailBlast.objects.get(pk=blast_pk)
recipients_list = models.RecipientList.objects.get(pk=recipients_list_pk)
send_backends.sync_recipients_list(recipients_list, blast)
@task
def send_recipients(recipient_pk, recipients_list_pk, blast_pk):
blast = models.DailyEmailBlast.objects.get(pk=blast_pk)
recipients_list = models.RecipientList.objects.get(pk=recipients_list_pk)
recipient = models.Recipient.objects.get(pk=recipient_pk)
send_backends.sync_recipient(recipient, recipients_list, blast)
<commit_msg>Fix every single async task was broken<commit_after>from celery.task import task
from . import models
from .send_backends import sync
@task
def send_daily_email_blasts(blast_pk):
blast = models.DailyEmailBlast.objects.get(pk=blast_pk)
sync.sync_daily_email_blasts(blast)
@task
def send_recipients_list(recipients_list_pk, blast_pk):
blast = models.DailyEmailBlast.objects.get(pk=blast_pk)
recipients_list = models.RecipientList.objects.get(pk=recipients_list_pk)
sync.sync_recipients_list(recipients_list, blast)
@task
def send_recipients(recipient_pk, recipients_list_pk, blast_pk):
blast = models.DailyEmailBlast.objects.get(pk=blast_pk)
recipients_list = models.RecipientList.objects.get(pk=recipients_list_pk)
recipient = models.Recipient.objects.get(pk=recipient_pk)
sync.sync_recipient(recipient, recipients_list, blast)
|
d33a624fa6aedb93ae43ba1d2c0f6a76d90ff4a6
|
foldermd5sums.py
|
foldermd5sums.py
|
#!/usr/bin/env python
"""Script to read data files in a directory, compute their md5sums, and output
them to a JSON file."""
import json
import os
import sys
import hashlib
def get_md5sums(directory):
md5sums = []
for filename in os.listdir(directory):
md5 = hashlib.md5()
with open(os.path.join(directory, filename), 'rb') as fp:
for chunk in iter(lambda: fp.read(128 * md5.block_size), b''):
md5.update(chunk)
md5hash = md5.hexdigest()
md5sums.append((filename, md5hash))
return md5sums
if __name__ == '__main__':
if len(sys.argv) < 3:
print('Usage: ' + sys.argv[0] + ' input_directory output.json')
sys.exit(1)
directory = sys.argv[1]
if not os.path.exists(directory):
print('Directory does not exist!')
sys.exit(1)
output_json = sys.argv[2]
md5sums = get_md5sums(directory)
with open(output_json, 'w') as fp:
json.dump(md5sums, fp, indent=0)
|
#!/usr/bin/env python
"""Script to read data files in a directory, compute their md5sums, and output
them to a JSON file."""
import json
import os
import sys
import hashlib
def get_relative_filepaths(base_directory):
""" Return a list of file paths without the base_directory prefix"""
file_list = []
for root, subFolders, files in os.walk('Data'):
relative_path="/".join(root.split('/')[1:])
for file in files:
file_list.append(os.path.join(relative_path,file))
return file_list
def get_md5sums(base_directory):
md5sums = []
for filename in get_relative_filepaths(base_directory):
md5 = hashlib.md5()
full_filepath = os.path.join(base_directory, filename)
with open(full_filepath, 'rb') as fp:
for chunk in iter(lambda: fp.read(128 * md5.block_size), b''):
md5.update(chunk)
md5hash = md5.hexdigest()
md5sums.append((filename, md5hash))
return md5sums
if __name__ == '__main__':
if len(sys.argv) < 3:
print('Usage: ' + sys.argv[0] + ' input_directory output.json')
sys.exit(1)
directory = sys.argv[1]
if not os.path.exists(directory):
print('Directory does not exist!')
sys.exit(1)
output_json = sys.argv[2]
md5sums = get_md5sums(directory)
with open(output_json, 'w') as fp:
json.dump(md5sums, fp, indent=0)
|
Allow directory of files to be indexed
|
ENH: Allow directory of files to be indexed
In the Data directory, there may be sub-directories of files that need to
be kept separate, but all of them need to be indexed.
|
Python
|
apache-2.0
|
zivy/SimpleITK-Notebooks,InsightSoftwareConsortium/SimpleITK-Notebooks,InsightSoftwareConsortium/SimpleITK-Notebooks,zivy/SimpleITK-Notebooks,InsightSoftwareConsortium/SimpleITK-Notebooks,thewtex/SimpleITK-Notebooks,zivy/SimpleITK-Notebooks,thewtex/SimpleITK-Notebooks,thewtex/SimpleITK-Notebooks
|
#!/usr/bin/env python
"""Script to read data files in a directory, compute their md5sums, and output
them to a JSON file."""
import json
import os
import sys
import hashlib
def get_md5sums(directory):
md5sums = []
for filename in os.listdir(directory):
md5 = hashlib.md5()
with open(os.path.join(directory, filename), 'rb') as fp:
for chunk in iter(lambda: fp.read(128 * md5.block_size), b''):
md5.update(chunk)
md5hash = md5.hexdigest()
md5sums.append((filename, md5hash))
return md5sums
if __name__ == '__main__':
if len(sys.argv) < 3:
print('Usage: ' + sys.argv[0] + ' input_directory output.json')
sys.exit(1)
directory = sys.argv[1]
if not os.path.exists(directory):
print('Directory does not exist!')
sys.exit(1)
output_json = sys.argv[2]
md5sums = get_md5sums(directory)
with open(output_json, 'w') as fp:
json.dump(md5sums, fp, indent=0)
ENH: Allow directory of files to be indexed
In the Data directory, there may be sub-directories of files that need to
be kept separate, but all of them need to be indexed.
|
#!/usr/bin/env python
"""Script to read data files in a directory, compute their md5sums, and output
them to a JSON file."""
import json
import os
import sys
import hashlib
def get_relative_filepaths(base_directory):
""" Return a list of file paths without the base_directory prefix"""
file_list = []
for root, subFolders, files in os.walk('Data'):
relative_path="/".join(root.split('/')[1:])
for file in files:
file_list.append(os.path.join(relative_path,file))
return file_list
def get_md5sums(base_directory):
md5sums = []
for filename in get_relative_filepaths(base_directory):
md5 = hashlib.md5()
full_filepath = os.path.join(base_directory, filename)
with open(full_filepath, 'rb') as fp:
for chunk in iter(lambda: fp.read(128 * md5.block_size), b''):
md5.update(chunk)
md5hash = md5.hexdigest()
md5sums.append((filename, md5hash))
return md5sums
if __name__ == '__main__':
if len(sys.argv) < 3:
print('Usage: ' + sys.argv[0] + ' input_directory output.json')
sys.exit(1)
directory = sys.argv[1]
if not os.path.exists(directory):
print('Directory does not exist!')
sys.exit(1)
output_json = sys.argv[2]
md5sums = get_md5sums(directory)
with open(output_json, 'w') as fp:
json.dump(md5sums, fp, indent=0)
|
<commit_before>#!/usr/bin/env python
"""Script to read data files in a directory, compute their md5sums, and output
them to a JSON file."""
import json
import os
import sys
import hashlib
def get_md5sums(directory):
md5sums = []
for filename in os.listdir(directory):
md5 = hashlib.md5()
with open(os.path.join(directory, filename), 'rb') as fp:
for chunk in iter(lambda: fp.read(128 * md5.block_size), b''):
md5.update(chunk)
md5hash = md5.hexdigest()
md5sums.append((filename, md5hash))
return md5sums
if __name__ == '__main__':
if len(sys.argv) < 3:
print('Usage: ' + sys.argv[0] + ' input_directory output.json')
sys.exit(1)
directory = sys.argv[1]
if not os.path.exists(directory):
print('Directory does not exist!')
sys.exit(1)
output_json = sys.argv[2]
md5sums = get_md5sums(directory)
with open(output_json, 'w') as fp:
json.dump(md5sums, fp, indent=0)
<commit_msg>ENH: Allow directory of files to be indexed
In the Data directory, there may be sub-directories of files that need to
be kept separate, but all of them need to be indexed.<commit_after>
|
#!/usr/bin/env python
"""Script to read data files in a directory, compute their md5sums, and output
them to a JSON file."""
import json
import os
import sys
import hashlib
def get_relative_filepaths(base_directory):
""" Return a list of file paths without the base_directory prefix"""
file_list = []
for root, subFolders, files in os.walk('Data'):
relative_path="/".join(root.split('/')[1:])
for file in files:
file_list.append(os.path.join(relative_path,file))
return file_list
def get_md5sums(base_directory):
md5sums = []
for filename in get_relative_filepaths(base_directory):
md5 = hashlib.md5()
full_filepath = os.path.join(base_directory, filename)
with open(full_filepath, 'rb') as fp:
for chunk in iter(lambda: fp.read(128 * md5.block_size), b''):
md5.update(chunk)
md5hash = md5.hexdigest()
md5sums.append((filename, md5hash))
return md5sums
if __name__ == '__main__':
if len(sys.argv) < 3:
print('Usage: ' + sys.argv[0] + ' input_directory output.json')
sys.exit(1)
directory = sys.argv[1]
if not os.path.exists(directory):
print('Directory does not exist!')
sys.exit(1)
output_json = sys.argv[2]
md5sums = get_md5sums(directory)
with open(output_json, 'w') as fp:
json.dump(md5sums, fp, indent=0)
|
#!/usr/bin/env python
"""Script to read data files in a directory, compute their md5sums, and output
them to a JSON file."""
import json
import os
import sys
import hashlib
def get_md5sums(directory):
md5sums = []
for filename in os.listdir(directory):
md5 = hashlib.md5()
with open(os.path.join(directory, filename), 'rb') as fp:
for chunk in iter(lambda: fp.read(128 * md5.block_size), b''):
md5.update(chunk)
md5hash = md5.hexdigest()
md5sums.append((filename, md5hash))
return md5sums
if __name__ == '__main__':
if len(sys.argv) < 3:
print('Usage: ' + sys.argv[0] + ' input_directory output.json')
sys.exit(1)
directory = sys.argv[1]
if not os.path.exists(directory):
print('Directory does not exist!')
sys.exit(1)
output_json = sys.argv[2]
md5sums = get_md5sums(directory)
with open(output_json, 'w') as fp:
json.dump(md5sums, fp, indent=0)
ENH: Allow directory of files to be indexed
In the Data directory, there may be sub-directories of files that need to
be kept separate, but all of them need to be indexed.#!/usr/bin/env python
"""Script to read data files in a directory, compute their md5sums, and output
them to a JSON file."""
import json
import os
import sys
import hashlib
def get_relative_filepaths(base_directory):
""" Return a list of file paths without the base_directory prefix"""
file_list = []
for root, subFolders, files in os.walk('Data'):
relative_path="/".join(root.split('/')[1:])
for file in files:
file_list.append(os.path.join(relative_path,file))
return file_list
def get_md5sums(base_directory):
md5sums = []
for filename in get_relative_filepaths(base_directory):
md5 = hashlib.md5()
full_filepath = os.path.join(base_directory, filename)
with open(full_filepath, 'rb') as fp:
for chunk in iter(lambda: fp.read(128 * md5.block_size), b''):
md5.update(chunk)
md5hash = md5.hexdigest()
md5sums.append((filename, md5hash))
return md5sums
if __name__ == '__main__':
if len(sys.argv) < 3:
print('Usage: ' + sys.argv[0] + ' input_directory output.json')
sys.exit(1)
directory = sys.argv[1]
if not os.path.exists(directory):
print('Directory does not exist!')
sys.exit(1)
output_json = sys.argv[2]
md5sums = get_md5sums(directory)
with open(output_json, 'w') as fp:
json.dump(md5sums, fp, indent=0)
|
<commit_before>#!/usr/bin/env python
"""Script to read data files in a directory, compute their md5sums, and output
them to a JSON file."""
import json
import os
import sys
import hashlib
def get_md5sums(directory):
md5sums = []
for filename in os.listdir(directory):
md5 = hashlib.md5()
with open(os.path.join(directory, filename), 'rb') as fp:
for chunk in iter(lambda: fp.read(128 * md5.block_size), b''):
md5.update(chunk)
md5hash = md5.hexdigest()
md5sums.append((filename, md5hash))
return md5sums
if __name__ == '__main__':
if len(sys.argv) < 3:
print('Usage: ' + sys.argv[0] + ' input_directory output.json')
sys.exit(1)
directory = sys.argv[1]
if not os.path.exists(directory):
print('Directory does not exist!')
sys.exit(1)
output_json = sys.argv[2]
md5sums = get_md5sums(directory)
with open(output_json, 'w') as fp:
json.dump(md5sums, fp, indent=0)
<commit_msg>ENH: Allow directory of files to be indexed
In the Data directory, there may be sub-directories of files that need to
be kept separate, but all of them need to be indexed.<commit_after>#!/usr/bin/env python
"""Script to read data files in a directory, compute their md5sums, and output
them to a JSON file."""
import json
import os
import sys
import hashlib
def get_relative_filepaths(base_directory):
""" Return a list of file paths without the base_directory prefix"""
file_list = []
for root, subFolders, files in os.walk('Data'):
relative_path="/".join(root.split('/')[1:])
for file in files:
file_list.append(os.path.join(relative_path,file))
return file_list
def get_md5sums(base_directory):
md5sums = []
for filename in get_relative_filepaths(base_directory):
md5 = hashlib.md5()
full_filepath = os.path.join(base_directory, filename)
with open(full_filepath, 'rb') as fp:
for chunk in iter(lambda: fp.read(128 * md5.block_size), b''):
md5.update(chunk)
md5hash = md5.hexdigest()
md5sums.append((filename, md5hash))
return md5sums
if __name__ == '__main__':
if len(sys.argv) < 3:
print('Usage: ' + sys.argv[0] + ' input_directory output.json')
sys.exit(1)
directory = sys.argv[1]
if not os.path.exists(directory):
print('Directory does not exist!')
sys.exit(1)
output_json = sys.argv[2]
md5sums = get_md5sums(directory)
with open(output_json, 'w') as fp:
json.dump(md5sums, fp, indent=0)
|
cbefb84542d9dfddd0f2fdf8bd0cb2fc89d5b824
|
jupytext/__init__.py
|
jupytext/__init__.py
|
"""Read and write Jupyter notebooks as text files"""
from .jupytext import readf, writef, writes, reads
from .formats import NOTEBOOK_EXTENSIONS, guess_format, get_format_implementation
from .version import __version__
try:
from .contentsmanager import TextFileContentsManager
except ImportError as err:
class TextFileContentsManager:
"""A class that raises the previous ImportError"""
err = err
def __init__(self):
raise self.err
__all__ = ['readf', 'writef', 'writes', 'reads',
'NOTEBOOK_EXTENSIONS', 'guess_format', 'get_format_implementation',
'TextFileContentsManager', '__version__']
|
"""Read and write Jupyter notebooks as text files"""
from .jupytext import readf, writef, writes, reads
from .formats import NOTEBOOK_EXTENSIONS, guess_format, get_format_implementation
from .version import __version__
try:
from .contentsmanager import TextFileContentsManager
except ImportError as err:
class TextFileContentsManager:
"""A class that raises the previous ImportError"""
err = err
def __init__(self):
raise self.err
def _jupyter_nbextension_paths():
"""Allows commands like
jupyter nbextension install --py jupytext
jupyter nbextension enable --py jupytext
jupyter labextension install jupyterlab-jupytext"""
return [dict(
section="notebook",
# the path is relative to the `jupytext` directory
src="nbextension",
# directory in the `nbextension/` namespace
dest="jupytext",
# _also_ in the `nbextension/` namespace
require="jupytext/index")]
__all__ = ['readf', 'writef', 'writes', 'reads',
'NOTEBOOK_EXTENSIONS', 'guess_format', 'get_format_implementation',
'TextFileContentsManager', '__version__']
|
Allow "jupyter nbextension install/enable --py jupytext"
|
Allow "jupyter nbextension install/enable --py jupytext"
|
Python
|
mit
|
mwouts/jupytext,mwouts/jupytext,mwouts/jupytext,mwouts/jupytext,mwouts/jupytext,mwouts/jupytext,mwouts/jupytext,mwouts/jupytext,mwouts/jupytext,mwouts/jupytext
|
"""Read and write Jupyter notebooks as text files"""
from .jupytext import readf, writef, writes, reads
from .formats import NOTEBOOK_EXTENSIONS, guess_format, get_format_implementation
from .version import __version__
try:
from .contentsmanager import TextFileContentsManager
except ImportError as err:
class TextFileContentsManager:
"""A class that raises the previous ImportError"""
err = err
def __init__(self):
raise self.err
__all__ = ['readf', 'writef', 'writes', 'reads',
'NOTEBOOK_EXTENSIONS', 'guess_format', 'get_format_implementation',
'TextFileContentsManager', '__version__']
Allow "jupyter nbextension install/enable --py jupytext"
|
"""Read and write Jupyter notebooks as text files"""
from .jupytext import readf, writef, writes, reads
from .formats import NOTEBOOK_EXTENSIONS, guess_format, get_format_implementation
from .version import __version__
try:
from .contentsmanager import TextFileContentsManager
except ImportError as err:
class TextFileContentsManager:
"""A class that raises the previous ImportError"""
err = err
def __init__(self):
raise self.err
def _jupyter_nbextension_paths():
"""Allows commands like
jupyter nbextension install --py jupytext
jupyter nbextension enable --py jupytext
jupyter labextension install jupyterlab-jupytext"""
return [dict(
section="notebook",
# the path is relative to the `jupytext` directory
src="nbextension",
# directory in the `nbextension/` namespace
dest="jupytext",
# _also_ in the `nbextension/` namespace
require="jupytext/index")]
__all__ = ['readf', 'writef', 'writes', 'reads',
'NOTEBOOK_EXTENSIONS', 'guess_format', 'get_format_implementation',
'TextFileContentsManager', '__version__']
|
<commit_before>"""Read and write Jupyter notebooks as text files"""
from .jupytext import readf, writef, writes, reads
from .formats import NOTEBOOK_EXTENSIONS, guess_format, get_format_implementation
from .version import __version__
try:
from .contentsmanager import TextFileContentsManager
except ImportError as err:
class TextFileContentsManager:
"""A class that raises the previous ImportError"""
err = err
def __init__(self):
raise self.err
__all__ = ['readf', 'writef', 'writes', 'reads',
'NOTEBOOK_EXTENSIONS', 'guess_format', 'get_format_implementation',
'TextFileContentsManager', '__version__']
<commit_msg>Allow "jupyter nbextension install/enable --py jupytext"<commit_after>
|
"""Read and write Jupyter notebooks as text files"""
from .jupytext import readf, writef, writes, reads
from .formats import NOTEBOOK_EXTENSIONS, guess_format, get_format_implementation
from .version import __version__
try:
from .contentsmanager import TextFileContentsManager
except ImportError as err:
class TextFileContentsManager:
"""A class that raises the previous ImportError"""
err = err
def __init__(self):
raise self.err
def _jupyter_nbextension_paths():
"""Allows commands like
jupyter nbextension install --py jupytext
jupyter nbextension enable --py jupytext
jupyter labextension install jupyterlab-jupytext"""
return [dict(
section="notebook",
# the path is relative to the `jupytext` directory
src="nbextension",
# directory in the `nbextension/` namespace
dest="jupytext",
# _also_ in the `nbextension/` namespace
require="jupytext/index")]
__all__ = ['readf', 'writef', 'writes', 'reads',
'NOTEBOOK_EXTENSIONS', 'guess_format', 'get_format_implementation',
'TextFileContentsManager', '__version__']
|
"""Read and write Jupyter notebooks as text files"""
from .jupytext import readf, writef, writes, reads
from .formats import NOTEBOOK_EXTENSIONS, guess_format, get_format_implementation
from .version import __version__
try:
from .contentsmanager import TextFileContentsManager
except ImportError as err:
class TextFileContentsManager:
"""A class that raises the previous ImportError"""
err = err
def __init__(self):
raise self.err
__all__ = ['readf', 'writef', 'writes', 'reads',
'NOTEBOOK_EXTENSIONS', 'guess_format', 'get_format_implementation',
'TextFileContentsManager', '__version__']
Allow "jupyter nbextension install/enable --py jupytext""""Read and write Jupyter notebooks as text files"""
from .jupytext import readf, writef, writes, reads
from .formats import NOTEBOOK_EXTENSIONS, guess_format, get_format_implementation
from .version import __version__
try:
from .contentsmanager import TextFileContentsManager
except ImportError as err:
class TextFileContentsManager:
"""A class that raises the previous ImportError"""
err = err
def __init__(self):
raise self.err
def _jupyter_nbextension_paths():
"""Allows commands like
jupyter nbextension install --py jupytext
jupyter nbextension enable --py jupytext
jupyter labextension install jupyterlab-jupytext"""
return [dict(
section="notebook",
# the path is relative to the `jupytext` directory
src="nbextension",
# directory in the `nbextension/` namespace
dest="jupytext",
# _also_ in the `nbextension/` namespace
require="jupytext/index")]
__all__ = ['readf', 'writef', 'writes', 'reads',
'NOTEBOOK_EXTENSIONS', 'guess_format', 'get_format_implementation',
'TextFileContentsManager', '__version__']
|
<commit_before>"""Read and write Jupyter notebooks as text files"""
from .jupytext import readf, writef, writes, reads
from .formats import NOTEBOOK_EXTENSIONS, guess_format, get_format_implementation
from .version import __version__
try:
from .contentsmanager import TextFileContentsManager
except ImportError as err:
class TextFileContentsManager:
"""A class that raises the previous ImportError"""
err = err
def __init__(self):
raise self.err
__all__ = ['readf', 'writef', 'writes', 'reads',
'NOTEBOOK_EXTENSIONS', 'guess_format', 'get_format_implementation',
'TextFileContentsManager', '__version__']
<commit_msg>Allow "jupyter nbextension install/enable --py jupytext"<commit_after>"""Read and write Jupyter notebooks as text files"""
from .jupytext import readf, writef, writes, reads
from .formats import NOTEBOOK_EXTENSIONS, guess_format, get_format_implementation
from .version import __version__
try:
from .contentsmanager import TextFileContentsManager
except ImportError as err:
class TextFileContentsManager:
"""A class that raises the previous ImportError"""
err = err
def __init__(self):
raise self.err
def _jupyter_nbextension_paths():
"""Allows commands like
jupyter nbextension install --py jupytext
jupyter nbextension enable --py jupytext
jupyter labextension install jupyterlab-jupytext"""
return [dict(
section="notebook",
# the path is relative to the `jupytext` directory
src="nbextension",
# directory in the `nbextension/` namespace
dest="jupytext",
# _also_ in the `nbextension/` namespace
require="jupytext/index")]
__all__ = ['readf', 'writef', 'writes', 'reads',
'NOTEBOOK_EXTENSIONS', 'guess_format', 'get_format_implementation',
'TextFileContentsManager', '__version__']
|
20d21b851d02bbcf8c6a0f065b9f05f5e0bfc662
|
geodj/youtube.py
|
geodj/youtube.py
|
from gdata.youtube.service import YouTubeService, YouTubeVideoQuery
from django.utils.encoding import smart_str
import re
class YoutubeMusic:
def __init__(self):
self.service = YouTubeService()
def search(self, artist):
query = YouTubeVideoQuery()
query.vq = artist
query.orderby = 'relevance'
query.racy = 'exclude'
query.categories.append("/Music")
feed = self.service.YouTubeQuery(query)
results = []
for entry in feed.entry:
if not self.is_valid_entry(artist, entry):
continue
results.append({
'url': entry.media.player.url,
'title': smart_str(entry.media.title.text),
'duration': int(entry.media.duration.seconds),
})
return {'artist': artist, 'results': results}
def is_valid_entry(self, artist, entry):
duration = int(entry.media.duration.seconds)
title = smart_str(entry.media.title.text).lower()
if entry.rating is not None and float(entry.rating.average) < 3.5:
return False
if duration < (2 * 60) or duration > (9 * 60):
return False
if artist.lower() not in title:
return False
if re.search("\b(concert|cover)\b", title):
return False
return True
|
from gdata.youtube.service import YouTubeService, YouTubeVideoQuery
from django.utils.encoding import smart_str
import re
class YoutubeMusic:
def __init__(self):
self.service = YouTubeService()
def search(self, artist):
query = YouTubeVideoQuery()
query.vq = artist
query.orderby = 'relevance'
query.racy = 'exclude'
query.format = '5'
query.categories.append("/Music")
feed = self.service.YouTubeQuery(query)
results = []
for entry in feed.entry:
if not self.is_valid_entry(artist, entry):
continue
results.append({
'url': entry.media.player.url,
'title': smart_str(entry.media.title.text),
'duration': int(entry.media.duration.seconds),
})
return {'artist': artist, 'results': results}
def is_valid_entry(self, artist, entry):
duration = int(entry.media.duration.seconds)
title = smart_str(entry.media.title.text).lower()
if entry.rating is not None and float(entry.rating.average) < 3.5:
return False
if duration < (2 * 60) or duration > (9 * 60):
return False
if artist.lower() not in title:
return False
if re.search("\b(concert|cover)\b", title):
return False
return True
|
Use format=5 in YT search to prevent "embedding disabled"
|
Use format=5 in YT search to prevent "embedding disabled"
|
Python
|
mit
|
6/GeoDJ,6/GeoDJ
|
from gdata.youtube.service import YouTubeService, YouTubeVideoQuery
from django.utils.encoding import smart_str
import re
class YoutubeMusic:
def __init__(self):
self.service = YouTubeService()
def search(self, artist):
query = YouTubeVideoQuery()
query.vq = artist
query.orderby = 'relevance'
query.racy = 'exclude'
query.categories.append("/Music")
feed = self.service.YouTubeQuery(query)
results = []
for entry in feed.entry:
if not self.is_valid_entry(artist, entry):
continue
results.append({
'url': entry.media.player.url,
'title': smart_str(entry.media.title.text),
'duration': int(entry.media.duration.seconds),
})
return {'artist': artist, 'results': results}
def is_valid_entry(self, artist, entry):
duration = int(entry.media.duration.seconds)
title = smart_str(entry.media.title.text).lower()
if entry.rating is not None and float(entry.rating.average) < 3.5:
return False
if duration < (2 * 60) or duration > (9 * 60):
return False
if artist.lower() not in title:
return False
if re.search("\b(concert|cover)\b", title):
return False
return True
Use format=5 in YT search to prevent "embedding disabled"
|
from gdata.youtube.service import YouTubeService, YouTubeVideoQuery
from django.utils.encoding import smart_str
import re
class YoutubeMusic:
def __init__(self):
self.service = YouTubeService()
def search(self, artist):
query = YouTubeVideoQuery()
query.vq = artist
query.orderby = 'relevance'
query.racy = 'exclude'
query.format = '5'
query.categories.append("/Music")
feed = self.service.YouTubeQuery(query)
results = []
for entry in feed.entry:
if not self.is_valid_entry(artist, entry):
continue
results.append({
'url': entry.media.player.url,
'title': smart_str(entry.media.title.text),
'duration': int(entry.media.duration.seconds),
})
return {'artist': artist, 'results': results}
def is_valid_entry(self, artist, entry):
duration = int(entry.media.duration.seconds)
title = smart_str(entry.media.title.text).lower()
if entry.rating is not None and float(entry.rating.average) < 3.5:
return False
if duration < (2 * 60) or duration > (9 * 60):
return False
if artist.lower() not in title:
return False
if re.search("\b(concert|cover)\b", title):
return False
return True
|
<commit_before>from gdata.youtube.service import YouTubeService, YouTubeVideoQuery
from django.utils.encoding import smart_str
import re
class YoutubeMusic:
def __init__(self):
self.service = YouTubeService()
def search(self, artist):
query = YouTubeVideoQuery()
query.vq = artist
query.orderby = 'relevance'
query.racy = 'exclude'
query.categories.append("/Music")
feed = self.service.YouTubeQuery(query)
results = []
for entry in feed.entry:
if not self.is_valid_entry(artist, entry):
continue
results.append({
'url': entry.media.player.url,
'title': smart_str(entry.media.title.text),
'duration': int(entry.media.duration.seconds),
})
return {'artist': artist, 'results': results}
def is_valid_entry(self, artist, entry):
duration = int(entry.media.duration.seconds)
title = smart_str(entry.media.title.text).lower()
if entry.rating is not None and float(entry.rating.average) < 3.5:
return False
if duration < (2 * 60) or duration > (9 * 60):
return False
if artist.lower() not in title:
return False
if re.search("\b(concert|cover)\b", title):
return False
return True
<commit_msg>Use format=5 in YT search to prevent "embedding disabled"<commit_after>
|
from gdata.youtube.service import YouTubeService, YouTubeVideoQuery
from django.utils.encoding import smart_str
import re
class YoutubeMusic:
def __init__(self):
self.service = YouTubeService()
def search(self, artist):
query = YouTubeVideoQuery()
query.vq = artist
query.orderby = 'relevance'
query.racy = 'exclude'
query.format = '5'
query.categories.append("/Music")
feed = self.service.YouTubeQuery(query)
results = []
for entry in feed.entry:
if not self.is_valid_entry(artist, entry):
continue
results.append({
'url': entry.media.player.url,
'title': smart_str(entry.media.title.text),
'duration': int(entry.media.duration.seconds),
})
return {'artist': artist, 'results': results}
def is_valid_entry(self, artist, entry):
duration = int(entry.media.duration.seconds)
title = smart_str(entry.media.title.text).lower()
if entry.rating is not None and float(entry.rating.average) < 3.5:
return False
if duration < (2 * 60) or duration > (9 * 60):
return False
if artist.lower() not in title:
return False
if re.search("\b(concert|cover)\b", title):
return False
return True
|
from gdata.youtube.service import YouTubeService, YouTubeVideoQuery
from django.utils.encoding import smart_str
import re
class YoutubeMusic:
def __init__(self):
self.service = YouTubeService()
def search(self, artist):
query = YouTubeVideoQuery()
query.vq = artist
query.orderby = 'relevance'
query.racy = 'exclude'
query.categories.append("/Music")
feed = self.service.YouTubeQuery(query)
results = []
for entry in feed.entry:
if not self.is_valid_entry(artist, entry):
continue
results.append({
'url': entry.media.player.url,
'title': smart_str(entry.media.title.text),
'duration': int(entry.media.duration.seconds),
})
return {'artist': artist, 'results': results}
def is_valid_entry(self, artist, entry):
duration = int(entry.media.duration.seconds)
title = smart_str(entry.media.title.text).lower()
if entry.rating is not None and float(entry.rating.average) < 3.5:
return False
if duration < (2 * 60) or duration > (9 * 60):
return False
if artist.lower() not in title:
return False
if re.search("\b(concert|cover)\b", title):
return False
return True
Use format=5 in YT search to prevent "embedding disabled"from gdata.youtube.service import YouTubeService, YouTubeVideoQuery
from django.utils.encoding import smart_str
import re
class YoutubeMusic:
def __init__(self):
self.service = YouTubeService()
def search(self, artist):
query = YouTubeVideoQuery()
query.vq = artist
query.orderby = 'relevance'
query.racy = 'exclude'
query.format = '5'
query.categories.append("/Music")
feed = self.service.YouTubeQuery(query)
results = []
for entry in feed.entry:
if not self.is_valid_entry(artist, entry):
continue
results.append({
'url': entry.media.player.url,
'title': smart_str(entry.media.title.text),
'duration': int(entry.media.duration.seconds),
})
return {'artist': artist, 'results': results}
def is_valid_entry(self, artist, entry):
duration = int(entry.media.duration.seconds)
title = smart_str(entry.media.title.text).lower()
if entry.rating is not None and float(entry.rating.average) < 3.5:
return False
if duration < (2 * 60) or duration > (9 * 60):
return False
if artist.lower() not in title:
return False
if re.search("\b(concert|cover)\b", title):
return False
return True
|
<commit_before>from gdata.youtube.service import YouTubeService, YouTubeVideoQuery
from django.utils.encoding import smart_str
import re
class YoutubeMusic:
def __init__(self):
self.service = YouTubeService()
def search(self, artist):
query = YouTubeVideoQuery()
query.vq = artist
query.orderby = 'relevance'
query.racy = 'exclude'
query.categories.append("/Music")
feed = self.service.YouTubeQuery(query)
results = []
for entry in feed.entry:
if not self.is_valid_entry(artist, entry):
continue
results.append({
'url': entry.media.player.url,
'title': smart_str(entry.media.title.text),
'duration': int(entry.media.duration.seconds),
})
return {'artist': artist, 'results': results}
def is_valid_entry(self, artist, entry):
duration = int(entry.media.duration.seconds)
title = smart_str(entry.media.title.text).lower()
if entry.rating is not None and float(entry.rating.average) < 3.5:
return False
if duration < (2 * 60) or duration > (9 * 60):
return False
if artist.lower() not in title:
return False
if re.search("\b(concert|cover)\b", title):
return False
return True
<commit_msg>Use format=5 in YT search to prevent "embedding disabled"<commit_after>from gdata.youtube.service import YouTubeService, YouTubeVideoQuery
from django.utils.encoding import smart_str
import re
class YoutubeMusic:
def __init__(self):
self.service = YouTubeService()
def search(self, artist):
query = YouTubeVideoQuery()
query.vq = artist
query.orderby = 'relevance'
query.racy = 'exclude'
query.format = '5'
query.categories.append("/Music")
feed = self.service.YouTubeQuery(query)
results = []
for entry in feed.entry:
if not self.is_valid_entry(artist, entry):
continue
results.append({
'url': entry.media.player.url,
'title': smart_str(entry.media.title.text),
'duration': int(entry.media.duration.seconds),
})
return {'artist': artist, 'results': results}
def is_valid_entry(self, artist, entry):
duration = int(entry.media.duration.seconds)
title = smart_str(entry.media.title.text).lower()
if entry.rating is not None and float(entry.rating.average) < 3.5:
return False
if duration < (2 * 60) or duration > (9 * 60):
return False
if artist.lower() not in title:
return False
if re.search("\b(concert|cover)\b", title):
return False
return True
|
1d6bb5e7ce706c8f54599f98744f3a5d62ce104e
|
src/config.py
|
src/config.py
|
import os
import ConfigParser as configparser
class Config(object):
def __init__(self):
self.config = configparser.RawConfigParser()
self.configfile = os.path.expanduser('~/.mmetering-clirc')
if not os.path.isfile(self.configfile):
# setup a new config file
self.init_file()
def init_file(self):
self.config.add_section('mmetering')
with open(self.configfile, 'a+') as configfile:
self.config.write(configfile)
def read(self):
self.config.read(self.configfile)
def write(self):
with open(self.configfile, 'wb') as configfile:
self.config.write(configfile)
def get_base_dir(self):
self.read()
try:
return self.config.get('mmetering', 'base_dir')
except configparser.NoOptionError:
print 'No base_dir specified in %s' % self.configfile
print 'Use mmetering-cli setup'
def set_base_dir(self, path):
base_dir = os.path.expanduser(path)
self.read()
self.config.set('mmetering', 'base_dir', base_dir)
self.write()
|
import os
import ConfigParser as configparser
class Config(object):
def __init__(self):
self.config = configparser.RawConfigParser()
self.configfile = os.path.expanduser('~/.mmetering-clirc')
if not os.path.isfile(self.configfile):
# setup a new config file
self.init_file()
def init_file(self):
self.config.add_section('mmetering')
with open(self.configfile, 'a+') as configfile:
self.config.write(configfile)
def read(self):
self.config.read(self.configfile)
def write(self):
with open(self.configfile, 'wb') as configfile:
self.config.write(configfile)
def set(self, section, name, value):
val = os.path.expanduser(value)
self.read()
self.config.set(section, name, val)
self.write()
def get(self, section, name):
self.read()
try:
return self.config.get(section, name)
except configparser.NoOptionError:
print 'No %s specified in %s' % (name, self.configfile)
print 'Try \'mmetering-cli setup\''
|
Replace get_base_dir and set_base_dir with more abstract methods get and set
|
Replace get_base_dir and set_base_dir with more abstract methods get and set
|
Python
|
mit
|
mmetering/mmetering-cli
|
import os
import ConfigParser as configparser
class Config(object):
def __init__(self):
self.config = configparser.RawConfigParser()
self.configfile = os.path.expanduser('~/.mmetering-clirc')
if not os.path.isfile(self.configfile):
# setup a new config file
self.init_file()
def init_file(self):
self.config.add_section('mmetering')
with open(self.configfile, 'a+') as configfile:
self.config.write(configfile)
def read(self):
self.config.read(self.configfile)
def write(self):
with open(self.configfile, 'wb') as configfile:
self.config.write(configfile)
def get_base_dir(self):
self.read()
try:
return self.config.get('mmetering', 'base_dir')
except configparser.NoOptionError:
print 'No base_dir specified in %s' % self.configfile
print 'Use mmetering-cli setup'
def set_base_dir(self, path):
base_dir = os.path.expanduser(path)
self.read()
self.config.set('mmetering', 'base_dir', base_dir)
self.write()
Replace get_base_dir and set_base_dir with more abstract methods get and set
|
import os
import ConfigParser as configparser
class Config(object):
def __init__(self):
self.config = configparser.RawConfigParser()
self.configfile = os.path.expanduser('~/.mmetering-clirc')
if not os.path.isfile(self.configfile):
# setup a new config file
self.init_file()
def init_file(self):
self.config.add_section('mmetering')
with open(self.configfile, 'a+') as configfile:
self.config.write(configfile)
def read(self):
self.config.read(self.configfile)
def write(self):
with open(self.configfile, 'wb') as configfile:
self.config.write(configfile)
def set(self, section, name, value):
val = os.path.expanduser(value)
self.read()
self.config.set(section, name, val)
self.write()
def get(self, section, name):
self.read()
try:
return self.config.get(section, name)
except configparser.NoOptionError:
print 'No %s specified in %s' % (name, self.configfile)
print 'Try \'mmetering-cli setup\''
|
<commit_before>import os
import ConfigParser as configparser
class Config(object):
def __init__(self):
self.config = configparser.RawConfigParser()
self.configfile = os.path.expanduser('~/.mmetering-clirc')
if not os.path.isfile(self.configfile):
# setup a new config file
self.init_file()
def init_file(self):
self.config.add_section('mmetering')
with open(self.configfile, 'a+') as configfile:
self.config.write(configfile)
def read(self):
self.config.read(self.configfile)
def write(self):
with open(self.configfile, 'wb') as configfile:
self.config.write(configfile)
def get_base_dir(self):
self.read()
try:
return self.config.get('mmetering', 'base_dir')
except configparser.NoOptionError:
print 'No base_dir specified in %s' % self.configfile
print 'Use mmetering-cli setup'
def set_base_dir(self, path):
base_dir = os.path.expanduser(path)
self.read()
self.config.set('mmetering', 'base_dir', base_dir)
self.write()
<commit_msg>Replace get_base_dir and set_base_dir with more abstract methods get and set<commit_after>
|
import os
import ConfigParser as configparser
class Config(object):
def __init__(self):
self.config = configparser.RawConfigParser()
self.configfile = os.path.expanduser('~/.mmetering-clirc')
if not os.path.isfile(self.configfile):
# setup a new config file
self.init_file()
def init_file(self):
self.config.add_section('mmetering')
with open(self.configfile, 'a+') as configfile:
self.config.write(configfile)
def read(self):
self.config.read(self.configfile)
def write(self):
with open(self.configfile, 'wb') as configfile:
self.config.write(configfile)
def set(self, section, name, value):
val = os.path.expanduser(value)
self.read()
self.config.set(section, name, val)
self.write()
def get(self, section, name):
self.read()
try:
return self.config.get(section, name)
except configparser.NoOptionError:
print 'No %s specified in %s' % (name, self.configfile)
print 'Try \'mmetering-cli setup\''
|
import os
import ConfigParser as configparser
class Config(object):
def __init__(self):
self.config = configparser.RawConfigParser()
self.configfile = os.path.expanduser('~/.mmetering-clirc')
if not os.path.isfile(self.configfile):
# setup a new config file
self.init_file()
def init_file(self):
self.config.add_section('mmetering')
with open(self.configfile, 'a+') as configfile:
self.config.write(configfile)
def read(self):
self.config.read(self.configfile)
def write(self):
with open(self.configfile, 'wb') as configfile:
self.config.write(configfile)
def get_base_dir(self):
self.read()
try:
return self.config.get('mmetering', 'base_dir')
except configparser.NoOptionError:
print 'No base_dir specified in %s' % self.configfile
print 'Use mmetering-cli setup'
def set_base_dir(self, path):
base_dir = os.path.expanduser(path)
self.read()
self.config.set('mmetering', 'base_dir', base_dir)
self.write()
Replace get_base_dir and set_base_dir with more abstract methods get and setimport os
import ConfigParser as configparser
class Config(object):
def __init__(self):
self.config = configparser.RawConfigParser()
self.configfile = os.path.expanduser('~/.mmetering-clirc')
if not os.path.isfile(self.configfile):
# setup a new config file
self.init_file()
def init_file(self):
self.config.add_section('mmetering')
with open(self.configfile, 'a+') as configfile:
self.config.write(configfile)
def read(self):
self.config.read(self.configfile)
def write(self):
with open(self.configfile, 'wb') as configfile:
self.config.write(configfile)
def set(self, section, name, value):
val = os.path.expanduser(value)
self.read()
self.config.set(section, name, val)
self.write()
def get(self, section, name):
self.read()
try:
return self.config.get(section, name)
except configparser.NoOptionError:
print 'No %s specified in %s' % (name, self.configfile)
print 'Try \'mmetering-cli setup\''
|
<commit_before>import os
import ConfigParser as configparser
class Config(object):
def __init__(self):
self.config = configparser.RawConfigParser()
self.configfile = os.path.expanduser('~/.mmetering-clirc')
if not os.path.isfile(self.configfile):
# setup a new config file
self.init_file()
def init_file(self):
self.config.add_section('mmetering')
with open(self.configfile, 'a+') as configfile:
self.config.write(configfile)
def read(self):
self.config.read(self.configfile)
def write(self):
with open(self.configfile, 'wb') as configfile:
self.config.write(configfile)
def get_base_dir(self):
self.read()
try:
return self.config.get('mmetering', 'base_dir')
except configparser.NoOptionError:
print 'No base_dir specified in %s' % self.configfile
print 'Use mmetering-cli setup'
def set_base_dir(self, path):
base_dir = os.path.expanduser(path)
self.read()
self.config.set('mmetering', 'base_dir', base_dir)
self.write()
<commit_msg>Replace get_base_dir and set_base_dir with more abstract methods get and set<commit_after>import os
import ConfigParser as configparser
class Config(object):
def __init__(self):
self.config = configparser.RawConfigParser()
self.configfile = os.path.expanduser('~/.mmetering-clirc')
if not os.path.isfile(self.configfile):
# setup a new config file
self.init_file()
def init_file(self):
self.config.add_section('mmetering')
with open(self.configfile, 'a+') as configfile:
self.config.write(configfile)
def read(self):
self.config.read(self.configfile)
def write(self):
with open(self.configfile, 'wb') as configfile:
self.config.write(configfile)
def set(self, section, name, value):
val = os.path.expanduser(value)
self.read()
self.config.set(section, name, val)
self.write()
def get(self, section, name):
self.read()
try:
return self.config.get(section, name)
except configparser.NoOptionError:
print 'No %s specified in %s' % (name, self.configfile)
print 'Try \'mmetering-cli setup\''
|
e966ddd804eee2f1b053de6f0bbf943d80dccc59
|
django_elastipymemcache/client.py
|
django_elastipymemcache/client.py
|
from pymemcache.client.hash import HashClient
class Client(HashClient):
def get_many(self, keys, gets=False, *args, **kwargs):
# pymemcache's HashClient may returns {'key': False}
end = super(Client, self).get_many(keys, gets, args, kwargs)
return {key: end[key] for key in end if end[key]}
get_multi = get_many
|
from pymemcache.client.hash import HashClient
class Client(HashClient):
def get_many(self, keys, gets=False, *args, **kwargs):
# pymemcache's HashClient may returns {'key': False}
end = super(Client, self).get_many(keys, gets, args, kwargs)
return {key: end.get(key) for key in end if end.get(key)}
get_multi = get_many
|
Fix get value more safe
|
Fix get value more safe
|
Python
|
mit
|
uncovertruth/django-elastipymemcache
|
from pymemcache.client.hash import HashClient
class Client(HashClient):
def get_many(self, keys, gets=False, *args, **kwargs):
# pymemcache's HashClient may returns {'key': False}
end = super(Client, self).get_many(keys, gets, args, kwargs)
return {key: end[key] for key in end if end[key]}
get_multi = get_many
Fix get value more safe
|
from pymemcache.client.hash import HashClient
class Client(HashClient):
def get_many(self, keys, gets=False, *args, **kwargs):
# pymemcache's HashClient may returns {'key': False}
end = super(Client, self).get_many(keys, gets, args, kwargs)
return {key: end.get(key) for key in end if end.get(key)}
get_multi = get_many
|
<commit_before>from pymemcache.client.hash import HashClient
class Client(HashClient):
def get_many(self, keys, gets=False, *args, **kwargs):
# pymemcache's HashClient may returns {'key': False}
end = super(Client, self).get_many(keys, gets, args, kwargs)
return {key: end[key] for key in end if end[key]}
get_multi = get_many
<commit_msg>Fix get value more safe<commit_after>
|
from pymemcache.client.hash import HashClient
class Client(HashClient):
def get_many(self, keys, gets=False, *args, **kwargs):
# pymemcache's HashClient may returns {'key': False}
end = super(Client, self).get_many(keys, gets, args, kwargs)
return {key: end.get(key) for key in end if end.get(key)}
get_multi = get_many
|
from pymemcache.client.hash import HashClient
class Client(HashClient):
def get_many(self, keys, gets=False, *args, **kwargs):
# pymemcache's HashClient may returns {'key': False}
end = super(Client, self).get_many(keys, gets, args, kwargs)
return {key: end[key] for key in end if end[key]}
get_multi = get_many
Fix get value more safefrom pymemcache.client.hash import HashClient
class Client(HashClient):
def get_many(self, keys, gets=False, *args, **kwargs):
# pymemcache's HashClient may returns {'key': False}
end = super(Client, self).get_many(keys, gets, args, kwargs)
return {key: end.get(key) for key in end if end.get(key)}
get_multi = get_many
|
<commit_before>from pymemcache.client.hash import HashClient
class Client(HashClient):
def get_many(self, keys, gets=False, *args, **kwargs):
# pymemcache's HashClient may returns {'key': False}
end = super(Client, self).get_many(keys, gets, args, kwargs)
return {key: end[key] for key in end if end[key]}
get_multi = get_many
<commit_msg>Fix get value more safe<commit_after>from pymemcache.client.hash import HashClient
class Client(HashClient):
def get_many(self, keys, gets=False, *args, **kwargs):
# pymemcache's HashClient may returns {'key': False}
end = super(Client, self).get_many(keys, gets, args, kwargs)
return {key: end.get(key) for key in end if end.get(key)}
get_multi = get_many
|
60daa277d5c3f1d9ab07ff5beccdaa323996068b
|
feincmstools/templatetags/feincmstools_tags.py
|
feincmstools/templatetags/feincmstools_tags.py
|
import os
from django import template
register = template.Library()
@register.filter
def is_parent_of(page1, page2):
"""
Determines whether a given page is the parent of another page
Example:
{% if page|is_parent_of:feincms_page %} ... {% endif %}
"""
if page1 is None:
return False
return (page1.tree_id == page2.tree_id and
page1.lft < page2.lft and
page1.rght > page2.rght)
@register.filter
def is_equal_or_parent_of(page1, page2):
return (page1.tree_id == page2.tree_id and
page1.lft <= page2.lft and
page1.rght >= page2.rght)
@register.filter
def is_sibling_of(page1, page2):
"""
Determines whether a given page is a sibling of another page
{% if page|is_sibling_of:feincms_page %} ... {% endif %}
"""
if page1 is None or page2 is None:
return False
return (page1.parent_id == page2.parent_id)
@register.filter
def get_extension(filename):
""" Return the extension from a file name """
return os.path.splitext(filename)[1][1:]
|
import os
from django import template
from feincms.templatetags.feincms_tags import feincms_render_content
register = template.Library()
@register.filter
def is_parent_of(page1, page2):
"""
Determines whether a given page is the parent of another page
Example:
{% if page|is_parent_of:feincms_page %} ... {% endif %}
"""
if page1 is None:
return False
return (page1.tree_id == page2.tree_id and
page1.lft < page2.lft and
page1.rght > page2.rght)
@register.filter
def is_equal_or_parent_of(page1, page2):
return (page1.tree_id == page2.tree_id and
page1.lft <= page2.lft and
page1.rght >= page2.rght)
@register.filter
def is_sibling_of(page1, page2):
"""
Determines whether a given page is a sibling of another page
{% if page|is_sibling_of:feincms_page %} ... {% endif %}
"""
if page1 is None or page2 is None:
return False
return (page1.parent_id == page2.parent_id)
@register.filter
def get_extension(filename):
""" Return the extension from a file name """
return os.path.splitext(filename)[1][1:]
@register.assignment_tag(takes_context=True)
def feincms_render_content_as(context, content, request=None):
return feincms_render_content(context, content, request)
|
Add assignment tag util for rendering chunks to tpl context
|
Add assignment tag util for rendering chunks to tpl context
|
Python
|
bsd-3-clause
|
ixc/glamkit-feincmstools,ixc/glamkit-feincmstools
|
import os
from django import template
register = template.Library()
@register.filter
def is_parent_of(page1, page2):
"""
Determines whether a given page is the parent of another page
Example:
{% if page|is_parent_of:feincms_page %} ... {% endif %}
"""
if page1 is None:
return False
return (page1.tree_id == page2.tree_id and
page1.lft < page2.lft and
page1.rght > page2.rght)
@register.filter
def is_equal_or_parent_of(page1, page2):
return (page1.tree_id == page2.tree_id and
page1.lft <= page2.lft and
page1.rght >= page2.rght)
@register.filter
def is_sibling_of(page1, page2):
"""
Determines whether a given page is a sibling of another page
{% if page|is_sibling_of:feincms_page %} ... {% endif %}
"""
if page1 is None or page2 is None:
return False
return (page1.parent_id == page2.parent_id)
@register.filter
def get_extension(filename):
""" Return the extension from a file name """
return os.path.splitext(filename)[1][1:]
Add assignment tag util for rendering chunks to tpl context
|
import os
from django import template
from feincms.templatetags.feincms_tags import feincms_render_content
register = template.Library()
@register.filter
def is_parent_of(page1, page2):
"""
Determines whether a given page is the parent of another page
Example:
{% if page|is_parent_of:feincms_page %} ... {% endif %}
"""
if page1 is None:
return False
return (page1.tree_id == page2.tree_id and
page1.lft < page2.lft and
page1.rght > page2.rght)
@register.filter
def is_equal_or_parent_of(page1, page2):
return (page1.tree_id == page2.tree_id and
page1.lft <= page2.lft and
page1.rght >= page2.rght)
@register.filter
def is_sibling_of(page1, page2):
"""
Determines whether a given page is a sibling of another page
{% if page|is_sibling_of:feincms_page %} ... {% endif %}
"""
if page1 is None or page2 is None:
return False
return (page1.parent_id == page2.parent_id)
@register.filter
def get_extension(filename):
""" Return the extension from a file name """
return os.path.splitext(filename)[1][1:]
@register.assignment_tag(takes_context=True)
def feincms_render_content_as(context, content, request=None):
return feincms_render_content(context, content, request)
|
<commit_before>import os
from django import template
register = template.Library()
@register.filter
def is_parent_of(page1, page2):
"""
Determines whether a given page is the parent of another page
Example:
{% if page|is_parent_of:feincms_page %} ... {% endif %}
"""
if page1 is None:
return False
return (page1.tree_id == page2.tree_id and
page1.lft < page2.lft and
page1.rght > page2.rght)
@register.filter
def is_equal_or_parent_of(page1, page2):
return (page1.tree_id == page2.tree_id and
page1.lft <= page2.lft and
page1.rght >= page2.rght)
@register.filter
def is_sibling_of(page1, page2):
"""
Determines whether a given page is a sibling of another page
{% if page|is_sibling_of:feincms_page %} ... {% endif %}
"""
if page1 is None or page2 is None:
return False
return (page1.parent_id == page2.parent_id)
@register.filter
def get_extension(filename):
""" Return the extension from a file name """
return os.path.splitext(filename)[1][1:]
<commit_msg>Add assignment tag util for rendering chunks to tpl context<commit_after>
|
import os
from django import template
from feincms.templatetags.feincms_tags import feincms_render_content
register = template.Library()
@register.filter
def is_parent_of(page1, page2):
"""
Determines whether a given page is the parent of another page
Example:
{% if page|is_parent_of:feincms_page %} ... {% endif %}
"""
if page1 is None:
return False
return (page1.tree_id == page2.tree_id and
page1.lft < page2.lft and
page1.rght > page2.rght)
@register.filter
def is_equal_or_parent_of(page1, page2):
return (page1.tree_id == page2.tree_id and
page1.lft <= page2.lft and
page1.rght >= page2.rght)
@register.filter
def is_sibling_of(page1, page2):
"""
Determines whether a given page is a sibling of another page
{% if page|is_sibling_of:feincms_page %} ... {% endif %}
"""
if page1 is None or page2 is None:
return False
return (page1.parent_id == page2.parent_id)
@register.filter
def get_extension(filename):
""" Return the extension from a file name """
return os.path.splitext(filename)[1][1:]
@register.assignment_tag(takes_context=True)
def feincms_render_content_as(context, content, request=None):
return feincms_render_content(context, content, request)
|
import os
from django import template
register = template.Library()
@register.filter
def is_parent_of(page1, page2):
"""
Determines whether a given page is the parent of another page
Example:
{% if page|is_parent_of:feincms_page %} ... {% endif %}
"""
if page1 is None:
return False
return (page1.tree_id == page2.tree_id and
page1.lft < page2.lft and
page1.rght > page2.rght)
@register.filter
def is_equal_or_parent_of(page1, page2):
return (page1.tree_id == page2.tree_id and
page1.lft <= page2.lft and
page1.rght >= page2.rght)
@register.filter
def is_sibling_of(page1, page2):
"""
Determines whether a given page is a sibling of another page
{% if page|is_sibling_of:feincms_page %} ... {% endif %}
"""
if page1 is None or page2 is None:
return False
return (page1.parent_id == page2.parent_id)
@register.filter
def get_extension(filename):
""" Return the extension from a file name """
return os.path.splitext(filename)[1][1:]
Add assignment tag util for rendering chunks to tpl contextimport os
from django import template
from feincms.templatetags.feincms_tags import feincms_render_content
register = template.Library()
@register.filter
def is_parent_of(page1, page2):
"""
Determines whether a given page is the parent of another page
Example:
{% if page|is_parent_of:feincms_page %} ... {% endif %}
"""
if page1 is None:
return False
return (page1.tree_id == page2.tree_id and
page1.lft < page2.lft and
page1.rght > page2.rght)
@register.filter
def is_equal_or_parent_of(page1, page2):
return (page1.tree_id == page2.tree_id and
page1.lft <= page2.lft and
page1.rght >= page2.rght)
@register.filter
def is_sibling_of(page1, page2):
"""
Determines whether a given page is a sibling of another page
{% if page|is_sibling_of:feincms_page %} ... {% endif %}
"""
if page1 is None or page2 is None:
return False
return (page1.parent_id == page2.parent_id)
@register.filter
def get_extension(filename):
""" Return the extension from a file name """
return os.path.splitext(filename)[1][1:]
@register.assignment_tag(takes_context=True)
def feincms_render_content_as(context, content, request=None):
return feincms_render_content(context, content, request)
|
<commit_before>import os
from django import template
register = template.Library()
@register.filter
def is_parent_of(page1, page2):
"""
Determines whether a given page is the parent of another page
Example:
{% if page|is_parent_of:feincms_page %} ... {% endif %}
"""
if page1 is None:
return False
return (page1.tree_id == page2.tree_id and
page1.lft < page2.lft and
page1.rght > page2.rght)
@register.filter
def is_equal_or_parent_of(page1, page2):
return (page1.tree_id == page2.tree_id and
page1.lft <= page2.lft and
page1.rght >= page2.rght)
@register.filter
def is_sibling_of(page1, page2):
"""
Determines whether a given page is a sibling of another page
{% if page|is_sibling_of:feincms_page %} ... {% endif %}
"""
if page1 is None or page2 is None:
return False
return (page1.parent_id == page2.parent_id)
@register.filter
def get_extension(filename):
""" Return the extension from a file name """
return os.path.splitext(filename)[1][1:]
<commit_msg>Add assignment tag util for rendering chunks to tpl context<commit_after>import os
from django import template
from feincms.templatetags.feincms_tags import feincms_render_content
register = template.Library()
@register.filter
def is_parent_of(page1, page2):
"""
Determines whether a given page is the parent of another page
Example:
{% if page|is_parent_of:feincms_page %} ... {% endif %}
"""
if page1 is None:
return False
return (page1.tree_id == page2.tree_id and
page1.lft < page2.lft and
page1.rght > page2.rght)
@register.filter
def is_equal_or_parent_of(page1, page2):
return (page1.tree_id == page2.tree_id and
page1.lft <= page2.lft and
page1.rght >= page2.rght)
@register.filter
def is_sibling_of(page1, page2):
"""
Determines whether a given page is a sibling of another page
{% if page|is_sibling_of:feincms_page %} ... {% endif %}
"""
if page1 is None or page2 is None:
return False
return (page1.parent_id == page2.parent_id)
@register.filter
def get_extension(filename):
""" Return the extension from a file name """
return os.path.splitext(filename)[1][1:]
@register.assignment_tag(takes_context=True)
def feincms_render_content_as(context, content, request=None):
return feincms_render_content(context, content, request)
|
1e5102d8bafb3b4d2cb07822129397aa56f30bbe
|
devicecloud/examples/example_helpers.py
|
devicecloud/examples/example_helpers.py
|
# This Source Code Form is subject to the terms of the Mozilla Public
# License, v. 2.0. If a copy of the MPL was not distributed with this
# file, You can obtain one at http://mozilla.org/MPL/2.0/.
#
# Copyright (c) 2015 Digi International, Inc.
from getpass import getpass
import os
from devicecloud import DeviceCloud
def get_authenticated_dc():
while True:
base_url = os.environ.get('DC_BASE_URL', 'https://login.etherios.com')
username = os.environ.get('DC_USERNAME', None)
if not username:
username = input("username: ")
password = os.environ.get('DC_PASSWORD', None)
if not password:
password = getpass("password: ")
dc = DeviceCloud(username, password, base_url=base_url)
if dc.has_valid_credentials():
print("Credentials accepted!")
return dc
else:
print("Invalid username or password provided, try again")
|
# This Source Code Form is subject to the terms of the Mozilla Public
# License, v. 2.0. If a copy of the MPL was not distributed with this
# file, You can obtain one at http://mozilla.org/MPL/2.0/.
#
# Copyright (c) 2015 Digi International, Inc.
from getpass import getpass
import os
from six.moves import input
from devicecloud import DeviceCloud
def get_authenticated_dc():
while True:
base_url = os.environ.get('DC_BASE_URL', 'https://login.etherios.com')
username = os.environ.get('DC_USERNAME', None)
if not username:
username = input("username: ")
password = os.environ.get('DC_PASSWORD', None)
if not password:
password = getpass("password: ")
dc = DeviceCloud(username, password, base_url=base_url)
if dc.has_valid_credentials():
print("Credentials accepted!")
return dc
else:
print("Invalid username or password provided, try again")
|
Handle using the input function in python 2 for getting username for examples
|
Handle using the input function in python 2 for getting username for examples
Previously this used the builtin input function to get the username. In
python 3 this is fine, but if python 2 this is equivalent to
eval(raw_input(prompt))
and thus tried to evaluate the username as a variable and typically failed.
|
Python
|
mpl-2.0
|
michaelcho/python-devicecloud,michaelcho/python-devicecloud,digidotcom/python-devicecloud,brucetsao/python-devicecloud,ctrlaltdel/python-devicecloud,digidotcom/python-devicecloud,brucetsao/python-devicecloud,ctrlaltdel/python-devicecloud
|
# This Source Code Form is subject to the terms of the Mozilla Public
# License, v. 2.0. If a copy of the MPL was not distributed with this
# file, You can obtain one at http://mozilla.org/MPL/2.0/.
#
# Copyright (c) 2015 Digi International, Inc.
from getpass import getpass
import os
from devicecloud import DeviceCloud
def get_authenticated_dc():
while True:
base_url = os.environ.get('DC_BASE_URL', 'https://login.etherios.com')
username = os.environ.get('DC_USERNAME', None)
if not username:
username = input("username: ")
password = os.environ.get('DC_PASSWORD', None)
if not password:
password = getpass("password: ")
dc = DeviceCloud(username, password, base_url=base_url)
if dc.has_valid_credentials():
print("Credentials accepted!")
return dc
else:
print("Invalid username or password provided, try again")
Handle using the input function in python 2 for getting username for examples
Previously this used the builtin input function to get the username. In
python 3 this is fine, but if python 2 this is equivalent to
eval(raw_input(prompt))
and thus tried to evaluate the username as a variable and typically failed.
|
# This Source Code Form is subject to the terms of the Mozilla Public
# License, v. 2.0. If a copy of the MPL was not distributed with this
# file, You can obtain one at http://mozilla.org/MPL/2.0/.
#
# Copyright (c) 2015 Digi International, Inc.
from getpass import getpass
import os
from six.moves import input
from devicecloud import DeviceCloud
def get_authenticated_dc():
while True:
base_url = os.environ.get('DC_BASE_URL', 'https://login.etherios.com')
username = os.environ.get('DC_USERNAME', None)
if not username:
username = input("username: ")
password = os.environ.get('DC_PASSWORD', None)
if not password:
password = getpass("password: ")
dc = DeviceCloud(username, password, base_url=base_url)
if dc.has_valid_credentials():
print("Credentials accepted!")
return dc
else:
print("Invalid username or password provided, try again")
|
<commit_before># This Source Code Form is subject to the terms of the Mozilla Public
# License, v. 2.0. If a copy of the MPL was not distributed with this
# file, You can obtain one at http://mozilla.org/MPL/2.0/.
#
# Copyright (c) 2015 Digi International, Inc.
from getpass import getpass
import os
from devicecloud import DeviceCloud
def get_authenticated_dc():
while True:
base_url = os.environ.get('DC_BASE_URL', 'https://login.etherios.com')
username = os.environ.get('DC_USERNAME', None)
if not username:
username = input("username: ")
password = os.environ.get('DC_PASSWORD', None)
if not password:
password = getpass("password: ")
dc = DeviceCloud(username, password, base_url=base_url)
if dc.has_valid_credentials():
print("Credentials accepted!")
return dc
else:
print("Invalid username or password provided, try again")
<commit_msg>Handle using the input function in python 2 for getting username for examples
Previously this used the builtin input function to get the username. In
python 3 this is fine, but if python 2 this is equivalent to
eval(raw_input(prompt))
and thus tried to evaluate the username as a variable and typically failed.<commit_after>
|
# This Source Code Form is subject to the terms of the Mozilla Public
# License, v. 2.0. If a copy of the MPL was not distributed with this
# file, You can obtain one at http://mozilla.org/MPL/2.0/.
#
# Copyright (c) 2015 Digi International, Inc.
from getpass import getpass
import os
from six.moves import input
from devicecloud import DeviceCloud
def get_authenticated_dc():
while True:
base_url = os.environ.get('DC_BASE_URL', 'https://login.etherios.com')
username = os.environ.get('DC_USERNAME', None)
if not username:
username = input("username: ")
password = os.environ.get('DC_PASSWORD', None)
if not password:
password = getpass("password: ")
dc = DeviceCloud(username, password, base_url=base_url)
if dc.has_valid_credentials():
print("Credentials accepted!")
return dc
else:
print("Invalid username or password provided, try again")
|
# This Source Code Form is subject to the terms of the Mozilla Public
# License, v. 2.0. If a copy of the MPL was not distributed with this
# file, You can obtain one at http://mozilla.org/MPL/2.0/.
#
# Copyright (c) 2015 Digi International, Inc.
from getpass import getpass
import os
from devicecloud import DeviceCloud
def get_authenticated_dc():
while True:
base_url = os.environ.get('DC_BASE_URL', 'https://login.etherios.com')
username = os.environ.get('DC_USERNAME', None)
if not username:
username = input("username: ")
password = os.environ.get('DC_PASSWORD', None)
if not password:
password = getpass("password: ")
dc = DeviceCloud(username, password, base_url=base_url)
if dc.has_valid_credentials():
print("Credentials accepted!")
return dc
else:
print("Invalid username or password provided, try again")
Handle using the input function in python 2 for getting username for examples
Previously this used the builtin input function to get the username. In
python 3 this is fine, but if python 2 this is equivalent to
eval(raw_input(prompt))
and thus tried to evaluate the username as a variable and typically failed.# This Source Code Form is subject to the terms of the Mozilla Public
# License, v. 2.0. If a copy of the MPL was not distributed with this
# file, You can obtain one at http://mozilla.org/MPL/2.0/.
#
# Copyright (c) 2015 Digi International, Inc.
from getpass import getpass
import os
from six.moves import input
from devicecloud import DeviceCloud
def get_authenticated_dc():
while True:
base_url = os.environ.get('DC_BASE_URL', 'https://login.etherios.com')
username = os.environ.get('DC_USERNAME', None)
if not username:
username = input("username: ")
password = os.environ.get('DC_PASSWORD', None)
if not password:
password = getpass("password: ")
dc = DeviceCloud(username, password, base_url=base_url)
if dc.has_valid_credentials():
print("Credentials accepted!")
return dc
else:
print("Invalid username or password provided, try again")
|
<commit_before># This Source Code Form is subject to the terms of the Mozilla Public
# License, v. 2.0. If a copy of the MPL was not distributed with this
# file, You can obtain one at http://mozilla.org/MPL/2.0/.
#
# Copyright (c) 2015 Digi International, Inc.
from getpass import getpass
import os
from devicecloud import DeviceCloud
def get_authenticated_dc():
while True:
base_url = os.environ.get('DC_BASE_URL', 'https://login.etherios.com')
username = os.environ.get('DC_USERNAME', None)
if not username:
username = input("username: ")
password = os.environ.get('DC_PASSWORD', None)
if not password:
password = getpass("password: ")
dc = DeviceCloud(username, password, base_url=base_url)
if dc.has_valid_credentials():
print("Credentials accepted!")
return dc
else:
print("Invalid username or password provided, try again")
<commit_msg>Handle using the input function in python 2 for getting username for examples
Previously this used the builtin input function to get the username. In
python 3 this is fine, but if python 2 this is equivalent to
eval(raw_input(prompt))
and thus tried to evaluate the username as a variable and typically failed.<commit_after># This Source Code Form is subject to the terms of the Mozilla Public
# License, v. 2.0. If a copy of the MPL was not distributed with this
# file, You can obtain one at http://mozilla.org/MPL/2.0/.
#
# Copyright (c) 2015 Digi International, Inc.
from getpass import getpass
import os
from six.moves import input
from devicecloud import DeviceCloud
def get_authenticated_dc():
while True:
base_url = os.environ.get('DC_BASE_URL', 'https://login.etherios.com')
username = os.environ.get('DC_USERNAME', None)
if not username:
username = input("username: ")
password = os.environ.get('DC_PASSWORD', None)
if not password:
password = getpass("password: ")
dc = DeviceCloud(username, password, base_url=base_url)
if dc.has_valid_credentials():
print("Credentials accepted!")
return dc
else:
print("Invalid username or password provided, try again")
|
4a41b33286cf881f0b3aa09c29a4aaa3568b5259
|
website/stats/plots/mimp.py
|
website/stats/plots/mimp.py
|
from analyses.mimp import glycosylation_sub_types, run_mimp
from helpers.plots import stacked_bar_plot
from ..store import counter
@counter
@stacked_bar_plot
def gains_and_losses_for_glycosylation_subtypes():
results = {}
effects = 'loss', 'gain'
for source_name in ['mc3', 'clinvar']:
for site_type_name in glycosylation_sub_types:
result = run_mimp(source_name, site_type_name, enzyme_type='catch-all')
if result.empty:
continue
effect_counts = result.effect.value_counts()
results[source_name] = effects, [effect_counts.get(effect, 0) for effect in effects]
return results
|
from analyses.mimp import glycosylation_sub_types, run_mimp
from helpers.plots import stacked_bar_plot
from ..store import counter
@counter
@stacked_bar_plot
def gains_and_losses_for_glycosylation_subtypes():
results = {}
effects = 'loss', 'gain'
for source_name in ['mc3', 'clinvar']:
for site_type_name in glycosylation_sub_types:
result = run_mimp(source_name, site_type_name, enzyme_type='catch-all')
if result.empty:
continue
effect_counts = result.effect.value_counts()
results[source_name] = effects, [
int(effect_counts.get(effect, 0))
for effect in effects
]
return results
|
Convert numpy int to native int for JSON serialization
|
Convert numpy int to native int for JSON serialization
|
Python
|
lgpl-2.1
|
reimandlab/Visualisation-Framework-for-Genome-Mutations,reimandlab/ActiveDriverDB,reimandlab/ActiveDriverDB,reimandlab/ActiveDriverDB,reimandlab/Visualisation-Framework-for-Genome-Mutations,reimandlab/Visualisation-Framework-for-Genome-Mutations,reimandlab/Visualistion-Framework-for-Genome-Mutations,reimandlab/Visualistion-Framework-for-Genome-Mutations,reimandlab/Visualistion-Framework-for-Genome-Mutations,reimandlab/Visualisation-Framework-for-Genome-Mutations,reimandlab/Visualisation-Framework-for-Genome-Mutations,reimandlab/ActiveDriverDB,reimandlab/Visualistion-Framework-for-Genome-Mutations
|
from analyses.mimp import glycosylation_sub_types, run_mimp
from helpers.plots import stacked_bar_plot
from ..store import counter
@counter
@stacked_bar_plot
def gains_and_losses_for_glycosylation_subtypes():
results = {}
effects = 'loss', 'gain'
for source_name in ['mc3', 'clinvar']:
for site_type_name in glycosylation_sub_types:
result = run_mimp(source_name, site_type_name, enzyme_type='catch-all')
if result.empty:
continue
effect_counts = result.effect.value_counts()
results[source_name] = effects, [effect_counts.get(effect, 0) for effect in effects]
return results
Convert numpy int to native int for JSON serialization
|
from analyses.mimp import glycosylation_sub_types, run_mimp
from helpers.plots import stacked_bar_plot
from ..store import counter
@counter
@stacked_bar_plot
def gains_and_losses_for_glycosylation_subtypes():
results = {}
effects = 'loss', 'gain'
for source_name in ['mc3', 'clinvar']:
for site_type_name in glycosylation_sub_types:
result = run_mimp(source_name, site_type_name, enzyme_type='catch-all')
if result.empty:
continue
effect_counts = result.effect.value_counts()
results[source_name] = effects, [
int(effect_counts.get(effect, 0))
for effect in effects
]
return results
|
<commit_before>from analyses.mimp import glycosylation_sub_types, run_mimp
from helpers.plots import stacked_bar_plot
from ..store import counter
@counter
@stacked_bar_plot
def gains_and_losses_for_glycosylation_subtypes():
results = {}
effects = 'loss', 'gain'
for source_name in ['mc3', 'clinvar']:
for site_type_name in glycosylation_sub_types:
result = run_mimp(source_name, site_type_name, enzyme_type='catch-all')
if result.empty:
continue
effect_counts = result.effect.value_counts()
results[source_name] = effects, [effect_counts.get(effect, 0) for effect in effects]
return results
<commit_msg>Convert numpy int to native int for JSON serialization<commit_after>
|
from analyses.mimp import glycosylation_sub_types, run_mimp
from helpers.plots import stacked_bar_plot
from ..store import counter
@counter
@stacked_bar_plot
def gains_and_losses_for_glycosylation_subtypes():
results = {}
effects = 'loss', 'gain'
for source_name in ['mc3', 'clinvar']:
for site_type_name in glycosylation_sub_types:
result = run_mimp(source_name, site_type_name, enzyme_type='catch-all')
if result.empty:
continue
effect_counts = result.effect.value_counts()
results[source_name] = effects, [
int(effect_counts.get(effect, 0))
for effect in effects
]
return results
|
from analyses.mimp import glycosylation_sub_types, run_mimp
from helpers.plots import stacked_bar_plot
from ..store import counter
@counter
@stacked_bar_plot
def gains_and_losses_for_glycosylation_subtypes():
results = {}
effects = 'loss', 'gain'
for source_name in ['mc3', 'clinvar']:
for site_type_name in glycosylation_sub_types:
result = run_mimp(source_name, site_type_name, enzyme_type='catch-all')
if result.empty:
continue
effect_counts = result.effect.value_counts()
results[source_name] = effects, [effect_counts.get(effect, 0) for effect in effects]
return results
Convert numpy int to native int for JSON serializationfrom analyses.mimp import glycosylation_sub_types, run_mimp
from helpers.plots import stacked_bar_plot
from ..store import counter
@counter
@stacked_bar_plot
def gains_and_losses_for_glycosylation_subtypes():
results = {}
effects = 'loss', 'gain'
for source_name in ['mc3', 'clinvar']:
for site_type_name in glycosylation_sub_types:
result = run_mimp(source_name, site_type_name, enzyme_type='catch-all')
if result.empty:
continue
effect_counts = result.effect.value_counts()
results[source_name] = effects, [
int(effect_counts.get(effect, 0))
for effect in effects
]
return results
|
<commit_before>from analyses.mimp import glycosylation_sub_types, run_mimp
from helpers.plots import stacked_bar_plot
from ..store import counter
@counter
@stacked_bar_plot
def gains_and_losses_for_glycosylation_subtypes():
results = {}
effects = 'loss', 'gain'
for source_name in ['mc3', 'clinvar']:
for site_type_name in glycosylation_sub_types:
result = run_mimp(source_name, site_type_name, enzyme_type='catch-all')
if result.empty:
continue
effect_counts = result.effect.value_counts()
results[source_name] = effects, [effect_counts.get(effect, 0) for effect in effects]
return results
<commit_msg>Convert numpy int to native int for JSON serialization<commit_after>from analyses.mimp import glycosylation_sub_types, run_mimp
from helpers.plots import stacked_bar_plot
from ..store import counter
@counter
@stacked_bar_plot
def gains_and_losses_for_glycosylation_subtypes():
results = {}
effects = 'loss', 'gain'
for source_name in ['mc3', 'clinvar']:
for site_type_name in glycosylation_sub_types:
result = run_mimp(source_name, site_type_name, enzyme_type='catch-all')
if result.empty:
continue
effect_counts = result.effect.value_counts()
results[source_name] = effects, [
int(effect_counts.get(effect, 0))
for effect in effects
]
return results
|
d7232d855d406a26b2485b5c1fcd587e90fddf39
|
tests/test_aio.py
|
tests/test_aio.py
|
import pytest
from ratelimiter import RateLimiter
@pytest.mark.asyncio
async def test_alock():
rl = RateLimiter(max_calls=10, period=0.01)
assert rl._alock is None
async with rl:
pass
alock = rl._alock
assert alock
async with rl:
pass
assert rl._alock is alock
|
import pytest
from ratelimiter import RateLimiter
@pytest.mark.asyncio
async def test_alock(event_loop):
rl = RateLimiter(max_calls=10, period=0.01)
assert rl._alock is None
async with rl:
pass
alock = rl._alock
assert alock
async with rl:
pass
assert rl._alock is alock
|
Fix Runtime warnings on async tests
|
Fix Runtime warnings on async tests
|
Python
|
apache-2.0
|
RazerM/ratelimiter
|
import pytest
from ratelimiter import RateLimiter
@pytest.mark.asyncio
async def test_alock():
rl = RateLimiter(max_calls=10, period=0.01)
assert rl._alock is None
async with rl:
pass
alock = rl._alock
assert alock
async with rl:
pass
assert rl._alock is alock
Fix Runtime warnings on async tests
|
import pytest
from ratelimiter import RateLimiter
@pytest.mark.asyncio
async def test_alock(event_loop):
rl = RateLimiter(max_calls=10, period=0.01)
assert rl._alock is None
async with rl:
pass
alock = rl._alock
assert alock
async with rl:
pass
assert rl._alock is alock
|
<commit_before>import pytest
from ratelimiter import RateLimiter
@pytest.mark.asyncio
async def test_alock():
rl = RateLimiter(max_calls=10, period=0.01)
assert rl._alock is None
async with rl:
pass
alock = rl._alock
assert alock
async with rl:
pass
assert rl._alock is alock
<commit_msg>Fix Runtime warnings on async tests<commit_after>
|
import pytest
from ratelimiter import RateLimiter
@pytest.mark.asyncio
async def test_alock(event_loop):
rl = RateLimiter(max_calls=10, period=0.01)
assert rl._alock is None
async with rl:
pass
alock = rl._alock
assert alock
async with rl:
pass
assert rl._alock is alock
|
import pytest
from ratelimiter import RateLimiter
@pytest.mark.asyncio
async def test_alock():
rl = RateLimiter(max_calls=10, period=0.01)
assert rl._alock is None
async with rl:
pass
alock = rl._alock
assert alock
async with rl:
pass
assert rl._alock is alock
Fix Runtime warnings on async testsimport pytest
from ratelimiter import RateLimiter
@pytest.mark.asyncio
async def test_alock(event_loop):
rl = RateLimiter(max_calls=10, period=0.01)
assert rl._alock is None
async with rl:
pass
alock = rl._alock
assert alock
async with rl:
pass
assert rl._alock is alock
|
<commit_before>import pytest
from ratelimiter import RateLimiter
@pytest.mark.asyncio
async def test_alock():
rl = RateLimiter(max_calls=10, period=0.01)
assert rl._alock is None
async with rl:
pass
alock = rl._alock
assert alock
async with rl:
pass
assert rl._alock is alock
<commit_msg>Fix Runtime warnings on async tests<commit_after>import pytest
from ratelimiter import RateLimiter
@pytest.mark.asyncio
async def test_alock(event_loop):
rl = RateLimiter(max_calls=10, period=0.01)
assert rl._alock is None
async with rl:
pass
alock = rl._alock
assert alock
async with rl:
pass
assert rl._alock is alock
|
d47d56525f85c5fa8b1f6b817a85479b9eb07582
|
sqlalchemy_utils/functions/__init__.py
|
sqlalchemy_utils/functions/__init__.py
|
from .defer_except import defer_except
from .mock import create_mock_engine, mock_engine
from .render import render_expression, render_statement
from .sort_query import sort_query, QuerySorterException
from .database import (
database_exists,
create_database,
drop_database,
escape_like,
is_auto_assigned_date_column,
is_indexed_foreign_key,
non_indexed_foreign_keys,
)
from .orm import (
declarative_base,
getdotattr,
has_changes,
identity,
naturally_equivalent,
primary_keys,
table_name,
)
__all__ = (
create_database,
create_mock_engine,
database_exists,
declarative_base,
defer_except,
drop_database,
escape_like,
getdotattr,
has_changes,
identity,
is_auto_assigned_date_column,
is_indexed_foreign_key,
mock_engine,
naturally_equivalent,
non_indexed_foreign_keys,
primary_keys,
QuerySorterException,
render_expression,
render_statement,
sort_query,
table_name,
)
|
from .defer_except import defer_except
from .mock import create_mock_engine, mock_engine
from .render import render_expression, render_statement
from .sort_query import sort_query, QuerySorterException
from .database import (
database_exists,
create_database,
drop_database,
escape_like,
is_auto_assigned_date_column,
is_indexed_foreign_key,
non_indexed_foreign_keys,
)
from .orm import (
declarative_base,
getdotattr,
has_changes,
identity,
naturally_equivalent,
query_entities,
primary_keys,
table_name,
)
__all__ = (
create_database,
create_mock_engine,
database_exists,
declarative_base,
defer_except,
drop_database,
escape_like,
getdotattr,
has_changes,
identity,
is_auto_assigned_date_column,
is_indexed_foreign_key,
mock_engine,
naturally_equivalent,
non_indexed_foreign_keys,
primary_keys,
QuerySorterException,
render_expression,
render_statement,
sort_query,
table_name,
)
|
Add query_entities to functions module import
|
Add query_entities to functions module import
|
Python
|
bsd-3-clause
|
joshfriend/sqlalchemy-utils,joshfriend/sqlalchemy-utils,cheungpat/sqlalchemy-utils,marrybird/sqlalchemy-utils,rmoorman/sqlalchemy-utils,spoqa/sqlalchemy-utils,tonyseek/sqlalchemy-utils,tonyseek/sqlalchemy-utils,JackWink/sqlalchemy-utils,konstantinoskostis/sqlalchemy-utils
|
from .defer_except import defer_except
from .mock import create_mock_engine, mock_engine
from .render import render_expression, render_statement
from .sort_query import sort_query, QuerySorterException
from .database import (
database_exists,
create_database,
drop_database,
escape_like,
is_auto_assigned_date_column,
is_indexed_foreign_key,
non_indexed_foreign_keys,
)
from .orm import (
declarative_base,
getdotattr,
has_changes,
identity,
naturally_equivalent,
primary_keys,
table_name,
)
__all__ = (
create_database,
create_mock_engine,
database_exists,
declarative_base,
defer_except,
drop_database,
escape_like,
getdotattr,
has_changes,
identity,
is_auto_assigned_date_column,
is_indexed_foreign_key,
mock_engine,
naturally_equivalent,
non_indexed_foreign_keys,
primary_keys,
QuerySorterException,
render_expression,
render_statement,
sort_query,
table_name,
)
Add query_entities to functions module import
|
from .defer_except import defer_except
from .mock import create_mock_engine, mock_engine
from .render import render_expression, render_statement
from .sort_query import sort_query, QuerySorterException
from .database import (
database_exists,
create_database,
drop_database,
escape_like,
is_auto_assigned_date_column,
is_indexed_foreign_key,
non_indexed_foreign_keys,
)
from .orm import (
declarative_base,
getdotattr,
has_changes,
identity,
naturally_equivalent,
query_entities,
primary_keys,
table_name,
)
__all__ = (
create_database,
create_mock_engine,
database_exists,
declarative_base,
defer_except,
drop_database,
escape_like,
getdotattr,
has_changes,
identity,
is_auto_assigned_date_column,
is_indexed_foreign_key,
mock_engine,
naturally_equivalent,
non_indexed_foreign_keys,
primary_keys,
QuerySorterException,
render_expression,
render_statement,
sort_query,
table_name,
)
|
<commit_before>from .defer_except import defer_except
from .mock import create_mock_engine, mock_engine
from .render import render_expression, render_statement
from .sort_query import sort_query, QuerySorterException
from .database import (
database_exists,
create_database,
drop_database,
escape_like,
is_auto_assigned_date_column,
is_indexed_foreign_key,
non_indexed_foreign_keys,
)
from .orm import (
declarative_base,
getdotattr,
has_changes,
identity,
naturally_equivalent,
primary_keys,
table_name,
)
__all__ = (
create_database,
create_mock_engine,
database_exists,
declarative_base,
defer_except,
drop_database,
escape_like,
getdotattr,
has_changes,
identity,
is_auto_assigned_date_column,
is_indexed_foreign_key,
mock_engine,
naturally_equivalent,
non_indexed_foreign_keys,
primary_keys,
QuerySorterException,
render_expression,
render_statement,
sort_query,
table_name,
)
<commit_msg>Add query_entities to functions module import<commit_after>
|
from .defer_except import defer_except
from .mock import create_mock_engine, mock_engine
from .render import render_expression, render_statement
from .sort_query import sort_query, QuerySorterException
from .database import (
database_exists,
create_database,
drop_database,
escape_like,
is_auto_assigned_date_column,
is_indexed_foreign_key,
non_indexed_foreign_keys,
)
from .orm import (
declarative_base,
getdotattr,
has_changes,
identity,
naturally_equivalent,
query_entities,
primary_keys,
table_name,
)
__all__ = (
create_database,
create_mock_engine,
database_exists,
declarative_base,
defer_except,
drop_database,
escape_like,
getdotattr,
has_changes,
identity,
is_auto_assigned_date_column,
is_indexed_foreign_key,
mock_engine,
naturally_equivalent,
non_indexed_foreign_keys,
primary_keys,
QuerySorterException,
render_expression,
render_statement,
sort_query,
table_name,
)
|
from .defer_except import defer_except
from .mock import create_mock_engine, mock_engine
from .render import render_expression, render_statement
from .sort_query import sort_query, QuerySorterException
from .database import (
database_exists,
create_database,
drop_database,
escape_like,
is_auto_assigned_date_column,
is_indexed_foreign_key,
non_indexed_foreign_keys,
)
from .orm import (
declarative_base,
getdotattr,
has_changes,
identity,
naturally_equivalent,
primary_keys,
table_name,
)
__all__ = (
create_database,
create_mock_engine,
database_exists,
declarative_base,
defer_except,
drop_database,
escape_like,
getdotattr,
has_changes,
identity,
is_auto_assigned_date_column,
is_indexed_foreign_key,
mock_engine,
naturally_equivalent,
non_indexed_foreign_keys,
primary_keys,
QuerySorterException,
render_expression,
render_statement,
sort_query,
table_name,
)
Add query_entities to functions module importfrom .defer_except import defer_except
from .mock import create_mock_engine, mock_engine
from .render import render_expression, render_statement
from .sort_query import sort_query, QuerySorterException
from .database import (
database_exists,
create_database,
drop_database,
escape_like,
is_auto_assigned_date_column,
is_indexed_foreign_key,
non_indexed_foreign_keys,
)
from .orm import (
declarative_base,
getdotattr,
has_changes,
identity,
naturally_equivalent,
query_entities,
primary_keys,
table_name,
)
__all__ = (
create_database,
create_mock_engine,
database_exists,
declarative_base,
defer_except,
drop_database,
escape_like,
getdotattr,
has_changes,
identity,
is_auto_assigned_date_column,
is_indexed_foreign_key,
mock_engine,
naturally_equivalent,
non_indexed_foreign_keys,
primary_keys,
QuerySorterException,
render_expression,
render_statement,
sort_query,
table_name,
)
|
<commit_before>from .defer_except import defer_except
from .mock import create_mock_engine, mock_engine
from .render import render_expression, render_statement
from .sort_query import sort_query, QuerySorterException
from .database import (
database_exists,
create_database,
drop_database,
escape_like,
is_auto_assigned_date_column,
is_indexed_foreign_key,
non_indexed_foreign_keys,
)
from .orm import (
declarative_base,
getdotattr,
has_changes,
identity,
naturally_equivalent,
primary_keys,
table_name,
)
__all__ = (
create_database,
create_mock_engine,
database_exists,
declarative_base,
defer_except,
drop_database,
escape_like,
getdotattr,
has_changes,
identity,
is_auto_assigned_date_column,
is_indexed_foreign_key,
mock_engine,
naturally_equivalent,
non_indexed_foreign_keys,
primary_keys,
QuerySorterException,
render_expression,
render_statement,
sort_query,
table_name,
)
<commit_msg>Add query_entities to functions module import<commit_after>from .defer_except import defer_except
from .mock import create_mock_engine, mock_engine
from .render import render_expression, render_statement
from .sort_query import sort_query, QuerySorterException
from .database import (
database_exists,
create_database,
drop_database,
escape_like,
is_auto_assigned_date_column,
is_indexed_foreign_key,
non_indexed_foreign_keys,
)
from .orm import (
declarative_base,
getdotattr,
has_changes,
identity,
naturally_equivalent,
query_entities,
primary_keys,
table_name,
)
__all__ = (
create_database,
create_mock_engine,
database_exists,
declarative_base,
defer_except,
drop_database,
escape_like,
getdotattr,
has_changes,
identity,
is_auto_assigned_date_column,
is_indexed_foreign_key,
mock_engine,
naturally_equivalent,
non_indexed_foreign_keys,
primary_keys,
QuerySorterException,
render_expression,
render_statement,
sort_query,
table_name,
)
|
244a8ef2d3976970f8647e5fdd3979932cebe6d7
|
webserver/celery.py
|
webserver/celery.py
|
from __future__ import absolute_import
import os
from celery import Celery
from django.conf import settings
# set the default Django settings module for the 'celery' program.
os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'webserver.settings')
app = Celery('webserver')
# Using a string here means the worker will not have to
# pickle the object when using Windows.
app.config_from_object('django.conf:settings')
app.autodiscover_tasks(lambda: settings.INSTALLED_APPS)
@app.task(bind=True)
def debug_task(self):
print('Request: {0!r}'.format(self.request))
|
from __future__ import absolute_import
import os
from celery import Celery
from django.conf import settings
# set the default Django settings module for the 'celery' program.
os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'webserver.settings')
app = Celery('webserver')
# Using a string here means the worker will not have to
# pickle the object when using Windows.
app.config_from_object('django.conf:settings')
app.autodiscover_tasks(lambda: settings.INSTALLED_APPS)
|
Remove debug task from Celery
|
Remove debug task from Celery
|
Python
|
mit
|
fengthedroid/heroes-of-the-storm-replay-parser,fengthedroid/heroes-of-the-storm-replay-parser,Oize/heroes-of-the-storm-replay-parser,Oize/heroes-of-the-storm-replay-parser,karlgluck/heroes-of-the-storm-replay-parser,Oize/heroes-of-the-storm-replay-parser
|
from __future__ import absolute_import
import os
from celery import Celery
from django.conf import settings
# set the default Django settings module for the 'celery' program.
os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'webserver.settings')
app = Celery('webserver')
# Using a string here means the worker will not have to
# pickle the object when using Windows.
app.config_from_object('django.conf:settings')
app.autodiscover_tasks(lambda: settings.INSTALLED_APPS)
@app.task(bind=True)
def debug_task(self):
print('Request: {0!r}'.format(self.request))
Remove debug task from Celery
|
from __future__ import absolute_import
import os
from celery import Celery
from django.conf import settings
# set the default Django settings module for the 'celery' program.
os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'webserver.settings')
app = Celery('webserver')
# Using a string here means the worker will not have to
# pickle the object when using Windows.
app.config_from_object('django.conf:settings')
app.autodiscover_tasks(lambda: settings.INSTALLED_APPS)
|
<commit_before>from __future__ import absolute_import
import os
from celery import Celery
from django.conf import settings
# set the default Django settings module for the 'celery' program.
os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'webserver.settings')
app = Celery('webserver')
# Using a string here means the worker will not have to
# pickle the object when using Windows.
app.config_from_object('django.conf:settings')
app.autodiscover_tasks(lambda: settings.INSTALLED_APPS)
@app.task(bind=True)
def debug_task(self):
print('Request: {0!r}'.format(self.request))
<commit_msg>Remove debug task from Celery<commit_after>
|
from __future__ import absolute_import
import os
from celery import Celery
from django.conf import settings
# set the default Django settings module for the 'celery' program.
os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'webserver.settings')
app = Celery('webserver')
# Using a string here means the worker will not have to
# pickle the object when using Windows.
app.config_from_object('django.conf:settings')
app.autodiscover_tasks(lambda: settings.INSTALLED_APPS)
|
from __future__ import absolute_import
import os
from celery import Celery
from django.conf import settings
# set the default Django settings module for the 'celery' program.
os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'webserver.settings')
app = Celery('webserver')
# Using a string here means the worker will not have to
# pickle the object when using Windows.
app.config_from_object('django.conf:settings')
app.autodiscover_tasks(lambda: settings.INSTALLED_APPS)
@app.task(bind=True)
def debug_task(self):
print('Request: {0!r}'.format(self.request))
Remove debug task from Celeryfrom __future__ import absolute_import
import os
from celery import Celery
from django.conf import settings
# set the default Django settings module for the 'celery' program.
os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'webserver.settings')
app = Celery('webserver')
# Using a string here means the worker will not have to
# pickle the object when using Windows.
app.config_from_object('django.conf:settings')
app.autodiscover_tasks(lambda: settings.INSTALLED_APPS)
|
<commit_before>from __future__ import absolute_import
import os
from celery import Celery
from django.conf import settings
# set the default Django settings module for the 'celery' program.
os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'webserver.settings')
app = Celery('webserver')
# Using a string here means the worker will not have to
# pickle the object when using Windows.
app.config_from_object('django.conf:settings')
app.autodiscover_tasks(lambda: settings.INSTALLED_APPS)
@app.task(bind=True)
def debug_task(self):
print('Request: {0!r}'.format(self.request))
<commit_msg>Remove debug task from Celery<commit_after>from __future__ import absolute_import
import os
from celery import Celery
from django.conf import settings
# set the default Django settings module for the 'celery' program.
os.environ.setdefault('DJANGO_SETTINGS_MODULE', 'webserver.settings')
app = Celery('webserver')
# Using a string here means the worker will not have to
# pickle the object when using Windows.
app.config_from_object('django.conf:settings')
app.autodiscover_tasks(lambda: settings.INSTALLED_APPS)
|
13ec50a7e2187edb03174ed4a9dbf8767f4c6ad4
|
version.py
|
version.py
|
major = 0
minor=0
patch=0
branch="dev"
timestamp=1376412824.91
|
major = 0
minor=0
patch=8
branch="master"
timestamp=1376412892.53
|
Tag commit for v0.0.8-master generated by gitmake.py
|
Tag commit for v0.0.8-master generated by gitmake.py
|
Python
|
mit
|
ryansturmer/gitmake
|
major = 0
minor=0
patch=0
branch="dev"
timestamp=1376412824.91Tag commit for v0.0.8-master generated by gitmake.py
|
major = 0
minor=0
patch=8
branch="master"
timestamp=1376412892.53
|
<commit_before>major = 0
minor=0
patch=0
branch="dev"
timestamp=1376412824.91<commit_msg>Tag commit for v0.0.8-master generated by gitmake.py<commit_after>
|
major = 0
minor=0
patch=8
branch="master"
timestamp=1376412892.53
|
major = 0
minor=0
patch=0
branch="dev"
timestamp=1376412824.91Tag commit for v0.0.8-master generated by gitmake.pymajor = 0
minor=0
patch=8
branch="master"
timestamp=1376412892.53
|
<commit_before>major = 0
minor=0
patch=0
branch="dev"
timestamp=1376412824.91<commit_msg>Tag commit for v0.0.8-master generated by gitmake.py<commit_after>major = 0
minor=0
patch=8
branch="master"
timestamp=1376412892.53
|
c18884b10f345a8a094a3c4bf589888027d43bd5
|
examples/django_app/example_app/urls.py
|
examples/django_app/example_app/urls.py
|
from django.conf.urls import include, url
from django.contrib import admin
from example_app.views import ChatterBotAppView, ChatterBotApiView
urlpatterns = [
url(r'^$', ChatterBotAppView.as_view(), name='main'),
url(r'^admin/', include(admin.site.urls), name='admin'),
url(r'^api/chatterbot/', ChatterBotApiView.as_view(), name='chatterbot'),
]
|
from django.conf.urls import url
from django.contrib import admin
from example_app.views import ChatterBotAppView, ChatterBotApiView
urlpatterns = [
url(r'^$', ChatterBotAppView.as_view(), name='main'),
url(r'^admin/', admin.site.urls, name='admin'),
url(r'^api/chatterbot/', ChatterBotApiView.as_view(), name='chatterbot'),
]
|
Remove url inlude for Django 2.0
|
Remove url inlude for Django 2.0
|
Python
|
bsd-3-clause
|
gunthercox/ChatterBot,vkosuri/ChatterBot
|
from django.conf.urls import include, url
from django.contrib import admin
from example_app.views import ChatterBotAppView, ChatterBotApiView
urlpatterns = [
url(r'^$', ChatterBotAppView.as_view(), name='main'),
url(r'^admin/', include(admin.site.urls), name='admin'),
url(r'^api/chatterbot/', ChatterBotApiView.as_view(), name='chatterbot'),
]
Remove url inlude for Django 2.0
|
from django.conf.urls import url
from django.contrib import admin
from example_app.views import ChatterBotAppView, ChatterBotApiView
urlpatterns = [
url(r'^$', ChatterBotAppView.as_view(), name='main'),
url(r'^admin/', admin.site.urls, name='admin'),
url(r'^api/chatterbot/', ChatterBotApiView.as_view(), name='chatterbot'),
]
|
<commit_before>from django.conf.urls import include, url
from django.contrib import admin
from example_app.views import ChatterBotAppView, ChatterBotApiView
urlpatterns = [
url(r'^$', ChatterBotAppView.as_view(), name='main'),
url(r'^admin/', include(admin.site.urls), name='admin'),
url(r'^api/chatterbot/', ChatterBotApiView.as_view(), name='chatterbot'),
]
<commit_msg>Remove url inlude for Django 2.0<commit_after>
|
from django.conf.urls import url
from django.contrib import admin
from example_app.views import ChatterBotAppView, ChatterBotApiView
urlpatterns = [
url(r'^$', ChatterBotAppView.as_view(), name='main'),
url(r'^admin/', admin.site.urls, name='admin'),
url(r'^api/chatterbot/', ChatterBotApiView.as_view(), name='chatterbot'),
]
|
from django.conf.urls import include, url
from django.contrib import admin
from example_app.views import ChatterBotAppView, ChatterBotApiView
urlpatterns = [
url(r'^$', ChatterBotAppView.as_view(), name='main'),
url(r'^admin/', include(admin.site.urls), name='admin'),
url(r'^api/chatterbot/', ChatterBotApiView.as_view(), name='chatterbot'),
]
Remove url inlude for Django 2.0from django.conf.urls import url
from django.contrib import admin
from example_app.views import ChatterBotAppView, ChatterBotApiView
urlpatterns = [
url(r'^$', ChatterBotAppView.as_view(), name='main'),
url(r'^admin/', admin.site.urls, name='admin'),
url(r'^api/chatterbot/', ChatterBotApiView.as_view(), name='chatterbot'),
]
|
<commit_before>from django.conf.urls import include, url
from django.contrib import admin
from example_app.views import ChatterBotAppView, ChatterBotApiView
urlpatterns = [
url(r'^$', ChatterBotAppView.as_view(), name='main'),
url(r'^admin/', include(admin.site.urls), name='admin'),
url(r'^api/chatterbot/', ChatterBotApiView.as_view(), name='chatterbot'),
]
<commit_msg>Remove url inlude for Django 2.0<commit_after>from django.conf.urls import url
from django.contrib import admin
from example_app.views import ChatterBotAppView, ChatterBotApiView
urlpatterns = [
url(r'^$', ChatterBotAppView.as_view(), name='main'),
url(r'^admin/', admin.site.urls, name='admin'),
url(r'^api/chatterbot/', ChatterBotApiView.as_view(), name='chatterbot'),
]
|
3243f199fb46d2d6f95ae9afd18b1570f9b5f529
|
astatsscraper/parsing.py
|
astatsscraper/parsing.py
|
def parse_app_page(response):
# Should always be able to grab a title
title = response.xpath('//div[@class = "panel panel-default panel-gameinfo"]/div[@class = "panel-heading"]/text()').extract()[0].strip()
# Parse times into floats
time_to_hundo = response.xpath('//table[@class = "Default1000"]/tr/td[span = "Hours to 100%"]/text()[last()]').extract()[0].strip()
time_to_hundo = time_to_hundo.replace(',', '.')
time_to_hundo = float(time_to_hundo)
# Points may or may not be present, default to 0 if absent
points = response.xpath('//table[@class = "Default1000"]/tr/td[span = "Points"]/text()[last()]').extract()
if not points:
points = 0
else:
points = int(points[0].strip())
yield {
'title': title,
'time to 100%': time_to_hundo,
'points': points,
}
def parse_search_result_for_apps(response):
for href in response.xpath('//table//table//a/@href'):
relative_url = href.extract()
if relative_url.startswith('Steam_Game_Info.php?AppID='):
yield relative_url[:len('Steam_Game_Info.php?AppID=')]
|
def parse_app_page(response):
# Should always be able to grab a title
title = response.xpath('//div[@class = "panel panel-default panel-gameinfo"]/div[@class = "panel-heading"]/text()').extract()[0].strip()
# Parse times into floats
time_to_hundo = response.xpath('//table[@class = "Default1000"]/tr/td[span = "Hours to 100%"]/text()[last()]').extract()[0].strip()
time_to_hundo = time_to_hundo.replace(',', '.')
time_to_hundo = float(time_to_hundo)
# Points may or may not be present, default to 0 if absent
points = response.xpath('//table[@class = "Default1000"]/tr/td[span = "Points"]/text()[last()]').extract()
if not points:
points = 0
else:
points = int(points[0].strip())
yield {
'title': title,
'time to 100%': time_to_hundo,
'points': points,
}
def parse_search_result_for_apps(response):
for href in response.xpath('//table//table//a/@href'):
relative_url = href.extract()
if relative_url.startswith('Steam_Game_Info.php?AppID='):
yield {
'app_id' : relative_url[len('Steam_Game_Info.php?AppID='):]
}
|
Fix up bad last commit
|
Fix up bad last commit
|
Python
|
mit
|
SingingTree/AStatsScraper,SingingTree/AStatsScraper
|
def parse_app_page(response):
# Should always be able to grab a title
title = response.xpath('//div[@class = "panel panel-default panel-gameinfo"]/div[@class = "panel-heading"]/text()').extract()[0].strip()
# Parse times into floats
time_to_hundo = response.xpath('//table[@class = "Default1000"]/tr/td[span = "Hours to 100%"]/text()[last()]').extract()[0].strip()
time_to_hundo = time_to_hundo.replace(',', '.')
time_to_hundo = float(time_to_hundo)
# Points may or may not be present, default to 0 if absent
points = response.xpath('//table[@class = "Default1000"]/tr/td[span = "Points"]/text()[last()]').extract()
if not points:
points = 0
else:
points = int(points[0].strip())
yield {
'title': title,
'time to 100%': time_to_hundo,
'points': points,
}
def parse_search_result_for_apps(response):
for href in response.xpath('//table//table//a/@href'):
relative_url = href.extract()
if relative_url.startswith('Steam_Game_Info.php?AppID='):
yield relative_url[:len('Steam_Game_Info.php?AppID=')]
Fix up bad last commit
|
def parse_app_page(response):
# Should always be able to grab a title
title = response.xpath('//div[@class = "panel panel-default panel-gameinfo"]/div[@class = "panel-heading"]/text()').extract()[0].strip()
# Parse times into floats
time_to_hundo = response.xpath('//table[@class = "Default1000"]/tr/td[span = "Hours to 100%"]/text()[last()]').extract()[0].strip()
time_to_hundo = time_to_hundo.replace(',', '.')
time_to_hundo = float(time_to_hundo)
# Points may or may not be present, default to 0 if absent
points = response.xpath('//table[@class = "Default1000"]/tr/td[span = "Points"]/text()[last()]').extract()
if not points:
points = 0
else:
points = int(points[0].strip())
yield {
'title': title,
'time to 100%': time_to_hundo,
'points': points,
}
def parse_search_result_for_apps(response):
for href in response.xpath('//table//table//a/@href'):
relative_url = href.extract()
if relative_url.startswith('Steam_Game_Info.php?AppID='):
yield {
'app_id' : relative_url[len('Steam_Game_Info.php?AppID='):]
}
|
<commit_before>def parse_app_page(response):
# Should always be able to grab a title
title = response.xpath('//div[@class = "panel panel-default panel-gameinfo"]/div[@class = "panel-heading"]/text()').extract()[0].strip()
# Parse times into floats
time_to_hundo = response.xpath('//table[@class = "Default1000"]/tr/td[span = "Hours to 100%"]/text()[last()]').extract()[0].strip()
time_to_hundo = time_to_hundo.replace(',', '.')
time_to_hundo = float(time_to_hundo)
# Points may or may not be present, default to 0 if absent
points = response.xpath('//table[@class = "Default1000"]/tr/td[span = "Points"]/text()[last()]').extract()
if not points:
points = 0
else:
points = int(points[0].strip())
yield {
'title': title,
'time to 100%': time_to_hundo,
'points': points,
}
def parse_search_result_for_apps(response):
for href in response.xpath('//table//table//a/@href'):
relative_url = href.extract()
if relative_url.startswith('Steam_Game_Info.php?AppID='):
yield relative_url[:len('Steam_Game_Info.php?AppID=')]
<commit_msg>Fix up bad last commit<commit_after>
|
def parse_app_page(response):
# Should always be able to grab a title
title = response.xpath('//div[@class = "panel panel-default panel-gameinfo"]/div[@class = "panel-heading"]/text()').extract()[0].strip()
# Parse times into floats
time_to_hundo = response.xpath('//table[@class = "Default1000"]/tr/td[span = "Hours to 100%"]/text()[last()]').extract()[0].strip()
time_to_hundo = time_to_hundo.replace(',', '.')
time_to_hundo = float(time_to_hundo)
# Points may or may not be present, default to 0 if absent
points = response.xpath('//table[@class = "Default1000"]/tr/td[span = "Points"]/text()[last()]').extract()
if not points:
points = 0
else:
points = int(points[0].strip())
yield {
'title': title,
'time to 100%': time_to_hundo,
'points': points,
}
def parse_search_result_for_apps(response):
for href in response.xpath('//table//table//a/@href'):
relative_url = href.extract()
if relative_url.startswith('Steam_Game_Info.php?AppID='):
yield {
'app_id' : relative_url[len('Steam_Game_Info.php?AppID='):]
}
|
def parse_app_page(response):
# Should always be able to grab a title
title = response.xpath('//div[@class = "panel panel-default panel-gameinfo"]/div[@class = "panel-heading"]/text()').extract()[0].strip()
# Parse times into floats
time_to_hundo = response.xpath('//table[@class = "Default1000"]/tr/td[span = "Hours to 100%"]/text()[last()]').extract()[0].strip()
time_to_hundo = time_to_hundo.replace(',', '.')
time_to_hundo = float(time_to_hundo)
# Points may or may not be present, default to 0 if absent
points = response.xpath('//table[@class = "Default1000"]/tr/td[span = "Points"]/text()[last()]').extract()
if not points:
points = 0
else:
points = int(points[0].strip())
yield {
'title': title,
'time to 100%': time_to_hundo,
'points': points,
}
def parse_search_result_for_apps(response):
for href in response.xpath('//table//table//a/@href'):
relative_url = href.extract()
if relative_url.startswith('Steam_Game_Info.php?AppID='):
yield relative_url[:len('Steam_Game_Info.php?AppID=')]
Fix up bad last commitdef parse_app_page(response):
# Should always be able to grab a title
title = response.xpath('//div[@class = "panel panel-default panel-gameinfo"]/div[@class = "panel-heading"]/text()').extract()[0].strip()
# Parse times into floats
time_to_hundo = response.xpath('//table[@class = "Default1000"]/tr/td[span = "Hours to 100%"]/text()[last()]').extract()[0].strip()
time_to_hundo = time_to_hundo.replace(',', '.')
time_to_hundo = float(time_to_hundo)
# Points may or may not be present, default to 0 if absent
points = response.xpath('//table[@class = "Default1000"]/tr/td[span = "Points"]/text()[last()]').extract()
if not points:
points = 0
else:
points = int(points[0].strip())
yield {
'title': title,
'time to 100%': time_to_hundo,
'points': points,
}
def parse_search_result_for_apps(response):
for href in response.xpath('//table//table//a/@href'):
relative_url = href.extract()
if relative_url.startswith('Steam_Game_Info.php?AppID='):
yield {
'app_id' : relative_url[len('Steam_Game_Info.php?AppID='):]
}
|
<commit_before>def parse_app_page(response):
# Should always be able to grab a title
title = response.xpath('//div[@class = "panel panel-default panel-gameinfo"]/div[@class = "panel-heading"]/text()').extract()[0].strip()
# Parse times into floats
time_to_hundo = response.xpath('//table[@class = "Default1000"]/tr/td[span = "Hours to 100%"]/text()[last()]').extract()[0].strip()
time_to_hundo = time_to_hundo.replace(',', '.')
time_to_hundo = float(time_to_hundo)
# Points may or may not be present, default to 0 if absent
points = response.xpath('//table[@class = "Default1000"]/tr/td[span = "Points"]/text()[last()]').extract()
if not points:
points = 0
else:
points = int(points[0].strip())
yield {
'title': title,
'time to 100%': time_to_hundo,
'points': points,
}
def parse_search_result_for_apps(response):
for href in response.xpath('//table//table//a/@href'):
relative_url = href.extract()
if relative_url.startswith('Steam_Game_Info.php?AppID='):
yield relative_url[:len('Steam_Game_Info.php?AppID=')]
<commit_msg>Fix up bad last commit<commit_after>def parse_app_page(response):
# Should always be able to grab a title
title = response.xpath('//div[@class = "panel panel-default panel-gameinfo"]/div[@class = "panel-heading"]/text()').extract()[0].strip()
# Parse times into floats
time_to_hundo = response.xpath('//table[@class = "Default1000"]/tr/td[span = "Hours to 100%"]/text()[last()]').extract()[0].strip()
time_to_hundo = time_to_hundo.replace(',', '.')
time_to_hundo = float(time_to_hundo)
# Points may or may not be present, default to 0 if absent
points = response.xpath('//table[@class = "Default1000"]/tr/td[span = "Points"]/text()[last()]').extract()
if not points:
points = 0
else:
points = int(points[0].strip())
yield {
'title': title,
'time to 100%': time_to_hundo,
'points': points,
}
def parse_search_result_for_apps(response):
for href in response.xpath('//table//table//a/@href'):
relative_url = href.extract()
if relative_url.startswith('Steam_Game_Info.php?AppID='):
yield {
'app_id' : relative_url[len('Steam_Game_Info.php?AppID='):]
}
|
cd41fdbdb53008c9701213d4f223bb8df0514ecb
|
byceps/util/datetime/timezone.py
|
byceps/util/datetime/timezone.py
|
"""
byceps.util.datetime.timezone
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
Timezone helpers
:Copyright: 2006-2021 Jochen Kupperschmidt
:License: Revised BSD (see `LICENSE` file for details)
"""
from datetime import datetime
from flask import current_app
import pendulum
def local_tz_to_utc(dt: datetime):
"""Convert date/time object from configured default local time to UTC."""
tz_str = get_timezone_string()
return (pendulum.instance(dt)
.set(tz=tz_str)
.in_tz(pendulum.UTC)
# Keep SQLAlchemy from converting it to another zone.
.replace(tzinfo=None))
def utc_to_local_tz(dt: datetime) -> datetime:
"""Convert naive date/time object from UTC to configured time zone."""
tz_str = get_timezone_string()
return pendulum.instance(dt).in_tz(tz_str)
def get_timezone_string() -> str:
"""Return the configured default timezone as a string."""
return current_app.config['TIMEZONE']
|
"""
byceps.util.datetime.timezone
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
Timezone helpers
:Copyright: 2006-2021 Jochen Kupperschmidt
:License: Revised BSD (see `LICENSE` file for details)
"""
from flask import current_app
def get_timezone_string() -> str:
"""Return the configured default timezone as a string."""
return current_app.config['TIMEZONE']
|
Remove unused custom functions `local_tz_to_utc`, `utc_to_local_tz`
|
Remove unused custom functions `local_tz_to_utc`, `utc_to_local_tz`
|
Python
|
bsd-3-clause
|
homeworkprod/byceps,homeworkprod/byceps,homeworkprod/byceps
|
"""
byceps.util.datetime.timezone
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
Timezone helpers
:Copyright: 2006-2021 Jochen Kupperschmidt
:License: Revised BSD (see `LICENSE` file for details)
"""
from datetime import datetime
from flask import current_app
import pendulum
def local_tz_to_utc(dt: datetime):
"""Convert date/time object from configured default local time to UTC."""
tz_str = get_timezone_string()
return (pendulum.instance(dt)
.set(tz=tz_str)
.in_tz(pendulum.UTC)
# Keep SQLAlchemy from converting it to another zone.
.replace(tzinfo=None))
def utc_to_local_tz(dt: datetime) -> datetime:
"""Convert naive date/time object from UTC to configured time zone."""
tz_str = get_timezone_string()
return pendulum.instance(dt).in_tz(tz_str)
def get_timezone_string() -> str:
"""Return the configured default timezone as a string."""
return current_app.config['TIMEZONE']
Remove unused custom functions `local_tz_to_utc`, `utc_to_local_tz`
|
"""
byceps.util.datetime.timezone
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
Timezone helpers
:Copyright: 2006-2021 Jochen Kupperschmidt
:License: Revised BSD (see `LICENSE` file for details)
"""
from flask import current_app
def get_timezone_string() -> str:
"""Return the configured default timezone as a string."""
return current_app.config['TIMEZONE']
|
<commit_before>"""
byceps.util.datetime.timezone
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
Timezone helpers
:Copyright: 2006-2021 Jochen Kupperschmidt
:License: Revised BSD (see `LICENSE` file for details)
"""
from datetime import datetime
from flask import current_app
import pendulum
def local_tz_to_utc(dt: datetime):
"""Convert date/time object from configured default local time to UTC."""
tz_str = get_timezone_string()
return (pendulum.instance(dt)
.set(tz=tz_str)
.in_tz(pendulum.UTC)
# Keep SQLAlchemy from converting it to another zone.
.replace(tzinfo=None))
def utc_to_local_tz(dt: datetime) -> datetime:
"""Convert naive date/time object from UTC to configured time zone."""
tz_str = get_timezone_string()
return pendulum.instance(dt).in_tz(tz_str)
def get_timezone_string() -> str:
"""Return the configured default timezone as a string."""
return current_app.config['TIMEZONE']
<commit_msg>Remove unused custom functions `local_tz_to_utc`, `utc_to_local_tz`<commit_after>
|
"""
byceps.util.datetime.timezone
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
Timezone helpers
:Copyright: 2006-2021 Jochen Kupperschmidt
:License: Revised BSD (see `LICENSE` file for details)
"""
from flask import current_app
def get_timezone_string() -> str:
"""Return the configured default timezone as a string."""
return current_app.config['TIMEZONE']
|
"""
byceps.util.datetime.timezone
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
Timezone helpers
:Copyright: 2006-2021 Jochen Kupperschmidt
:License: Revised BSD (see `LICENSE` file for details)
"""
from datetime import datetime
from flask import current_app
import pendulum
def local_tz_to_utc(dt: datetime):
"""Convert date/time object from configured default local time to UTC."""
tz_str = get_timezone_string()
return (pendulum.instance(dt)
.set(tz=tz_str)
.in_tz(pendulum.UTC)
# Keep SQLAlchemy from converting it to another zone.
.replace(tzinfo=None))
def utc_to_local_tz(dt: datetime) -> datetime:
"""Convert naive date/time object from UTC to configured time zone."""
tz_str = get_timezone_string()
return pendulum.instance(dt).in_tz(tz_str)
def get_timezone_string() -> str:
"""Return the configured default timezone as a string."""
return current_app.config['TIMEZONE']
Remove unused custom functions `local_tz_to_utc`, `utc_to_local_tz`"""
byceps.util.datetime.timezone
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
Timezone helpers
:Copyright: 2006-2021 Jochen Kupperschmidt
:License: Revised BSD (see `LICENSE` file for details)
"""
from flask import current_app
def get_timezone_string() -> str:
"""Return the configured default timezone as a string."""
return current_app.config['TIMEZONE']
|
<commit_before>"""
byceps.util.datetime.timezone
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
Timezone helpers
:Copyright: 2006-2021 Jochen Kupperschmidt
:License: Revised BSD (see `LICENSE` file for details)
"""
from datetime import datetime
from flask import current_app
import pendulum
def local_tz_to_utc(dt: datetime):
"""Convert date/time object from configured default local time to UTC."""
tz_str = get_timezone_string()
return (pendulum.instance(dt)
.set(tz=tz_str)
.in_tz(pendulum.UTC)
# Keep SQLAlchemy from converting it to another zone.
.replace(tzinfo=None))
def utc_to_local_tz(dt: datetime) -> datetime:
"""Convert naive date/time object from UTC to configured time zone."""
tz_str = get_timezone_string()
return pendulum.instance(dt).in_tz(tz_str)
def get_timezone_string() -> str:
"""Return the configured default timezone as a string."""
return current_app.config['TIMEZONE']
<commit_msg>Remove unused custom functions `local_tz_to_utc`, `utc_to_local_tz`<commit_after>"""
byceps.util.datetime.timezone
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
Timezone helpers
:Copyright: 2006-2021 Jochen Kupperschmidt
:License: Revised BSD (see `LICENSE` file for details)
"""
from flask import current_app
def get_timezone_string() -> str:
"""Return the configured default timezone as a string."""
return current_app.config['TIMEZONE']
|
61fdbe0dba79dc19cda5320a0ad1352facf12d3d
|
twine/__init__.py
|
twine/__init__.py
|
# Copyright 2018 Donald Stufft and individual contributors
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# https://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
__all__ = (
"__title__", "__summary__", "__uri__", "__version__", "__author__",
"__email__", "__license__", "__copyright__",
)
__copyright__ = "Copyright 2019 Donald Stufft and individual contributors"
try:
import importlib.metadata as importlib_metadata
except ImportError:
import importlib_metadata
metadata = importlib_metadata.metadata('twine')
__title__ = metadata['name']
__summary__ = metadata['summary']
__uri__ = metadata['home-page']
__version__ = metadata['version']
__author__ = metadata['author']
__email__ = metadata['author-email']
__license__ = metadata['license']
|
# Copyright 2018 Donald Stufft and individual contributors
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# https://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
__all__ = (
"__title__", "__summary__", "__uri__", "__version__", "__author__",
"__email__", "__license__", "__copyright__",
)
__copyright__ = "Copyright 2019 Donald Stufft and individual contributors"
try:
# https://github.com/python/mypy/issues/1393
from importlib.metadata import metadata # type: ignore
except ImportError:
# https://github.com/python/mypy/issues/1153
from importlib_metadata import metadata # type: ignore
twine_metadata = metadata('twine')
__title__ = twine_metadata['name']
__summary__ = twine_metadata['summary']
__uri__ = twine_metadata['home-page']
__version__ = twine_metadata['version']
__author__ = twine_metadata['author']
__email__ = twine_metadata['author-email']
__license__ = twine_metadata['license']
|
Rework imports and ignore known mypy issues
|
Rework imports and ignore known mypy issues
|
Python
|
apache-2.0
|
pypa/twine
|
# Copyright 2018 Donald Stufft and individual contributors
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# https://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
__all__ = (
"__title__", "__summary__", "__uri__", "__version__", "__author__",
"__email__", "__license__", "__copyright__",
)
__copyright__ = "Copyright 2019 Donald Stufft and individual contributors"
try:
import importlib.metadata as importlib_metadata
except ImportError:
import importlib_metadata
metadata = importlib_metadata.metadata('twine')
__title__ = metadata['name']
__summary__ = metadata['summary']
__uri__ = metadata['home-page']
__version__ = metadata['version']
__author__ = metadata['author']
__email__ = metadata['author-email']
__license__ = metadata['license']
Rework imports and ignore known mypy issues
|
# Copyright 2018 Donald Stufft and individual contributors
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# https://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
__all__ = (
"__title__", "__summary__", "__uri__", "__version__", "__author__",
"__email__", "__license__", "__copyright__",
)
__copyright__ = "Copyright 2019 Donald Stufft and individual contributors"
try:
# https://github.com/python/mypy/issues/1393
from importlib.metadata import metadata # type: ignore
except ImportError:
# https://github.com/python/mypy/issues/1153
from importlib_metadata import metadata # type: ignore
twine_metadata = metadata('twine')
__title__ = twine_metadata['name']
__summary__ = twine_metadata['summary']
__uri__ = twine_metadata['home-page']
__version__ = twine_metadata['version']
__author__ = twine_metadata['author']
__email__ = twine_metadata['author-email']
__license__ = twine_metadata['license']
|
<commit_before># Copyright 2018 Donald Stufft and individual contributors
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# https://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
__all__ = (
"__title__", "__summary__", "__uri__", "__version__", "__author__",
"__email__", "__license__", "__copyright__",
)
__copyright__ = "Copyright 2019 Donald Stufft and individual contributors"
try:
import importlib.metadata as importlib_metadata
except ImportError:
import importlib_metadata
metadata = importlib_metadata.metadata('twine')
__title__ = metadata['name']
__summary__ = metadata['summary']
__uri__ = metadata['home-page']
__version__ = metadata['version']
__author__ = metadata['author']
__email__ = metadata['author-email']
__license__ = metadata['license']
<commit_msg>Rework imports and ignore known mypy issues<commit_after>
|
# Copyright 2018 Donald Stufft and individual contributors
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# https://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
__all__ = (
"__title__", "__summary__", "__uri__", "__version__", "__author__",
"__email__", "__license__", "__copyright__",
)
__copyright__ = "Copyright 2019 Donald Stufft and individual contributors"
try:
# https://github.com/python/mypy/issues/1393
from importlib.metadata import metadata # type: ignore
except ImportError:
# https://github.com/python/mypy/issues/1153
from importlib_metadata import metadata # type: ignore
twine_metadata = metadata('twine')
__title__ = twine_metadata['name']
__summary__ = twine_metadata['summary']
__uri__ = twine_metadata['home-page']
__version__ = twine_metadata['version']
__author__ = twine_metadata['author']
__email__ = twine_metadata['author-email']
__license__ = twine_metadata['license']
|
# Copyright 2018 Donald Stufft and individual contributors
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# https://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
__all__ = (
"__title__", "__summary__", "__uri__", "__version__", "__author__",
"__email__", "__license__", "__copyright__",
)
__copyright__ = "Copyright 2019 Donald Stufft and individual contributors"
try:
import importlib.metadata as importlib_metadata
except ImportError:
import importlib_metadata
metadata = importlib_metadata.metadata('twine')
__title__ = metadata['name']
__summary__ = metadata['summary']
__uri__ = metadata['home-page']
__version__ = metadata['version']
__author__ = metadata['author']
__email__ = metadata['author-email']
__license__ = metadata['license']
Rework imports and ignore known mypy issues# Copyright 2018 Donald Stufft and individual contributors
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# https://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
__all__ = (
"__title__", "__summary__", "__uri__", "__version__", "__author__",
"__email__", "__license__", "__copyright__",
)
__copyright__ = "Copyright 2019 Donald Stufft and individual contributors"
try:
# https://github.com/python/mypy/issues/1393
from importlib.metadata import metadata # type: ignore
except ImportError:
# https://github.com/python/mypy/issues/1153
from importlib_metadata import metadata # type: ignore
twine_metadata = metadata('twine')
__title__ = twine_metadata['name']
__summary__ = twine_metadata['summary']
__uri__ = twine_metadata['home-page']
__version__ = twine_metadata['version']
__author__ = twine_metadata['author']
__email__ = twine_metadata['author-email']
__license__ = twine_metadata['license']
|
<commit_before># Copyright 2018 Donald Stufft and individual contributors
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# https://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
__all__ = (
"__title__", "__summary__", "__uri__", "__version__", "__author__",
"__email__", "__license__", "__copyright__",
)
__copyright__ = "Copyright 2019 Donald Stufft and individual contributors"
try:
import importlib.metadata as importlib_metadata
except ImportError:
import importlib_metadata
metadata = importlib_metadata.metadata('twine')
__title__ = metadata['name']
__summary__ = metadata['summary']
__uri__ = metadata['home-page']
__version__ = metadata['version']
__author__ = metadata['author']
__email__ = metadata['author-email']
__license__ = metadata['license']
<commit_msg>Rework imports and ignore known mypy issues<commit_after># Copyright 2018 Donald Stufft and individual contributors
#
# Licensed under the Apache License, Version 2.0 (the "License");
# you may not use this file except in compliance with the License.
# You may obtain a copy of the License at
#
# https://www.apache.org/licenses/LICENSE-2.0
#
# Unless required by applicable law or agreed to in writing, software
# distributed under the License is distributed on an "AS IS" BASIS,
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
# See the License for the specific language governing permissions and
# limitations under the License.
__all__ = (
"__title__", "__summary__", "__uri__", "__version__", "__author__",
"__email__", "__license__", "__copyright__",
)
__copyright__ = "Copyright 2019 Donald Stufft and individual contributors"
try:
# https://github.com/python/mypy/issues/1393
from importlib.metadata import metadata # type: ignore
except ImportError:
# https://github.com/python/mypy/issues/1153
from importlib_metadata import metadata # type: ignore
twine_metadata = metadata('twine')
__title__ = twine_metadata['name']
__summary__ = twine_metadata['summary']
__uri__ = twine_metadata['home-page']
__version__ = twine_metadata['version']
__author__ = twine_metadata['author']
__email__ = twine_metadata['author-email']
__license__ = twine_metadata['license']
|
1bc4507234d87b1ed246501165fa1d8138bf5ca6
|
cheddar/exceptions.py
|
cheddar/exceptions.py
|
"""
Shared exception.
"""
class BadRequestError(Exception):
pass
class ConflictError(Exception):
pass
class NotFoundError(Exception):
def __init__(self, status_code=None):
self.status_code = status_code
|
"""
Shared exception.
"""
class BadRequestError(Exception):
pass
class ConflictError(Exception):
pass
class NotFoundError(Exception):
def __init__(self, status_code=None):
super(NotFoundError, self).__init__()
self.status_code = status_code
|
Fix for pypy compatibility: must super's __init__
|
Fix for pypy compatibility: must super's __init__
|
Python
|
apache-2.0
|
jessemyers/cheddar,jessemyers/cheddar
|
"""
Shared exception.
"""
class BadRequestError(Exception):
pass
class ConflictError(Exception):
pass
class NotFoundError(Exception):
def __init__(self, status_code=None):
self.status_code = status_code
Fix for pypy compatibility: must super's __init__
|
"""
Shared exception.
"""
class BadRequestError(Exception):
pass
class ConflictError(Exception):
pass
class NotFoundError(Exception):
def __init__(self, status_code=None):
super(NotFoundError, self).__init__()
self.status_code = status_code
|
<commit_before>"""
Shared exception.
"""
class BadRequestError(Exception):
pass
class ConflictError(Exception):
pass
class NotFoundError(Exception):
def __init__(self, status_code=None):
self.status_code = status_code
<commit_msg>Fix for pypy compatibility: must super's __init__<commit_after>
|
"""
Shared exception.
"""
class BadRequestError(Exception):
pass
class ConflictError(Exception):
pass
class NotFoundError(Exception):
def __init__(self, status_code=None):
super(NotFoundError, self).__init__()
self.status_code = status_code
|
"""
Shared exception.
"""
class BadRequestError(Exception):
pass
class ConflictError(Exception):
pass
class NotFoundError(Exception):
def __init__(self, status_code=None):
self.status_code = status_code
Fix for pypy compatibility: must super's __init__"""
Shared exception.
"""
class BadRequestError(Exception):
pass
class ConflictError(Exception):
pass
class NotFoundError(Exception):
def __init__(self, status_code=None):
super(NotFoundError, self).__init__()
self.status_code = status_code
|
<commit_before>"""
Shared exception.
"""
class BadRequestError(Exception):
pass
class ConflictError(Exception):
pass
class NotFoundError(Exception):
def __init__(self, status_code=None):
self.status_code = status_code
<commit_msg>Fix for pypy compatibility: must super's __init__<commit_after>"""
Shared exception.
"""
class BadRequestError(Exception):
pass
class ConflictError(Exception):
pass
class NotFoundError(Exception):
def __init__(self, status_code=None):
super(NotFoundError, self).__init__()
self.status_code = status_code
|
ac9cd5ff007ee131e97f70c49c763f79f06ebf5a
|
green/test/test_integration.py
|
green/test/test_integration.py
|
import multiprocessing
import os
from pathlib import PurePath
import subprocess
import sys
import tempfile
from textwrap import dedent
import unittest
try:
from unittest.mock import MagicMock
except:
from mock import MagicMock
from green import cmdline
class TestFinalizer(unittest.TestCase):
def setUp(self):
self.tmpdir = tempfile.mkdtemp()
def test_finalizer(self):
"""
Test that the finalizer works on Python 3.8+
"""
sub_tmpdir = tempfile.mkdtemp(dir=self.tmpdir)
for i in range(multiprocessing.cpu_count() * 2):
fh = open(os.path.join(sub_tmpdir, f"test_finalizer{i}.py"), "w")
fh.write(
dedent(
f"""
import unittest
class Pass{i}(unittest.TestCase):
def test_pass{i}(self):
pass
def msg():
print("finalizer worked")
"""
)
)
fh.close()
args = [
sys.executable,
"-m",
"green.cmdline",
"--finalizer=test_finalizer0.msg",
"--maxtasksperchild=1",
]
pythonpath = str(PurePath(__file__).parent.parent.parent)
output = subprocess.run(
args,
cwd=sub_tmpdir,
stdout=subprocess.PIPE,
stderr=subprocess.STDOUT,
env={"PYTHONPATH": pythonpath},
timeout=10,
).stdout.decode("utf-8")
self.assertIn("finalizer worked", output)
|
import copy
import multiprocessing
import os
from pathlib import PurePath
import subprocess
import sys
import tempfile
from textwrap import dedent
import unittest
try:
from unittest.mock import MagicMock
except:
from mock import MagicMock
from green import cmdline
class TestFinalizer(unittest.TestCase):
def setUp(self):
self.tmpdir = tempfile.mkdtemp()
def test_finalizer(self):
"""
Test that the finalizer works on Python 3.8+
"""
sub_tmpdir = tempfile.mkdtemp(dir=self.tmpdir)
for i in range(multiprocessing.cpu_count() * 2):
fh = open(os.path.join(sub_tmpdir, f"test_finalizer{i}.py"), "w")
fh.write(
dedent(
f"""
import unittest
class Pass{i}(unittest.TestCase):
def test_pass{i}(self):
pass
def msg():
print("finalizer worked")
"""
)
)
fh.close()
args = [
sys.executable,
"-m",
"green.cmdline",
"--finalizer=test_finalizer0.msg",
"--maxtasksperchild=1",
]
pythonpath = str(PurePath(__file__).parent.parent.parent)
env = copy.deepcopy(os.environ)
env["PYTHONPATH"] = pythonpath
output = subprocess.run(
args,
cwd=sub_tmpdir,
stdout=subprocess.PIPE,
stderr=subprocess.STDOUT,
env=env,
timeout=10,
).stdout.decode("utf-8")
self.assertIn("finalizer worked", output)
|
Include the entire existing environment for integration tests subprocesses
|
Include the entire existing environment for integration tests subprocesses
|
Python
|
mit
|
CleanCut/green,CleanCut/green
|
import multiprocessing
import os
from pathlib import PurePath
import subprocess
import sys
import tempfile
from textwrap import dedent
import unittest
try:
from unittest.mock import MagicMock
except:
from mock import MagicMock
from green import cmdline
class TestFinalizer(unittest.TestCase):
def setUp(self):
self.tmpdir = tempfile.mkdtemp()
def test_finalizer(self):
"""
Test that the finalizer works on Python 3.8+
"""
sub_tmpdir = tempfile.mkdtemp(dir=self.tmpdir)
for i in range(multiprocessing.cpu_count() * 2):
fh = open(os.path.join(sub_tmpdir, f"test_finalizer{i}.py"), "w")
fh.write(
dedent(
f"""
import unittest
class Pass{i}(unittest.TestCase):
def test_pass{i}(self):
pass
def msg():
print("finalizer worked")
"""
)
)
fh.close()
args = [
sys.executable,
"-m",
"green.cmdline",
"--finalizer=test_finalizer0.msg",
"--maxtasksperchild=1",
]
pythonpath = str(PurePath(__file__).parent.parent.parent)
output = subprocess.run(
args,
cwd=sub_tmpdir,
stdout=subprocess.PIPE,
stderr=subprocess.STDOUT,
env={"PYTHONPATH": pythonpath},
timeout=10,
).stdout.decode("utf-8")
self.assertIn("finalizer worked", output)
Include the entire existing environment for integration tests subprocesses
|
import copy
import multiprocessing
import os
from pathlib import PurePath
import subprocess
import sys
import tempfile
from textwrap import dedent
import unittest
try:
from unittest.mock import MagicMock
except:
from mock import MagicMock
from green import cmdline
class TestFinalizer(unittest.TestCase):
def setUp(self):
self.tmpdir = tempfile.mkdtemp()
def test_finalizer(self):
"""
Test that the finalizer works on Python 3.8+
"""
sub_tmpdir = tempfile.mkdtemp(dir=self.tmpdir)
for i in range(multiprocessing.cpu_count() * 2):
fh = open(os.path.join(sub_tmpdir, f"test_finalizer{i}.py"), "w")
fh.write(
dedent(
f"""
import unittest
class Pass{i}(unittest.TestCase):
def test_pass{i}(self):
pass
def msg():
print("finalizer worked")
"""
)
)
fh.close()
args = [
sys.executable,
"-m",
"green.cmdline",
"--finalizer=test_finalizer0.msg",
"--maxtasksperchild=1",
]
pythonpath = str(PurePath(__file__).parent.parent.parent)
env = copy.deepcopy(os.environ)
env["PYTHONPATH"] = pythonpath
output = subprocess.run(
args,
cwd=sub_tmpdir,
stdout=subprocess.PIPE,
stderr=subprocess.STDOUT,
env=env,
timeout=10,
).stdout.decode("utf-8")
self.assertIn("finalizer worked", output)
|
<commit_before>import multiprocessing
import os
from pathlib import PurePath
import subprocess
import sys
import tempfile
from textwrap import dedent
import unittest
try:
from unittest.mock import MagicMock
except:
from mock import MagicMock
from green import cmdline
class TestFinalizer(unittest.TestCase):
def setUp(self):
self.tmpdir = tempfile.mkdtemp()
def test_finalizer(self):
"""
Test that the finalizer works on Python 3.8+
"""
sub_tmpdir = tempfile.mkdtemp(dir=self.tmpdir)
for i in range(multiprocessing.cpu_count() * 2):
fh = open(os.path.join(sub_tmpdir, f"test_finalizer{i}.py"), "w")
fh.write(
dedent(
f"""
import unittest
class Pass{i}(unittest.TestCase):
def test_pass{i}(self):
pass
def msg():
print("finalizer worked")
"""
)
)
fh.close()
args = [
sys.executable,
"-m",
"green.cmdline",
"--finalizer=test_finalizer0.msg",
"--maxtasksperchild=1",
]
pythonpath = str(PurePath(__file__).parent.parent.parent)
output = subprocess.run(
args,
cwd=sub_tmpdir,
stdout=subprocess.PIPE,
stderr=subprocess.STDOUT,
env={"PYTHONPATH": pythonpath},
timeout=10,
).stdout.decode("utf-8")
self.assertIn("finalizer worked", output)
<commit_msg>Include the entire existing environment for integration tests subprocesses<commit_after>
|
import copy
import multiprocessing
import os
from pathlib import PurePath
import subprocess
import sys
import tempfile
from textwrap import dedent
import unittest
try:
from unittest.mock import MagicMock
except:
from mock import MagicMock
from green import cmdline
class TestFinalizer(unittest.TestCase):
def setUp(self):
self.tmpdir = tempfile.mkdtemp()
def test_finalizer(self):
"""
Test that the finalizer works on Python 3.8+
"""
sub_tmpdir = tempfile.mkdtemp(dir=self.tmpdir)
for i in range(multiprocessing.cpu_count() * 2):
fh = open(os.path.join(sub_tmpdir, f"test_finalizer{i}.py"), "w")
fh.write(
dedent(
f"""
import unittest
class Pass{i}(unittest.TestCase):
def test_pass{i}(self):
pass
def msg():
print("finalizer worked")
"""
)
)
fh.close()
args = [
sys.executable,
"-m",
"green.cmdline",
"--finalizer=test_finalizer0.msg",
"--maxtasksperchild=1",
]
pythonpath = str(PurePath(__file__).parent.parent.parent)
env = copy.deepcopy(os.environ)
env["PYTHONPATH"] = pythonpath
output = subprocess.run(
args,
cwd=sub_tmpdir,
stdout=subprocess.PIPE,
stderr=subprocess.STDOUT,
env=env,
timeout=10,
).stdout.decode("utf-8")
self.assertIn("finalizer worked", output)
|
import multiprocessing
import os
from pathlib import PurePath
import subprocess
import sys
import tempfile
from textwrap import dedent
import unittest
try:
from unittest.mock import MagicMock
except:
from mock import MagicMock
from green import cmdline
class TestFinalizer(unittest.TestCase):
def setUp(self):
self.tmpdir = tempfile.mkdtemp()
def test_finalizer(self):
"""
Test that the finalizer works on Python 3.8+
"""
sub_tmpdir = tempfile.mkdtemp(dir=self.tmpdir)
for i in range(multiprocessing.cpu_count() * 2):
fh = open(os.path.join(sub_tmpdir, f"test_finalizer{i}.py"), "w")
fh.write(
dedent(
f"""
import unittest
class Pass{i}(unittest.TestCase):
def test_pass{i}(self):
pass
def msg():
print("finalizer worked")
"""
)
)
fh.close()
args = [
sys.executable,
"-m",
"green.cmdline",
"--finalizer=test_finalizer0.msg",
"--maxtasksperchild=1",
]
pythonpath = str(PurePath(__file__).parent.parent.parent)
output = subprocess.run(
args,
cwd=sub_tmpdir,
stdout=subprocess.PIPE,
stderr=subprocess.STDOUT,
env={"PYTHONPATH": pythonpath},
timeout=10,
).stdout.decode("utf-8")
self.assertIn("finalizer worked", output)
Include the entire existing environment for integration tests subprocessesimport copy
import multiprocessing
import os
from pathlib import PurePath
import subprocess
import sys
import tempfile
from textwrap import dedent
import unittest
try:
from unittest.mock import MagicMock
except:
from mock import MagicMock
from green import cmdline
class TestFinalizer(unittest.TestCase):
def setUp(self):
self.tmpdir = tempfile.mkdtemp()
def test_finalizer(self):
"""
Test that the finalizer works on Python 3.8+
"""
sub_tmpdir = tempfile.mkdtemp(dir=self.tmpdir)
for i in range(multiprocessing.cpu_count() * 2):
fh = open(os.path.join(sub_tmpdir, f"test_finalizer{i}.py"), "w")
fh.write(
dedent(
f"""
import unittest
class Pass{i}(unittest.TestCase):
def test_pass{i}(self):
pass
def msg():
print("finalizer worked")
"""
)
)
fh.close()
args = [
sys.executable,
"-m",
"green.cmdline",
"--finalizer=test_finalizer0.msg",
"--maxtasksperchild=1",
]
pythonpath = str(PurePath(__file__).parent.parent.parent)
env = copy.deepcopy(os.environ)
env["PYTHONPATH"] = pythonpath
output = subprocess.run(
args,
cwd=sub_tmpdir,
stdout=subprocess.PIPE,
stderr=subprocess.STDOUT,
env=env,
timeout=10,
).stdout.decode("utf-8")
self.assertIn("finalizer worked", output)
|
<commit_before>import multiprocessing
import os
from pathlib import PurePath
import subprocess
import sys
import tempfile
from textwrap import dedent
import unittest
try:
from unittest.mock import MagicMock
except:
from mock import MagicMock
from green import cmdline
class TestFinalizer(unittest.TestCase):
def setUp(self):
self.tmpdir = tempfile.mkdtemp()
def test_finalizer(self):
"""
Test that the finalizer works on Python 3.8+
"""
sub_tmpdir = tempfile.mkdtemp(dir=self.tmpdir)
for i in range(multiprocessing.cpu_count() * 2):
fh = open(os.path.join(sub_tmpdir, f"test_finalizer{i}.py"), "w")
fh.write(
dedent(
f"""
import unittest
class Pass{i}(unittest.TestCase):
def test_pass{i}(self):
pass
def msg():
print("finalizer worked")
"""
)
)
fh.close()
args = [
sys.executable,
"-m",
"green.cmdline",
"--finalizer=test_finalizer0.msg",
"--maxtasksperchild=1",
]
pythonpath = str(PurePath(__file__).parent.parent.parent)
output = subprocess.run(
args,
cwd=sub_tmpdir,
stdout=subprocess.PIPE,
stderr=subprocess.STDOUT,
env={"PYTHONPATH": pythonpath},
timeout=10,
).stdout.decode("utf-8")
self.assertIn("finalizer worked", output)
<commit_msg>Include the entire existing environment for integration tests subprocesses<commit_after>import copy
import multiprocessing
import os
from pathlib import PurePath
import subprocess
import sys
import tempfile
from textwrap import dedent
import unittest
try:
from unittest.mock import MagicMock
except:
from mock import MagicMock
from green import cmdline
class TestFinalizer(unittest.TestCase):
def setUp(self):
self.tmpdir = tempfile.mkdtemp()
def test_finalizer(self):
"""
Test that the finalizer works on Python 3.8+
"""
sub_tmpdir = tempfile.mkdtemp(dir=self.tmpdir)
for i in range(multiprocessing.cpu_count() * 2):
fh = open(os.path.join(sub_tmpdir, f"test_finalizer{i}.py"), "w")
fh.write(
dedent(
f"""
import unittest
class Pass{i}(unittest.TestCase):
def test_pass{i}(self):
pass
def msg():
print("finalizer worked")
"""
)
)
fh.close()
args = [
sys.executable,
"-m",
"green.cmdline",
"--finalizer=test_finalizer0.msg",
"--maxtasksperchild=1",
]
pythonpath = str(PurePath(__file__).parent.parent.parent)
env = copy.deepcopy(os.environ)
env["PYTHONPATH"] = pythonpath
output = subprocess.run(
args,
cwd=sub_tmpdir,
stdout=subprocess.PIPE,
stderr=subprocess.STDOUT,
env=env,
timeout=10,
).stdout.decode("utf-8")
self.assertIn("finalizer worked", output)
|
290a1f7a2c6860ec57bdb74b9c97207e93e611f0
|
visualize_data.py
|
visualize_data.py
|
from __future__ import division
import argparse
import cv2
import h5py
import util
def main():
parser = argparse.ArgumentParser()
parser.add_argument('hdf5_fname', type=str)
parser.add_argument('--vis_scale', '-r', type=int, default=10, metavar='R', help='rescale image by R for visualization')
args = parser.parse_args()
with h5py.File(args.hdf5_fname, 'r') as hdf5_file:
for image_curr, vel, image_diff in zip(hdf5_file['image_curr'], hdf5_file['vel'], hdf5_file['image_diff']):
image_next = image_curr + image_diff
vis_image, done = util.visualize_images_callback(image_curr, image_next, vis_scale=args.vis_scale, delay=0)
if done:
break
cv2.destroyAllWindows()
if __name__ == "__main__":
main()
|
from __future__ import division
import argparse
import cv2
import h5py
import util
def main():
parser = argparse.ArgumentParser()
parser.add_argument('hdf5_fname', type=str)
parser.add_argument('--vis_scale', '-r', type=int, default=10, metavar='R', help='rescale image by R for visualization')
parser.add_argument('--reverse', action='store_true')
args = parser.parse_args()
with h5py.File(args.hdf5_fname, 'r') as hdf5_file:
dsets = (hdf5_file['image_curr'], hdf5_file['vel'], hdf5_file['image_diff'])
if args.reverse:
dsets = tuple(dset[()][::-1] for dset in dsets)
for image_curr, vel, image_diff in zip(*dsets):
image_next = image_curr + image_diff
vis_image, done = util.visualize_images_callback(image_curr, image_next, vis_scale=args.vis_scale, delay=0)
if done:
break
cv2.destroyAllWindows()
if __name__ == "__main__":
main()
|
Add option to visualize data in reverse
|
Add option to visualize data in reverse
|
Python
|
mit
|
alexlee-gk/visual_dynamics
|
from __future__ import division
import argparse
import cv2
import h5py
import util
def main():
parser = argparse.ArgumentParser()
parser.add_argument('hdf5_fname', type=str)
parser.add_argument('--vis_scale', '-r', type=int, default=10, metavar='R', help='rescale image by R for visualization')
args = parser.parse_args()
with h5py.File(args.hdf5_fname, 'r') as hdf5_file:
for image_curr, vel, image_diff in zip(hdf5_file['image_curr'], hdf5_file['vel'], hdf5_file['image_diff']):
image_next = image_curr + image_diff
vis_image, done = util.visualize_images_callback(image_curr, image_next, vis_scale=args.vis_scale, delay=0)
if done:
break
cv2.destroyAllWindows()
if __name__ == "__main__":
main()
Add option to visualize data in reverse
|
from __future__ import division
import argparse
import cv2
import h5py
import util
def main():
parser = argparse.ArgumentParser()
parser.add_argument('hdf5_fname', type=str)
parser.add_argument('--vis_scale', '-r', type=int, default=10, metavar='R', help='rescale image by R for visualization')
parser.add_argument('--reverse', action='store_true')
args = parser.parse_args()
with h5py.File(args.hdf5_fname, 'r') as hdf5_file:
dsets = (hdf5_file['image_curr'], hdf5_file['vel'], hdf5_file['image_diff'])
if args.reverse:
dsets = tuple(dset[()][::-1] for dset in dsets)
for image_curr, vel, image_diff in zip(*dsets):
image_next = image_curr + image_diff
vis_image, done = util.visualize_images_callback(image_curr, image_next, vis_scale=args.vis_scale, delay=0)
if done:
break
cv2.destroyAllWindows()
if __name__ == "__main__":
main()
|
<commit_before>from __future__ import division
import argparse
import cv2
import h5py
import util
def main():
parser = argparse.ArgumentParser()
parser.add_argument('hdf5_fname', type=str)
parser.add_argument('--vis_scale', '-r', type=int, default=10, metavar='R', help='rescale image by R for visualization')
args = parser.parse_args()
with h5py.File(args.hdf5_fname, 'r') as hdf5_file:
for image_curr, vel, image_diff in zip(hdf5_file['image_curr'], hdf5_file['vel'], hdf5_file['image_diff']):
image_next = image_curr + image_diff
vis_image, done = util.visualize_images_callback(image_curr, image_next, vis_scale=args.vis_scale, delay=0)
if done:
break
cv2.destroyAllWindows()
if __name__ == "__main__":
main()
<commit_msg>Add option to visualize data in reverse<commit_after>
|
from __future__ import division
import argparse
import cv2
import h5py
import util
def main():
parser = argparse.ArgumentParser()
parser.add_argument('hdf5_fname', type=str)
parser.add_argument('--vis_scale', '-r', type=int, default=10, metavar='R', help='rescale image by R for visualization')
parser.add_argument('--reverse', action='store_true')
args = parser.parse_args()
with h5py.File(args.hdf5_fname, 'r') as hdf5_file:
dsets = (hdf5_file['image_curr'], hdf5_file['vel'], hdf5_file['image_diff'])
if args.reverse:
dsets = tuple(dset[()][::-1] for dset in dsets)
for image_curr, vel, image_diff in zip(*dsets):
image_next = image_curr + image_diff
vis_image, done = util.visualize_images_callback(image_curr, image_next, vis_scale=args.vis_scale, delay=0)
if done:
break
cv2.destroyAllWindows()
if __name__ == "__main__":
main()
|
from __future__ import division
import argparse
import cv2
import h5py
import util
def main():
parser = argparse.ArgumentParser()
parser.add_argument('hdf5_fname', type=str)
parser.add_argument('--vis_scale', '-r', type=int, default=10, metavar='R', help='rescale image by R for visualization')
args = parser.parse_args()
with h5py.File(args.hdf5_fname, 'r') as hdf5_file:
for image_curr, vel, image_diff in zip(hdf5_file['image_curr'], hdf5_file['vel'], hdf5_file['image_diff']):
image_next = image_curr + image_diff
vis_image, done = util.visualize_images_callback(image_curr, image_next, vis_scale=args.vis_scale, delay=0)
if done:
break
cv2.destroyAllWindows()
if __name__ == "__main__":
main()
Add option to visualize data in reversefrom __future__ import division
import argparse
import cv2
import h5py
import util
def main():
parser = argparse.ArgumentParser()
parser.add_argument('hdf5_fname', type=str)
parser.add_argument('--vis_scale', '-r', type=int, default=10, metavar='R', help='rescale image by R for visualization')
parser.add_argument('--reverse', action='store_true')
args = parser.parse_args()
with h5py.File(args.hdf5_fname, 'r') as hdf5_file:
dsets = (hdf5_file['image_curr'], hdf5_file['vel'], hdf5_file['image_diff'])
if args.reverse:
dsets = tuple(dset[()][::-1] for dset in dsets)
for image_curr, vel, image_diff in zip(*dsets):
image_next = image_curr + image_diff
vis_image, done = util.visualize_images_callback(image_curr, image_next, vis_scale=args.vis_scale, delay=0)
if done:
break
cv2.destroyAllWindows()
if __name__ == "__main__":
main()
|
<commit_before>from __future__ import division
import argparse
import cv2
import h5py
import util
def main():
parser = argparse.ArgumentParser()
parser.add_argument('hdf5_fname', type=str)
parser.add_argument('--vis_scale', '-r', type=int, default=10, metavar='R', help='rescale image by R for visualization')
args = parser.parse_args()
with h5py.File(args.hdf5_fname, 'r') as hdf5_file:
for image_curr, vel, image_diff in zip(hdf5_file['image_curr'], hdf5_file['vel'], hdf5_file['image_diff']):
image_next = image_curr + image_diff
vis_image, done = util.visualize_images_callback(image_curr, image_next, vis_scale=args.vis_scale, delay=0)
if done:
break
cv2.destroyAllWindows()
if __name__ == "__main__":
main()
<commit_msg>Add option to visualize data in reverse<commit_after>from __future__ import division
import argparse
import cv2
import h5py
import util
def main():
parser = argparse.ArgumentParser()
parser.add_argument('hdf5_fname', type=str)
parser.add_argument('--vis_scale', '-r', type=int, default=10, metavar='R', help='rescale image by R for visualization')
parser.add_argument('--reverse', action='store_true')
args = parser.parse_args()
with h5py.File(args.hdf5_fname, 'r') as hdf5_file:
dsets = (hdf5_file['image_curr'], hdf5_file['vel'], hdf5_file['image_diff'])
if args.reverse:
dsets = tuple(dset[()][::-1] for dset in dsets)
for image_curr, vel, image_diff in zip(*dsets):
image_next = image_curr + image_diff
vis_image, done = util.visualize_images_callback(image_curr, image_next, vis_scale=args.vis_scale, delay=0)
if done:
break
cv2.destroyAllWindows()
if __name__ == "__main__":
main()
|
bf6d6cdaf946af7ce8d1aa6831e7da9b47fef54f
|
user_deletion/managers.py
|
user_deletion/managers.py
|
from dateutil.relativedelta import relativedelta
from django.utils import timezone
class UserDeletionManagerMixin:
def users_to_notify(self):
"""Finds all users who have been inactive and not yet notified."""
from django.apps import apps
user_deletion_config = apps.get_app_config('user_deletion')
threshold = timezone.now() - relativedelta(
months=user_deletion_config.MONTH_NOTIFICATION,
)
return self.filter(last_login__lte=threshold, notified=False)
def users_to_delete(self):
"""Finds all users who have been inactive and were notified."""
from django.apps import apps
user_deletion_config = apps.get_app_config('user_deletion')
threshold = timezone.now() - relativedelta(
months=user_deletion_config.MONTH_DELETION,
)
return self.filter(last_login__lte=threshold, notified=True)
|
from dateutil.relativedelta import relativedelta
from django.apps import apps
from django.utils import timezone
class UserDeletionManagerMixin:
def users_to_notify(self):
"""Finds all users who have been inactive and not yet notified."""
user_deletion_config = apps.get_app_config('user_deletion')
threshold = timezone.now() - relativedelta(
months=user_deletion_config.MONTH_NOTIFICATION,
)
return self.filter(last_login__lte=threshold, notified=False)
def users_to_delete(self):
"""Finds all users who have been inactive and were notified."""
user_deletion_config = apps.get_app_config('user_deletion')
threshold = timezone.now() - relativedelta(
months=user_deletion_config.MONTH_DELETION,
)
return self.filter(last_login__lte=threshold, notified=True)
|
Put import back on top
|
Put import back on top
|
Python
|
bsd-2-clause
|
incuna/django-user-deletion
|
from dateutil.relativedelta import relativedelta
from django.utils import timezone
class UserDeletionManagerMixin:
def users_to_notify(self):
"""Finds all users who have been inactive and not yet notified."""
from django.apps import apps
user_deletion_config = apps.get_app_config('user_deletion')
threshold = timezone.now() - relativedelta(
months=user_deletion_config.MONTH_NOTIFICATION,
)
return self.filter(last_login__lte=threshold, notified=False)
def users_to_delete(self):
"""Finds all users who have been inactive and were notified."""
from django.apps import apps
user_deletion_config = apps.get_app_config('user_deletion')
threshold = timezone.now() - relativedelta(
months=user_deletion_config.MONTH_DELETION,
)
return self.filter(last_login__lte=threshold, notified=True)
Put import back on top
|
from dateutil.relativedelta import relativedelta
from django.apps import apps
from django.utils import timezone
class UserDeletionManagerMixin:
def users_to_notify(self):
"""Finds all users who have been inactive and not yet notified."""
user_deletion_config = apps.get_app_config('user_deletion')
threshold = timezone.now() - relativedelta(
months=user_deletion_config.MONTH_NOTIFICATION,
)
return self.filter(last_login__lte=threshold, notified=False)
def users_to_delete(self):
"""Finds all users who have been inactive and were notified."""
user_deletion_config = apps.get_app_config('user_deletion')
threshold = timezone.now() - relativedelta(
months=user_deletion_config.MONTH_DELETION,
)
return self.filter(last_login__lte=threshold, notified=True)
|
<commit_before>from dateutil.relativedelta import relativedelta
from django.utils import timezone
class UserDeletionManagerMixin:
def users_to_notify(self):
"""Finds all users who have been inactive and not yet notified."""
from django.apps import apps
user_deletion_config = apps.get_app_config('user_deletion')
threshold = timezone.now() - relativedelta(
months=user_deletion_config.MONTH_NOTIFICATION,
)
return self.filter(last_login__lte=threshold, notified=False)
def users_to_delete(self):
"""Finds all users who have been inactive and were notified."""
from django.apps import apps
user_deletion_config = apps.get_app_config('user_deletion')
threshold = timezone.now() - relativedelta(
months=user_deletion_config.MONTH_DELETION,
)
return self.filter(last_login__lte=threshold, notified=True)
<commit_msg>Put import back on top<commit_after>
|
from dateutil.relativedelta import relativedelta
from django.apps import apps
from django.utils import timezone
class UserDeletionManagerMixin:
def users_to_notify(self):
"""Finds all users who have been inactive and not yet notified."""
user_deletion_config = apps.get_app_config('user_deletion')
threshold = timezone.now() - relativedelta(
months=user_deletion_config.MONTH_NOTIFICATION,
)
return self.filter(last_login__lte=threshold, notified=False)
def users_to_delete(self):
"""Finds all users who have been inactive and were notified."""
user_deletion_config = apps.get_app_config('user_deletion')
threshold = timezone.now() - relativedelta(
months=user_deletion_config.MONTH_DELETION,
)
return self.filter(last_login__lte=threshold, notified=True)
|
from dateutil.relativedelta import relativedelta
from django.utils import timezone
class UserDeletionManagerMixin:
def users_to_notify(self):
"""Finds all users who have been inactive and not yet notified."""
from django.apps import apps
user_deletion_config = apps.get_app_config('user_deletion')
threshold = timezone.now() - relativedelta(
months=user_deletion_config.MONTH_NOTIFICATION,
)
return self.filter(last_login__lte=threshold, notified=False)
def users_to_delete(self):
"""Finds all users who have been inactive and were notified."""
from django.apps import apps
user_deletion_config = apps.get_app_config('user_deletion')
threshold = timezone.now() - relativedelta(
months=user_deletion_config.MONTH_DELETION,
)
return self.filter(last_login__lte=threshold, notified=True)
Put import back on topfrom dateutil.relativedelta import relativedelta
from django.apps import apps
from django.utils import timezone
class UserDeletionManagerMixin:
def users_to_notify(self):
"""Finds all users who have been inactive and not yet notified."""
user_deletion_config = apps.get_app_config('user_deletion')
threshold = timezone.now() - relativedelta(
months=user_deletion_config.MONTH_NOTIFICATION,
)
return self.filter(last_login__lte=threshold, notified=False)
def users_to_delete(self):
"""Finds all users who have been inactive and were notified."""
user_deletion_config = apps.get_app_config('user_deletion')
threshold = timezone.now() - relativedelta(
months=user_deletion_config.MONTH_DELETION,
)
return self.filter(last_login__lte=threshold, notified=True)
|
<commit_before>from dateutil.relativedelta import relativedelta
from django.utils import timezone
class UserDeletionManagerMixin:
def users_to_notify(self):
"""Finds all users who have been inactive and not yet notified."""
from django.apps import apps
user_deletion_config = apps.get_app_config('user_deletion')
threshold = timezone.now() - relativedelta(
months=user_deletion_config.MONTH_NOTIFICATION,
)
return self.filter(last_login__lte=threshold, notified=False)
def users_to_delete(self):
"""Finds all users who have been inactive and were notified."""
from django.apps import apps
user_deletion_config = apps.get_app_config('user_deletion')
threshold = timezone.now() - relativedelta(
months=user_deletion_config.MONTH_DELETION,
)
return self.filter(last_login__lte=threshold, notified=True)
<commit_msg>Put import back on top<commit_after>from dateutil.relativedelta import relativedelta
from django.apps import apps
from django.utils import timezone
class UserDeletionManagerMixin:
def users_to_notify(self):
"""Finds all users who have been inactive and not yet notified."""
user_deletion_config = apps.get_app_config('user_deletion')
threshold = timezone.now() - relativedelta(
months=user_deletion_config.MONTH_NOTIFICATION,
)
return self.filter(last_login__lte=threshold, notified=False)
def users_to_delete(self):
"""Finds all users who have been inactive and were notified."""
user_deletion_config = apps.get_app_config('user_deletion')
threshold = timezone.now() - relativedelta(
months=user_deletion_config.MONTH_DELETION,
)
return self.filter(last_login__lte=threshold, notified=True)
|
6a767780253ef981e78b00bb9937e9aaa0f9d1b8
|
motobot/core_plugins/network_handlers.py
|
motobot/core_plugins/network_handlers.py
|
from motobot import hook
from time import sleep
@hook('PING')
def handle_ping(bot, context, message):
""" Handle the server's pings. """
bot.send('PONG :' + message.params[-1])
@hook('439')
def handle_wait(bot, context, message):
""" Handles too fast for server message and waits 1 second. """
bot.identified = False
sleep(1)
@hook('NOTICE')
def handle_identification(bot, context, message):
""" Use the notice message to identify and register to the server. """
if not bot.identified:
bot.send('USER MotoBot localhost localhost MotoBot')
bot.send('NICK ' + bot.nick)
bot.identified = True
@hook('002')
def handle_nickserv_identification(bot, context, message):
""" At server welcome message 004 identify to nickserv and join channels. """
if bot.nickserv_password is not None:
bot.send('PRIVMSG nickserv :identify ' + bot.nickserv_password)
@hook('ERROR')
def handle_error(bot, context, message):
""" Handle an error message from the server. """
bot.connected = bot.identified = False
|
from motobot import hook
from time import sleep
@hook('PING')
def handle_ping(bot, context, message):
""" Handle the server's pings. """
bot.send('PONG :' + message.params[-1])
@hook('439')
def handle_wait(bot, context, message):
""" Handles too fast for server message and waits 1 second. """
bot.identified = False
sleep(1)
@hook('NOTICE')
def handle_identification(bot, context, message):
""" Use the notice message to identify and register to the server. """
if not bot.identified:
bot.send('USER MotoBot localhost localhost MotoBot')
bot.send('NICK ' + bot.nick)
bot.identified = True
@hook('002')
def handle_nickserv_identification(bot, context, message):
""" At server welcome message 004 identify to nickserv and join channels. """
if bot.nickserv_password is not None:
bot.send('PRIVMSG nickserv :identify ' + bot.nickserv_password)
sleep(1)
@hook('ERROR')
def handle_error(bot, context, message):
""" Handle an error message from the server. """
bot.connected = bot.identified = False
|
Add sleep after nickserv identify
|
Add sleep after nickserv identify
|
Python
|
mit
|
Motoko11/MotoBot
|
from motobot import hook
from time import sleep
@hook('PING')
def handle_ping(bot, context, message):
""" Handle the server's pings. """
bot.send('PONG :' + message.params[-1])
@hook('439')
def handle_wait(bot, context, message):
""" Handles too fast for server message and waits 1 second. """
bot.identified = False
sleep(1)
@hook('NOTICE')
def handle_identification(bot, context, message):
""" Use the notice message to identify and register to the server. """
if not bot.identified:
bot.send('USER MotoBot localhost localhost MotoBot')
bot.send('NICK ' + bot.nick)
bot.identified = True
@hook('002')
def handle_nickserv_identification(bot, context, message):
""" At server welcome message 004 identify to nickserv and join channels. """
if bot.nickserv_password is not None:
bot.send('PRIVMSG nickserv :identify ' + bot.nickserv_password)
@hook('ERROR')
def handle_error(bot, context, message):
""" Handle an error message from the server. """
bot.connected = bot.identified = False
Add sleep after nickserv identify
|
from motobot import hook
from time import sleep
@hook('PING')
def handle_ping(bot, context, message):
""" Handle the server's pings. """
bot.send('PONG :' + message.params[-1])
@hook('439')
def handle_wait(bot, context, message):
""" Handles too fast for server message and waits 1 second. """
bot.identified = False
sleep(1)
@hook('NOTICE')
def handle_identification(bot, context, message):
""" Use the notice message to identify and register to the server. """
if not bot.identified:
bot.send('USER MotoBot localhost localhost MotoBot')
bot.send('NICK ' + bot.nick)
bot.identified = True
@hook('002')
def handle_nickserv_identification(bot, context, message):
""" At server welcome message 004 identify to nickserv and join channels. """
if bot.nickserv_password is not None:
bot.send('PRIVMSG nickserv :identify ' + bot.nickserv_password)
sleep(1)
@hook('ERROR')
def handle_error(bot, context, message):
""" Handle an error message from the server. """
bot.connected = bot.identified = False
|
<commit_before>from motobot import hook
from time import sleep
@hook('PING')
def handle_ping(bot, context, message):
""" Handle the server's pings. """
bot.send('PONG :' + message.params[-1])
@hook('439')
def handle_wait(bot, context, message):
""" Handles too fast for server message and waits 1 second. """
bot.identified = False
sleep(1)
@hook('NOTICE')
def handle_identification(bot, context, message):
""" Use the notice message to identify and register to the server. """
if not bot.identified:
bot.send('USER MotoBot localhost localhost MotoBot')
bot.send('NICK ' + bot.nick)
bot.identified = True
@hook('002')
def handle_nickserv_identification(bot, context, message):
""" At server welcome message 004 identify to nickserv and join channels. """
if bot.nickserv_password is not None:
bot.send('PRIVMSG nickserv :identify ' + bot.nickserv_password)
@hook('ERROR')
def handle_error(bot, context, message):
""" Handle an error message from the server. """
bot.connected = bot.identified = False
<commit_msg>Add sleep after nickserv identify<commit_after>
|
from motobot import hook
from time import sleep
@hook('PING')
def handle_ping(bot, context, message):
""" Handle the server's pings. """
bot.send('PONG :' + message.params[-1])
@hook('439')
def handle_wait(bot, context, message):
""" Handles too fast for server message and waits 1 second. """
bot.identified = False
sleep(1)
@hook('NOTICE')
def handle_identification(bot, context, message):
""" Use the notice message to identify and register to the server. """
if not bot.identified:
bot.send('USER MotoBot localhost localhost MotoBot')
bot.send('NICK ' + bot.nick)
bot.identified = True
@hook('002')
def handle_nickserv_identification(bot, context, message):
""" At server welcome message 004 identify to nickserv and join channels. """
if bot.nickserv_password is not None:
bot.send('PRIVMSG nickserv :identify ' + bot.nickserv_password)
sleep(1)
@hook('ERROR')
def handle_error(bot, context, message):
""" Handle an error message from the server. """
bot.connected = bot.identified = False
|
from motobot import hook
from time import sleep
@hook('PING')
def handle_ping(bot, context, message):
""" Handle the server's pings. """
bot.send('PONG :' + message.params[-1])
@hook('439')
def handle_wait(bot, context, message):
""" Handles too fast for server message and waits 1 second. """
bot.identified = False
sleep(1)
@hook('NOTICE')
def handle_identification(bot, context, message):
""" Use the notice message to identify and register to the server. """
if not bot.identified:
bot.send('USER MotoBot localhost localhost MotoBot')
bot.send('NICK ' + bot.nick)
bot.identified = True
@hook('002')
def handle_nickserv_identification(bot, context, message):
""" At server welcome message 004 identify to nickserv and join channels. """
if bot.nickserv_password is not None:
bot.send('PRIVMSG nickserv :identify ' + bot.nickserv_password)
@hook('ERROR')
def handle_error(bot, context, message):
""" Handle an error message from the server. """
bot.connected = bot.identified = False
Add sleep after nickserv identifyfrom motobot import hook
from time import sleep
@hook('PING')
def handle_ping(bot, context, message):
""" Handle the server's pings. """
bot.send('PONG :' + message.params[-1])
@hook('439')
def handle_wait(bot, context, message):
""" Handles too fast for server message and waits 1 second. """
bot.identified = False
sleep(1)
@hook('NOTICE')
def handle_identification(bot, context, message):
""" Use the notice message to identify and register to the server. """
if not bot.identified:
bot.send('USER MotoBot localhost localhost MotoBot')
bot.send('NICK ' + bot.nick)
bot.identified = True
@hook('002')
def handle_nickserv_identification(bot, context, message):
""" At server welcome message 004 identify to nickserv and join channels. """
if bot.nickserv_password is not None:
bot.send('PRIVMSG nickserv :identify ' + bot.nickserv_password)
sleep(1)
@hook('ERROR')
def handle_error(bot, context, message):
""" Handle an error message from the server. """
bot.connected = bot.identified = False
|
<commit_before>from motobot import hook
from time import sleep
@hook('PING')
def handle_ping(bot, context, message):
""" Handle the server's pings. """
bot.send('PONG :' + message.params[-1])
@hook('439')
def handle_wait(bot, context, message):
""" Handles too fast for server message and waits 1 second. """
bot.identified = False
sleep(1)
@hook('NOTICE')
def handle_identification(bot, context, message):
""" Use the notice message to identify and register to the server. """
if not bot.identified:
bot.send('USER MotoBot localhost localhost MotoBot')
bot.send('NICK ' + bot.nick)
bot.identified = True
@hook('002')
def handle_nickserv_identification(bot, context, message):
""" At server welcome message 004 identify to nickserv and join channels. """
if bot.nickserv_password is not None:
bot.send('PRIVMSG nickserv :identify ' + bot.nickserv_password)
@hook('ERROR')
def handle_error(bot, context, message):
""" Handle an error message from the server. """
bot.connected = bot.identified = False
<commit_msg>Add sleep after nickserv identify<commit_after>from motobot import hook
from time import sleep
@hook('PING')
def handle_ping(bot, context, message):
""" Handle the server's pings. """
bot.send('PONG :' + message.params[-1])
@hook('439')
def handle_wait(bot, context, message):
""" Handles too fast for server message and waits 1 second. """
bot.identified = False
sleep(1)
@hook('NOTICE')
def handle_identification(bot, context, message):
""" Use the notice message to identify and register to the server. """
if not bot.identified:
bot.send('USER MotoBot localhost localhost MotoBot')
bot.send('NICK ' + bot.nick)
bot.identified = True
@hook('002')
def handle_nickserv_identification(bot, context, message):
""" At server welcome message 004 identify to nickserv and join channels. """
if bot.nickserv_password is not None:
bot.send('PRIVMSG nickserv :identify ' + bot.nickserv_password)
sleep(1)
@hook('ERROR')
def handle_error(bot, context, message):
""" Handle an error message from the server. """
bot.connected = bot.identified = False
|
c2d3c2c471dfb504626509a34256eb2d9898cfa2
|
rest_framework_nested/viewsets.py
|
rest_framework_nested/viewsets.py
|
class NestedViewSetMixin(object):
def get_queryset(self):
"""
Filter the `QuerySet` based on its parents as defined in the
`serializer_class.parent_lookup_kwargs`.
"""
queryset = super(NestedViewSetMixin, self).get_queryset()
if hasattr(self.serializer_class, 'parent_lookup_kwargs'):
orm_filters = {}
for query_param, field_name in self.serializer_class.parent_lookup_kwargs.items():
orm_filters[field_name] = self.kwargs[query_param]
return queryset.filter(**orm_filters)
return queryset
|
class NestedViewSetMixin(object):
def get_queryset(self):
"""
Filter the `QuerySet` based on its parents as defined in the
`serializer_class.parent_lookup_kwargs`.
"""
queryset = super(NestedViewSetMixin, self).get_queryset()
serializer_class = self.get_serializer_class()
if hasattr(serializer_class, 'parent_lookup_kwargs'):
orm_filters = {}
for query_param, field_name in serializer_class.parent_lookup_kwargs.items():
orm_filters[field_name] = self.kwargs[query_param]
return queryset.filter(**orm_filters)
return queryset
|
Fix to use get_serializer_class method instead of serializer_class
|
Fix to use get_serializer_class method instead of serializer_class
|
Python
|
apache-2.0
|
alanjds/drf-nested-routers
|
class NestedViewSetMixin(object):
def get_queryset(self):
"""
Filter the `QuerySet` based on its parents as defined in the
`serializer_class.parent_lookup_kwargs`.
"""
queryset = super(NestedViewSetMixin, self).get_queryset()
if hasattr(self.serializer_class, 'parent_lookup_kwargs'):
orm_filters = {}
for query_param, field_name in self.serializer_class.parent_lookup_kwargs.items():
orm_filters[field_name] = self.kwargs[query_param]
return queryset.filter(**orm_filters)
return queryset
Fix to use get_serializer_class method instead of serializer_class
|
class NestedViewSetMixin(object):
def get_queryset(self):
"""
Filter the `QuerySet` based on its parents as defined in the
`serializer_class.parent_lookup_kwargs`.
"""
queryset = super(NestedViewSetMixin, self).get_queryset()
serializer_class = self.get_serializer_class()
if hasattr(serializer_class, 'parent_lookup_kwargs'):
orm_filters = {}
for query_param, field_name in serializer_class.parent_lookup_kwargs.items():
orm_filters[field_name] = self.kwargs[query_param]
return queryset.filter(**orm_filters)
return queryset
|
<commit_before>class NestedViewSetMixin(object):
def get_queryset(self):
"""
Filter the `QuerySet` based on its parents as defined in the
`serializer_class.parent_lookup_kwargs`.
"""
queryset = super(NestedViewSetMixin, self).get_queryset()
if hasattr(self.serializer_class, 'parent_lookup_kwargs'):
orm_filters = {}
for query_param, field_name in self.serializer_class.parent_lookup_kwargs.items():
orm_filters[field_name] = self.kwargs[query_param]
return queryset.filter(**orm_filters)
return queryset
<commit_msg>Fix to use get_serializer_class method instead of serializer_class<commit_after>
|
class NestedViewSetMixin(object):
def get_queryset(self):
"""
Filter the `QuerySet` based on its parents as defined in the
`serializer_class.parent_lookup_kwargs`.
"""
queryset = super(NestedViewSetMixin, self).get_queryset()
serializer_class = self.get_serializer_class()
if hasattr(serializer_class, 'parent_lookup_kwargs'):
orm_filters = {}
for query_param, field_name in serializer_class.parent_lookup_kwargs.items():
orm_filters[field_name] = self.kwargs[query_param]
return queryset.filter(**orm_filters)
return queryset
|
class NestedViewSetMixin(object):
def get_queryset(self):
"""
Filter the `QuerySet` based on its parents as defined in the
`serializer_class.parent_lookup_kwargs`.
"""
queryset = super(NestedViewSetMixin, self).get_queryset()
if hasattr(self.serializer_class, 'parent_lookup_kwargs'):
orm_filters = {}
for query_param, field_name in self.serializer_class.parent_lookup_kwargs.items():
orm_filters[field_name] = self.kwargs[query_param]
return queryset.filter(**orm_filters)
return queryset
Fix to use get_serializer_class method instead of serializer_classclass NestedViewSetMixin(object):
def get_queryset(self):
"""
Filter the `QuerySet` based on its parents as defined in the
`serializer_class.parent_lookup_kwargs`.
"""
queryset = super(NestedViewSetMixin, self).get_queryset()
serializer_class = self.get_serializer_class()
if hasattr(serializer_class, 'parent_lookup_kwargs'):
orm_filters = {}
for query_param, field_name in serializer_class.parent_lookup_kwargs.items():
orm_filters[field_name] = self.kwargs[query_param]
return queryset.filter(**orm_filters)
return queryset
|
<commit_before>class NestedViewSetMixin(object):
def get_queryset(self):
"""
Filter the `QuerySet` based on its parents as defined in the
`serializer_class.parent_lookup_kwargs`.
"""
queryset = super(NestedViewSetMixin, self).get_queryset()
if hasattr(self.serializer_class, 'parent_lookup_kwargs'):
orm_filters = {}
for query_param, field_name in self.serializer_class.parent_lookup_kwargs.items():
orm_filters[field_name] = self.kwargs[query_param]
return queryset.filter(**orm_filters)
return queryset
<commit_msg>Fix to use get_serializer_class method instead of serializer_class<commit_after>class NestedViewSetMixin(object):
def get_queryset(self):
"""
Filter the `QuerySet` based on its parents as defined in the
`serializer_class.parent_lookup_kwargs`.
"""
queryset = super(NestedViewSetMixin, self).get_queryset()
serializer_class = self.get_serializer_class()
if hasattr(serializer_class, 'parent_lookup_kwargs'):
orm_filters = {}
for query_param, field_name in serializer_class.parent_lookup_kwargs.items():
orm_filters[field_name] = self.kwargs[query_param]
return queryset.filter(**orm_filters)
return queryset
|
3b7328dd7d9d235bf32b3cfb836b49e50b70be77
|
oz/plugins/redis_sessions/__init__.py
|
oz/plugins/redis_sessions/__init__.py
|
from __future__ import absolute_import, division, print_function, with_statement, unicode_literals
import os
import binascii
import hashlib
import oz.app
from .middleware import *
from .options import *
from .tests import *
def random_hex(length):
"""Generates a random hex string"""
return binascii.hexlify(os.urandom(length))[length:]
def password_hash(password, password_salt=None):
"""Hashes a specified password"""
password_salt = password_salt or oz.app.settings["session_salt"]
return u"sha256!%s" % hashlib.sha256(unicode(password_salt) + unicode(password)).hexdigest()
|
from __future__ import absolute_import, division, print_function, with_statement, unicode_literals
import os
import binascii
import hashlib
import oz.app
from .middleware import *
from .options import *
from .tests import *
def random_hex(length):
"""Generates a random hex string"""
return binascii.hexlify(os.urandom(length))[length:]
def password_hash(password, password_salt=None):
"""Hashes a specified password"""
password_salt = password_salt or oz.app.settings["session_salt"]
salted_password = "".join([unicode(password_salt), password])
return "sha256!%s" % unicode(hashlib.sha256(salted_password.encode("utf-8")).hexdigest())
|
Allow for non-ascii characters in password_hash
|
Allow for non-ascii characters in password_hash
|
Python
|
bsd-3-clause
|
dailymuse/oz,dailymuse/oz,dailymuse/oz
|
from __future__ import absolute_import, division, print_function, with_statement, unicode_literals
import os
import binascii
import hashlib
import oz.app
from .middleware import *
from .options import *
from .tests import *
def random_hex(length):
"""Generates a random hex string"""
return binascii.hexlify(os.urandom(length))[length:]
def password_hash(password, password_salt=None):
"""Hashes a specified password"""
password_salt = password_salt or oz.app.settings["session_salt"]
return u"sha256!%s" % hashlib.sha256(unicode(password_salt) + unicode(password)).hexdigest()
Allow for non-ascii characters in password_hash
|
from __future__ import absolute_import, division, print_function, with_statement, unicode_literals
import os
import binascii
import hashlib
import oz.app
from .middleware import *
from .options import *
from .tests import *
def random_hex(length):
"""Generates a random hex string"""
return binascii.hexlify(os.urandom(length))[length:]
def password_hash(password, password_salt=None):
"""Hashes a specified password"""
password_salt = password_salt or oz.app.settings["session_salt"]
salted_password = "".join([unicode(password_salt), password])
return "sha256!%s" % unicode(hashlib.sha256(salted_password.encode("utf-8")).hexdigest())
|
<commit_before>from __future__ import absolute_import, division, print_function, with_statement, unicode_literals
import os
import binascii
import hashlib
import oz.app
from .middleware import *
from .options import *
from .tests import *
def random_hex(length):
"""Generates a random hex string"""
return binascii.hexlify(os.urandom(length))[length:]
def password_hash(password, password_salt=None):
"""Hashes a specified password"""
password_salt = password_salt or oz.app.settings["session_salt"]
return u"sha256!%s" % hashlib.sha256(unicode(password_salt) + unicode(password)).hexdigest()
<commit_msg>Allow for non-ascii characters in password_hash<commit_after>
|
from __future__ import absolute_import, division, print_function, with_statement, unicode_literals
import os
import binascii
import hashlib
import oz.app
from .middleware import *
from .options import *
from .tests import *
def random_hex(length):
"""Generates a random hex string"""
return binascii.hexlify(os.urandom(length))[length:]
def password_hash(password, password_salt=None):
"""Hashes a specified password"""
password_salt = password_salt or oz.app.settings["session_salt"]
salted_password = "".join([unicode(password_salt), password])
return "sha256!%s" % unicode(hashlib.sha256(salted_password.encode("utf-8")).hexdigest())
|
from __future__ import absolute_import, division, print_function, with_statement, unicode_literals
import os
import binascii
import hashlib
import oz.app
from .middleware import *
from .options import *
from .tests import *
def random_hex(length):
"""Generates a random hex string"""
return binascii.hexlify(os.urandom(length))[length:]
def password_hash(password, password_salt=None):
"""Hashes a specified password"""
password_salt = password_salt or oz.app.settings["session_salt"]
return u"sha256!%s" % hashlib.sha256(unicode(password_salt) + unicode(password)).hexdigest()
Allow for non-ascii characters in password_hashfrom __future__ import absolute_import, division, print_function, with_statement, unicode_literals
import os
import binascii
import hashlib
import oz.app
from .middleware import *
from .options import *
from .tests import *
def random_hex(length):
"""Generates a random hex string"""
return binascii.hexlify(os.urandom(length))[length:]
def password_hash(password, password_salt=None):
"""Hashes a specified password"""
password_salt = password_salt or oz.app.settings["session_salt"]
salted_password = "".join([unicode(password_salt), password])
return "sha256!%s" % unicode(hashlib.sha256(salted_password.encode("utf-8")).hexdigest())
|
<commit_before>from __future__ import absolute_import, division, print_function, with_statement, unicode_literals
import os
import binascii
import hashlib
import oz.app
from .middleware import *
from .options import *
from .tests import *
def random_hex(length):
"""Generates a random hex string"""
return binascii.hexlify(os.urandom(length))[length:]
def password_hash(password, password_salt=None):
"""Hashes a specified password"""
password_salt = password_salt or oz.app.settings["session_salt"]
return u"sha256!%s" % hashlib.sha256(unicode(password_salt) + unicode(password)).hexdigest()
<commit_msg>Allow for non-ascii characters in password_hash<commit_after>from __future__ import absolute_import, division, print_function, with_statement, unicode_literals
import os
import binascii
import hashlib
import oz.app
from .middleware import *
from .options import *
from .tests import *
def random_hex(length):
"""Generates a random hex string"""
return binascii.hexlify(os.urandom(length))[length:]
def password_hash(password, password_salt=None):
"""Hashes a specified password"""
password_salt = password_salt or oz.app.settings["session_salt"]
salted_password = "".join([unicode(password_salt), password])
return "sha256!%s" % unicode(hashlib.sha256(salted_password.encode("utf-8")).hexdigest())
|
442f21bfde16f72d4480fa7fd9dea2eac741a857
|
src/analyses/views.py
|
src/analyses/views.py
|
from django.contrib import messages
from django.contrib.auth import get_user_model
from django.contrib.auth.mixins import LoginRequiredMixin
from django.utils.translation import ugettext_lazy as _
from django.views.generic import CreateView, TemplateView
from .forms import AbstractAnalysisCreateForm
from .pipelines import AVAILABLE_PIPELINES
User = get_user_model()
class SelectNewAnalysisTypeView(LoginRequiredMixin, TemplateView):
template_name = "analyses/new_analysis_by_type.html"
def get_context_data(self, **kwargs):
context = super().get_context_data(**kwargs)
context['available_pipelines'] = AVAILABLE_PIPELINES
return context
class AbstractAnalysisFormView(LoginRequiredMixin, CreateView):
form_class = AbstractAnalysisCreateForm
template_name = None
analysis_type = 'AbstractAnalysis'
analysis_description = ''
analysis_create_url = None
def get_form_kwargs(self):
"""Pass request object for form creation"""
kwargs = super().get_form_kwargs()
kwargs['request'] = self.request
return kwargs
def form_valid(self, form):
response = super().form_valid(form)
messages.add_message(
self.request, messages.INFO,
_('You just created a %(analysis_type)s analysis!') % {
'analysis_type': self.analysis_type
}
)
return response
|
from django.contrib import messages
from django.contrib.auth import get_user_model
from django.contrib.auth.mixins import LoginRequiredMixin
from django.utils.translation import ugettext_lazy as _
from django.views.generic import CreateView, TemplateView
from .forms import AbstractAnalysisCreateForm
from .pipelines import AVAILABLE_PIPELINES
User = get_user_model()
class SelectNewAnalysisTypeView(LoginRequiredMixin, TemplateView):
template_name = "analyses/new_analysis_by_type.html"
def get_context_data(self, **kwargs):
context = super().get_context_data(**kwargs)
context['available_pipelines'] = AVAILABLE_PIPELINES
return context
class AbstractAnalysisFormView(LoginRequiredMixin, CreateView):
form_class = AbstractAnalysisCreateForm
template_name = None
analysis_type = 'AbstractAnalysis'
analysis_description = ''
analysis_create_url = None
def get_form_kwargs(self):
"""Pass request object for form creation"""
kwargs = super().get_form_kwargs()
kwargs['request'] = self.request
return kwargs
def form_valid(self, form):
response = super().form_valid(form)
messages.add_message(
self.request, messages.INFO,
_(
'You just created a %(analysis_type)s analysis! '
'View its detail <a href="%(analysis_detail_url)s">here</a>.'
) % {
'analysis_type': self.analysis_type,
'analysis_detail_url': self.object.get_absolute_url(),
},
extra_tags='safe',
)
return response
|
Include analysis detail view URL in message
|
Include analysis detail view URL in message
|
Python
|
mit
|
ccwang002/biocloud-server-kai,ccwang002/biocloud-server-kai,ccwang002/biocloud-server-kai
|
from django.contrib import messages
from django.contrib.auth import get_user_model
from django.contrib.auth.mixins import LoginRequiredMixin
from django.utils.translation import ugettext_lazy as _
from django.views.generic import CreateView, TemplateView
from .forms import AbstractAnalysisCreateForm
from .pipelines import AVAILABLE_PIPELINES
User = get_user_model()
class SelectNewAnalysisTypeView(LoginRequiredMixin, TemplateView):
template_name = "analyses/new_analysis_by_type.html"
def get_context_data(self, **kwargs):
context = super().get_context_data(**kwargs)
context['available_pipelines'] = AVAILABLE_PIPELINES
return context
class AbstractAnalysisFormView(LoginRequiredMixin, CreateView):
form_class = AbstractAnalysisCreateForm
template_name = None
analysis_type = 'AbstractAnalysis'
analysis_description = ''
analysis_create_url = None
def get_form_kwargs(self):
"""Pass request object for form creation"""
kwargs = super().get_form_kwargs()
kwargs['request'] = self.request
return kwargs
def form_valid(self, form):
response = super().form_valid(form)
messages.add_message(
self.request, messages.INFO,
_('You just created a %(analysis_type)s analysis!') % {
'analysis_type': self.analysis_type
}
)
return response
Include analysis detail view URL in message
|
from django.contrib import messages
from django.contrib.auth import get_user_model
from django.contrib.auth.mixins import LoginRequiredMixin
from django.utils.translation import ugettext_lazy as _
from django.views.generic import CreateView, TemplateView
from .forms import AbstractAnalysisCreateForm
from .pipelines import AVAILABLE_PIPELINES
User = get_user_model()
class SelectNewAnalysisTypeView(LoginRequiredMixin, TemplateView):
template_name = "analyses/new_analysis_by_type.html"
def get_context_data(self, **kwargs):
context = super().get_context_data(**kwargs)
context['available_pipelines'] = AVAILABLE_PIPELINES
return context
class AbstractAnalysisFormView(LoginRequiredMixin, CreateView):
form_class = AbstractAnalysisCreateForm
template_name = None
analysis_type = 'AbstractAnalysis'
analysis_description = ''
analysis_create_url = None
def get_form_kwargs(self):
"""Pass request object for form creation"""
kwargs = super().get_form_kwargs()
kwargs['request'] = self.request
return kwargs
def form_valid(self, form):
response = super().form_valid(form)
messages.add_message(
self.request, messages.INFO,
_(
'You just created a %(analysis_type)s analysis! '
'View its detail <a href="%(analysis_detail_url)s">here</a>.'
) % {
'analysis_type': self.analysis_type,
'analysis_detail_url': self.object.get_absolute_url(),
},
extra_tags='safe',
)
return response
|
<commit_before>from django.contrib import messages
from django.contrib.auth import get_user_model
from django.contrib.auth.mixins import LoginRequiredMixin
from django.utils.translation import ugettext_lazy as _
from django.views.generic import CreateView, TemplateView
from .forms import AbstractAnalysisCreateForm
from .pipelines import AVAILABLE_PIPELINES
User = get_user_model()
class SelectNewAnalysisTypeView(LoginRequiredMixin, TemplateView):
template_name = "analyses/new_analysis_by_type.html"
def get_context_data(self, **kwargs):
context = super().get_context_data(**kwargs)
context['available_pipelines'] = AVAILABLE_PIPELINES
return context
class AbstractAnalysisFormView(LoginRequiredMixin, CreateView):
form_class = AbstractAnalysisCreateForm
template_name = None
analysis_type = 'AbstractAnalysis'
analysis_description = ''
analysis_create_url = None
def get_form_kwargs(self):
"""Pass request object for form creation"""
kwargs = super().get_form_kwargs()
kwargs['request'] = self.request
return kwargs
def form_valid(self, form):
response = super().form_valid(form)
messages.add_message(
self.request, messages.INFO,
_('You just created a %(analysis_type)s analysis!') % {
'analysis_type': self.analysis_type
}
)
return response
<commit_msg>Include analysis detail view URL in message<commit_after>
|
from django.contrib import messages
from django.contrib.auth import get_user_model
from django.contrib.auth.mixins import LoginRequiredMixin
from django.utils.translation import ugettext_lazy as _
from django.views.generic import CreateView, TemplateView
from .forms import AbstractAnalysisCreateForm
from .pipelines import AVAILABLE_PIPELINES
User = get_user_model()
class SelectNewAnalysisTypeView(LoginRequiredMixin, TemplateView):
template_name = "analyses/new_analysis_by_type.html"
def get_context_data(self, **kwargs):
context = super().get_context_data(**kwargs)
context['available_pipelines'] = AVAILABLE_PIPELINES
return context
class AbstractAnalysisFormView(LoginRequiredMixin, CreateView):
form_class = AbstractAnalysisCreateForm
template_name = None
analysis_type = 'AbstractAnalysis'
analysis_description = ''
analysis_create_url = None
def get_form_kwargs(self):
"""Pass request object for form creation"""
kwargs = super().get_form_kwargs()
kwargs['request'] = self.request
return kwargs
def form_valid(self, form):
response = super().form_valid(form)
messages.add_message(
self.request, messages.INFO,
_(
'You just created a %(analysis_type)s analysis! '
'View its detail <a href="%(analysis_detail_url)s">here</a>.'
) % {
'analysis_type': self.analysis_type,
'analysis_detail_url': self.object.get_absolute_url(),
},
extra_tags='safe',
)
return response
|
from django.contrib import messages
from django.contrib.auth import get_user_model
from django.contrib.auth.mixins import LoginRequiredMixin
from django.utils.translation import ugettext_lazy as _
from django.views.generic import CreateView, TemplateView
from .forms import AbstractAnalysisCreateForm
from .pipelines import AVAILABLE_PIPELINES
User = get_user_model()
class SelectNewAnalysisTypeView(LoginRequiredMixin, TemplateView):
template_name = "analyses/new_analysis_by_type.html"
def get_context_data(self, **kwargs):
context = super().get_context_data(**kwargs)
context['available_pipelines'] = AVAILABLE_PIPELINES
return context
class AbstractAnalysisFormView(LoginRequiredMixin, CreateView):
form_class = AbstractAnalysisCreateForm
template_name = None
analysis_type = 'AbstractAnalysis'
analysis_description = ''
analysis_create_url = None
def get_form_kwargs(self):
"""Pass request object for form creation"""
kwargs = super().get_form_kwargs()
kwargs['request'] = self.request
return kwargs
def form_valid(self, form):
response = super().form_valid(form)
messages.add_message(
self.request, messages.INFO,
_('You just created a %(analysis_type)s analysis!') % {
'analysis_type': self.analysis_type
}
)
return response
Include analysis detail view URL in messagefrom django.contrib import messages
from django.contrib.auth import get_user_model
from django.contrib.auth.mixins import LoginRequiredMixin
from django.utils.translation import ugettext_lazy as _
from django.views.generic import CreateView, TemplateView
from .forms import AbstractAnalysisCreateForm
from .pipelines import AVAILABLE_PIPELINES
User = get_user_model()
class SelectNewAnalysisTypeView(LoginRequiredMixin, TemplateView):
template_name = "analyses/new_analysis_by_type.html"
def get_context_data(self, **kwargs):
context = super().get_context_data(**kwargs)
context['available_pipelines'] = AVAILABLE_PIPELINES
return context
class AbstractAnalysisFormView(LoginRequiredMixin, CreateView):
form_class = AbstractAnalysisCreateForm
template_name = None
analysis_type = 'AbstractAnalysis'
analysis_description = ''
analysis_create_url = None
def get_form_kwargs(self):
"""Pass request object for form creation"""
kwargs = super().get_form_kwargs()
kwargs['request'] = self.request
return kwargs
def form_valid(self, form):
response = super().form_valid(form)
messages.add_message(
self.request, messages.INFO,
_(
'You just created a %(analysis_type)s analysis! '
'View its detail <a href="%(analysis_detail_url)s">here</a>.'
) % {
'analysis_type': self.analysis_type,
'analysis_detail_url': self.object.get_absolute_url(),
},
extra_tags='safe',
)
return response
|
<commit_before>from django.contrib import messages
from django.contrib.auth import get_user_model
from django.contrib.auth.mixins import LoginRequiredMixin
from django.utils.translation import ugettext_lazy as _
from django.views.generic import CreateView, TemplateView
from .forms import AbstractAnalysisCreateForm
from .pipelines import AVAILABLE_PIPELINES
User = get_user_model()
class SelectNewAnalysisTypeView(LoginRequiredMixin, TemplateView):
template_name = "analyses/new_analysis_by_type.html"
def get_context_data(self, **kwargs):
context = super().get_context_data(**kwargs)
context['available_pipelines'] = AVAILABLE_PIPELINES
return context
class AbstractAnalysisFormView(LoginRequiredMixin, CreateView):
form_class = AbstractAnalysisCreateForm
template_name = None
analysis_type = 'AbstractAnalysis'
analysis_description = ''
analysis_create_url = None
def get_form_kwargs(self):
"""Pass request object for form creation"""
kwargs = super().get_form_kwargs()
kwargs['request'] = self.request
return kwargs
def form_valid(self, form):
response = super().form_valid(form)
messages.add_message(
self.request, messages.INFO,
_('You just created a %(analysis_type)s analysis!') % {
'analysis_type': self.analysis_type
}
)
return response
<commit_msg>Include analysis detail view URL in message<commit_after>from django.contrib import messages
from django.contrib.auth import get_user_model
from django.contrib.auth.mixins import LoginRequiredMixin
from django.utils.translation import ugettext_lazy as _
from django.views.generic import CreateView, TemplateView
from .forms import AbstractAnalysisCreateForm
from .pipelines import AVAILABLE_PIPELINES
User = get_user_model()
class SelectNewAnalysisTypeView(LoginRequiredMixin, TemplateView):
template_name = "analyses/new_analysis_by_type.html"
def get_context_data(self, **kwargs):
context = super().get_context_data(**kwargs)
context['available_pipelines'] = AVAILABLE_PIPELINES
return context
class AbstractAnalysisFormView(LoginRequiredMixin, CreateView):
form_class = AbstractAnalysisCreateForm
template_name = None
analysis_type = 'AbstractAnalysis'
analysis_description = ''
analysis_create_url = None
def get_form_kwargs(self):
"""Pass request object for form creation"""
kwargs = super().get_form_kwargs()
kwargs['request'] = self.request
return kwargs
def form_valid(self, form):
response = super().form_valid(form)
messages.add_message(
self.request, messages.INFO,
_(
'You just created a %(analysis_type)s analysis! '
'View its detail <a href="%(analysis_detail_url)s">here</a>.'
) % {
'analysis_type': self.analysis_type,
'analysis_detail_url': self.object.get_absolute_url(),
},
extra_tags='safe',
)
return response
|
2bc95d90db15160f9c4869c03f9dadb6cd8d56fa
|
seleniumbase/config/proxy_list.py
|
seleniumbase/config/proxy_list.py
|
"""
Proxy Server "Phone Book".
Simplify running browser tests through a proxy server
by adding your frequently-used proxies here.
Now you can do something like this on the command line:
"pytest SOME_TEST.py --proxy=proxy1"
Format of PROXY_LIST server entries:
* "ip_address:port" OR "username:password@ip_address:port"
* "server:port" OR "username:password@server:port"
(Do NOT include the http:// or https:// in your proxy string!)
Example proxies in PROXY_LIST below are not guaranteed to be active or secure.
If you don't already have a proxy server to connect to,
you can try finding one from one of following sites:
* https://www.us-proxy.org/
"""
PROXY_LIST = {
"example1": "159.122.164.163:8080", # (Example) - set your own proxy here
"proxy1": None,
"proxy2": None,
"proxy3": None,
"proxy4": None,
"proxy5": None,
}
|
"""
Proxy Server "Phone Book".
Simplify running browser tests through a proxy server
by adding your frequently-used proxies here.
Now you can do something like this on the command line:
"pytest SOME_TEST.py --proxy=proxy1"
Format of PROXY_LIST server entries:
* "ip_address:port" OR "username:password@ip_address:port"
* "server:port" OR "username:password@server:port"
(Do NOT include the http:// or https:// in your proxy string!)
Example proxies in PROXY_LIST below are not guaranteed to be active or secure.
If you don't already have a proxy server to connect to,
you can try finding one from one of following sites:
* https://www.us-proxy.org/
"""
PROXY_LIST = {
"example1": "159.122.164.163:8080", # (Example) - set your own proxy here
"example2": "158.69.138.8:1080", # (Example) - set your own proxy here
"proxy1": None,
"proxy2": None,
"proxy3": None,
"proxy4": None,
"proxy5": None,
}
|
Add another proxy server example string
|
Add another proxy server example string
|
Python
|
mit
|
seleniumbase/SeleniumBase,seleniumbase/SeleniumBase,mdmintz/SeleniumBase,seleniumbase/SeleniumBase,seleniumbase/SeleniumBase,mdmintz/SeleniumBase,mdmintz/SeleniumBase,mdmintz/SeleniumBase
|
"""
Proxy Server "Phone Book".
Simplify running browser tests through a proxy server
by adding your frequently-used proxies here.
Now you can do something like this on the command line:
"pytest SOME_TEST.py --proxy=proxy1"
Format of PROXY_LIST server entries:
* "ip_address:port" OR "username:password@ip_address:port"
* "server:port" OR "username:password@server:port"
(Do NOT include the http:// or https:// in your proxy string!)
Example proxies in PROXY_LIST below are not guaranteed to be active or secure.
If you don't already have a proxy server to connect to,
you can try finding one from one of following sites:
* https://www.us-proxy.org/
"""
PROXY_LIST = {
"example1": "159.122.164.163:8080", # (Example) - set your own proxy here
"proxy1": None,
"proxy2": None,
"proxy3": None,
"proxy4": None,
"proxy5": None,
}
Add another proxy server example string
|
"""
Proxy Server "Phone Book".
Simplify running browser tests through a proxy server
by adding your frequently-used proxies here.
Now you can do something like this on the command line:
"pytest SOME_TEST.py --proxy=proxy1"
Format of PROXY_LIST server entries:
* "ip_address:port" OR "username:password@ip_address:port"
* "server:port" OR "username:password@server:port"
(Do NOT include the http:// or https:// in your proxy string!)
Example proxies in PROXY_LIST below are not guaranteed to be active or secure.
If you don't already have a proxy server to connect to,
you can try finding one from one of following sites:
* https://www.us-proxy.org/
"""
PROXY_LIST = {
"example1": "159.122.164.163:8080", # (Example) - set your own proxy here
"example2": "158.69.138.8:1080", # (Example) - set your own proxy here
"proxy1": None,
"proxy2": None,
"proxy3": None,
"proxy4": None,
"proxy5": None,
}
|
<commit_before>"""
Proxy Server "Phone Book".
Simplify running browser tests through a proxy server
by adding your frequently-used proxies here.
Now you can do something like this on the command line:
"pytest SOME_TEST.py --proxy=proxy1"
Format of PROXY_LIST server entries:
* "ip_address:port" OR "username:password@ip_address:port"
* "server:port" OR "username:password@server:port"
(Do NOT include the http:// or https:// in your proxy string!)
Example proxies in PROXY_LIST below are not guaranteed to be active or secure.
If you don't already have a proxy server to connect to,
you can try finding one from one of following sites:
* https://www.us-proxy.org/
"""
PROXY_LIST = {
"example1": "159.122.164.163:8080", # (Example) - set your own proxy here
"proxy1": None,
"proxy2": None,
"proxy3": None,
"proxy4": None,
"proxy5": None,
}
<commit_msg>Add another proxy server example string<commit_after>
|
"""
Proxy Server "Phone Book".
Simplify running browser tests through a proxy server
by adding your frequently-used proxies here.
Now you can do something like this on the command line:
"pytest SOME_TEST.py --proxy=proxy1"
Format of PROXY_LIST server entries:
* "ip_address:port" OR "username:password@ip_address:port"
* "server:port" OR "username:password@server:port"
(Do NOT include the http:// or https:// in your proxy string!)
Example proxies in PROXY_LIST below are not guaranteed to be active or secure.
If you don't already have a proxy server to connect to,
you can try finding one from one of following sites:
* https://www.us-proxy.org/
"""
PROXY_LIST = {
"example1": "159.122.164.163:8080", # (Example) - set your own proxy here
"example2": "158.69.138.8:1080", # (Example) - set your own proxy here
"proxy1": None,
"proxy2": None,
"proxy3": None,
"proxy4": None,
"proxy5": None,
}
|
"""
Proxy Server "Phone Book".
Simplify running browser tests through a proxy server
by adding your frequently-used proxies here.
Now you can do something like this on the command line:
"pytest SOME_TEST.py --proxy=proxy1"
Format of PROXY_LIST server entries:
* "ip_address:port" OR "username:password@ip_address:port"
* "server:port" OR "username:password@server:port"
(Do NOT include the http:// or https:// in your proxy string!)
Example proxies in PROXY_LIST below are not guaranteed to be active or secure.
If you don't already have a proxy server to connect to,
you can try finding one from one of following sites:
* https://www.us-proxy.org/
"""
PROXY_LIST = {
"example1": "159.122.164.163:8080", # (Example) - set your own proxy here
"proxy1": None,
"proxy2": None,
"proxy3": None,
"proxy4": None,
"proxy5": None,
}
Add another proxy server example string"""
Proxy Server "Phone Book".
Simplify running browser tests through a proxy server
by adding your frequently-used proxies here.
Now you can do something like this on the command line:
"pytest SOME_TEST.py --proxy=proxy1"
Format of PROXY_LIST server entries:
* "ip_address:port" OR "username:password@ip_address:port"
* "server:port" OR "username:password@server:port"
(Do NOT include the http:// or https:// in your proxy string!)
Example proxies in PROXY_LIST below are not guaranteed to be active or secure.
If you don't already have a proxy server to connect to,
you can try finding one from one of following sites:
* https://www.us-proxy.org/
"""
PROXY_LIST = {
"example1": "159.122.164.163:8080", # (Example) - set your own proxy here
"example2": "158.69.138.8:1080", # (Example) - set your own proxy here
"proxy1": None,
"proxy2": None,
"proxy3": None,
"proxy4": None,
"proxy5": None,
}
|
<commit_before>"""
Proxy Server "Phone Book".
Simplify running browser tests through a proxy server
by adding your frequently-used proxies here.
Now you can do something like this on the command line:
"pytest SOME_TEST.py --proxy=proxy1"
Format of PROXY_LIST server entries:
* "ip_address:port" OR "username:password@ip_address:port"
* "server:port" OR "username:password@server:port"
(Do NOT include the http:// or https:// in your proxy string!)
Example proxies in PROXY_LIST below are not guaranteed to be active or secure.
If you don't already have a proxy server to connect to,
you can try finding one from one of following sites:
* https://www.us-proxy.org/
"""
PROXY_LIST = {
"example1": "159.122.164.163:8080", # (Example) - set your own proxy here
"proxy1": None,
"proxy2": None,
"proxy3": None,
"proxy4": None,
"proxy5": None,
}
<commit_msg>Add another proxy server example string<commit_after>"""
Proxy Server "Phone Book".
Simplify running browser tests through a proxy server
by adding your frequently-used proxies here.
Now you can do something like this on the command line:
"pytest SOME_TEST.py --proxy=proxy1"
Format of PROXY_LIST server entries:
* "ip_address:port" OR "username:password@ip_address:port"
* "server:port" OR "username:password@server:port"
(Do NOT include the http:// or https:// in your proxy string!)
Example proxies in PROXY_LIST below are not guaranteed to be active or secure.
If you don't already have a proxy server to connect to,
you can try finding one from one of following sites:
* https://www.us-proxy.org/
"""
PROXY_LIST = {
"example1": "159.122.164.163:8080", # (Example) - set your own proxy here
"example2": "158.69.138.8:1080", # (Example) - set your own proxy here
"proxy1": None,
"proxy2": None,
"proxy3": None,
"proxy4": None,
"proxy5": None,
}
|
1a15a08abd7c7b5313402be4574ca6811044fd75
|
launch_control/models/hw_device.py
|
launch_control/models/hw_device.py
|
"""
Module with the HardwareDevice model.
"""
from launch_control.utils.json import PlainOldData
class HardwareDevice(PlainOldData):
"""
Model representing any HardwareDevice
A device is just a "device_type" attribute with a bag of properties
and a human readable description. Individual device types can be
freely added. For simplicity some common types of devices are
provided as class properties DEVICE_xxx.
Instances will come from a variety of factory classes, each capable
of enumerating devices that it understands. The upside of having a
common class like this is that it's easier to store it in the
database _and_ not have to agree on a common set of properties for,
say, all CPUs.
If you want you can create instances manually, like this:
>>> cpu = HardwareDevice(HardwareDevice.DEVICE_CPU)
>>> cpu.desc = "800MHz OMAP3 Processor"
>>> cpu.attributes['machine'] = 'arm'
>>> cpu.attributes['mhz'] = 800
>>> cpu.attributes['vendor'] = 'Texas Instruments'
"""
DEVICE_CPU = "device.cpu"
DEVICE_MEM = "device.mem"
DEVICE_USB = "device.usb"
DEVICE_PCI = "device.pci"
DEVICE_BOARD = "device.board"
__slots__ = ('device_type', 'desc', 'attributes')
def __init__(self, device_type, desc, attributes=None):
self.device_type = device_type
self.description = description
self.attributes = attributes or {}
|
"""
Module with the HardwareDevice model.
"""
from launch_control.utils.json import PlainOldData
class HardwareDevice(PlainOldData):
"""
Model representing any HardwareDevice
A device is just a "device_type" attribute with a bag of properties
and a human readable description. Individual device types can be
freely added. For simplicity some common types of devices are
provided as class properties DEVICE_xxx.
Instances will come from a variety of factory classes, each capable
of enumerating devices that it understands. The upside of having a
common class like this is that it's easier to store it in the
database _and_ not have to agree on a common set of properties for,
say, all CPUs.
If you want you can create instances manually, like this:
>>> cpu = HardwareDevice(HardwareDevice.DEVICE_CPU)
>>> cpu.desc = "800MHz OMAP3 Processor"
>>> cpu.attributes['machine'] = 'arm'
>>> cpu.attributes['mhz'] = 800
>>> cpu.attributes['vendor'] = 'Texas Instruments'
"""
DEVICE_CPU = "device.cpu"
DEVICE_MEM = "device.mem"
DEVICE_USB = "device.usb"
DEVICE_PCI = "device.pci"
DEVICE_BOARD = "device.board"
__slots__ = ('device_type', 'desc', 'attributes')
def __init__(self, device_type, description, attributes=None):
self.device_type = device_type
self.description = description
self.attributes = attributes or {}
|
Fix HardwareDevice constructor to provide 'description' argument
|
Fix HardwareDevice constructor to provide 'description' argument
|
Python
|
agpl-3.0
|
Linaro/lava-server,Linaro/lava-server,OSSystems/lava-server,Linaro/lava-server,Linaro/lava-server,OSSystems/lava-server,OSSystems/lava-server
|
"""
Module with the HardwareDevice model.
"""
from launch_control.utils.json import PlainOldData
class HardwareDevice(PlainOldData):
"""
Model representing any HardwareDevice
A device is just a "device_type" attribute with a bag of properties
and a human readable description. Individual device types can be
freely added. For simplicity some common types of devices are
provided as class properties DEVICE_xxx.
Instances will come from a variety of factory classes, each capable
of enumerating devices that it understands. The upside of having a
common class like this is that it's easier to store it in the
database _and_ not have to agree on a common set of properties for,
say, all CPUs.
If you want you can create instances manually, like this:
>>> cpu = HardwareDevice(HardwareDevice.DEVICE_CPU)
>>> cpu.desc = "800MHz OMAP3 Processor"
>>> cpu.attributes['machine'] = 'arm'
>>> cpu.attributes['mhz'] = 800
>>> cpu.attributes['vendor'] = 'Texas Instruments'
"""
DEVICE_CPU = "device.cpu"
DEVICE_MEM = "device.mem"
DEVICE_USB = "device.usb"
DEVICE_PCI = "device.pci"
DEVICE_BOARD = "device.board"
__slots__ = ('device_type', 'desc', 'attributes')
def __init__(self, device_type, desc, attributes=None):
self.device_type = device_type
self.description = description
self.attributes = attributes or {}
Fix HardwareDevice constructor to provide 'description' argument
|
"""
Module with the HardwareDevice model.
"""
from launch_control.utils.json import PlainOldData
class HardwareDevice(PlainOldData):
"""
Model representing any HardwareDevice
A device is just a "device_type" attribute with a bag of properties
and a human readable description. Individual device types can be
freely added. For simplicity some common types of devices are
provided as class properties DEVICE_xxx.
Instances will come from a variety of factory classes, each capable
of enumerating devices that it understands. The upside of having a
common class like this is that it's easier to store it in the
database _and_ not have to agree on a common set of properties for,
say, all CPUs.
If you want you can create instances manually, like this:
>>> cpu = HardwareDevice(HardwareDevice.DEVICE_CPU)
>>> cpu.desc = "800MHz OMAP3 Processor"
>>> cpu.attributes['machine'] = 'arm'
>>> cpu.attributes['mhz'] = 800
>>> cpu.attributes['vendor'] = 'Texas Instruments'
"""
DEVICE_CPU = "device.cpu"
DEVICE_MEM = "device.mem"
DEVICE_USB = "device.usb"
DEVICE_PCI = "device.pci"
DEVICE_BOARD = "device.board"
__slots__ = ('device_type', 'desc', 'attributes')
def __init__(self, device_type, description, attributes=None):
self.device_type = device_type
self.description = description
self.attributes = attributes or {}
|
<commit_before>"""
Module with the HardwareDevice model.
"""
from launch_control.utils.json import PlainOldData
class HardwareDevice(PlainOldData):
"""
Model representing any HardwareDevice
A device is just a "device_type" attribute with a bag of properties
and a human readable description. Individual device types can be
freely added. For simplicity some common types of devices are
provided as class properties DEVICE_xxx.
Instances will come from a variety of factory classes, each capable
of enumerating devices that it understands. The upside of having a
common class like this is that it's easier to store it in the
database _and_ not have to agree on a common set of properties for,
say, all CPUs.
If you want you can create instances manually, like this:
>>> cpu = HardwareDevice(HardwareDevice.DEVICE_CPU)
>>> cpu.desc = "800MHz OMAP3 Processor"
>>> cpu.attributes['machine'] = 'arm'
>>> cpu.attributes['mhz'] = 800
>>> cpu.attributes['vendor'] = 'Texas Instruments'
"""
DEVICE_CPU = "device.cpu"
DEVICE_MEM = "device.mem"
DEVICE_USB = "device.usb"
DEVICE_PCI = "device.pci"
DEVICE_BOARD = "device.board"
__slots__ = ('device_type', 'desc', 'attributes')
def __init__(self, device_type, desc, attributes=None):
self.device_type = device_type
self.description = description
self.attributes = attributes or {}
<commit_msg>Fix HardwareDevice constructor to provide 'description' argument<commit_after>
|
"""
Module with the HardwareDevice model.
"""
from launch_control.utils.json import PlainOldData
class HardwareDevice(PlainOldData):
"""
Model representing any HardwareDevice
A device is just a "device_type" attribute with a bag of properties
and a human readable description. Individual device types can be
freely added. For simplicity some common types of devices are
provided as class properties DEVICE_xxx.
Instances will come from a variety of factory classes, each capable
of enumerating devices that it understands. The upside of having a
common class like this is that it's easier to store it in the
database _and_ not have to agree on a common set of properties for,
say, all CPUs.
If you want you can create instances manually, like this:
>>> cpu = HardwareDevice(HardwareDevice.DEVICE_CPU)
>>> cpu.desc = "800MHz OMAP3 Processor"
>>> cpu.attributes['machine'] = 'arm'
>>> cpu.attributes['mhz'] = 800
>>> cpu.attributes['vendor'] = 'Texas Instruments'
"""
DEVICE_CPU = "device.cpu"
DEVICE_MEM = "device.mem"
DEVICE_USB = "device.usb"
DEVICE_PCI = "device.pci"
DEVICE_BOARD = "device.board"
__slots__ = ('device_type', 'desc', 'attributes')
def __init__(self, device_type, description, attributes=None):
self.device_type = device_type
self.description = description
self.attributes = attributes or {}
|
"""
Module with the HardwareDevice model.
"""
from launch_control.utils.json import PlainOldData
class HardwareDevice(PlainOldData):
"""
Model representing any HardwareDevice
A device is just a "device_type" attribute with a bag of properties
and a human readable description. Individual device types can be
freely added. For simplicity some common types of devices are
provided as class properties DEVICE_xxx.
Instances will come from a variety of factory classes, each capable
of enumerating devices that it understands. The upside of having a
common class like this is that it's easier to store it in the
database _and_ not have to agree on a common set of properties for,
say, all CPUs.
If you want you can create instances manually, like this:
>>> cpu = HardwareDevice(HardwareDevice.DEVICE_CPU)
>>> cpu.desc = "800MHz OMAP3 Processor"
>>> cpu.attributes['machine'] = 'arm'
>>> cpu.attributes['mhz'] = 800
>>> cpu.attributes['vendor'] = 'Texas Instruments'
"""
DEVICE_CPU = "device.cpu"
DEVICE_MEM = "device.mem"
DEVICE_USB = "device.usb"
DEVICE_PCI = "device.pci"
DEVICE_BOARD = "device.board"
__slots__ = ('device_type', 'desc', 'attributes')
def __init__(self, device_type, desc, attributes=None):
self.device_type = device_type
self.description = description
self.attributes = attributes or {}
Fix HardwareDevice constructor to provide 'description' argument"""
Module with the HardwareDevice model.
"""
from launch_control.utils.json import PlainOldData
class HardwareDevice(PlainOldData):
"""
Model representing any HardwareDevice
A device is just a "device_type" attribute with a bag of properties
and a human readable description. Individual device types can be
freely added. For simplicity some common types of devices are
provided as class properties DEVICE_xxx.
Instances will come from a variety of factory classes, each capable
of enumerating devices that it understands. The upside of having a
common class like this is that it's easier to store it in the
database _and_ not have to agree on a common set of properties for,
say, all CPUs.
If you want you can create instances manually, like this:
>>> cpu = HardwareDevice(HardwareDevice.DEVICE_CPU)
>>> cpu.desc = "800MHz OMAP3 Processor"
>>> cpu.attributes['machine'] = 'arm'
>>> cpu.attributes['mhz'] = 800
>>> cpu.attributes['vendor'] = 'Texas Instruments'
"""
DEVICE_CPU = "device.cpu"
DEVICE_MEM = "device.mem"
DEVICE_USB = "device.usb"
DEVICE_PCI = "device.pci"
DEVICE_BOARD = "device.board"
__slots__ = ('device_type', 'desc', 'attributes')
def __init__(self, device_type, description, attributes=None):
self.device_type = device_type
self.description = description
self.attributes = attributes or {}
|
<commit_before>"""
Module with the HardwareDevice model.
"""
from launch_control.utils.json import PlainOldData
class HardwareDevice(PlainOldData):
"""
Model representing any HardwareDevice
A device is just a "device_type" attribute with a bag of properties
and a human readable description. Individual device types can be
freely added. For simplicity some common types of devices are
provided as class properties DEVICE_xxx.
Instances will come from a variety of factory classes, each capable
of enumerating devices that it understands. The upside of having a
common class like this is that it's easier to store it in the
database _and_ not have to agree on a common set of properties for,
say, all CPUs.
If you want you can create instances manually, like this:
>>> cpu = HardwareDevice(HardwareDevice.DEVICE_CPU)
>>> cpu.desc = "800MHz OMAP3 Processor"
>>> cpu.attributes['machine'] = 'arm'
>>> cpu.attributes['mhz'] = 800
>>> cpu.attributes['vendor'] = 'Texas Instruments'
"""
DEVICE_CPU = "device.cpu"
DEVICE_MEM = "device.mem"
DEVICE_USB = "device.usb"
DEVICE_PCI = "device.pci"
DEVICE_BOARD = "device.board"
__slots__ = ('device_type', 'desc', 'attributes')
def __init__(self, device_type, desc, attributes=None):
self.device_type = device_type
self.description = description
self.attributes = attributes or {}
<commit_msg>Fix HardwareDevice constructor to provide 'description' argument<commit_after>"""
Module with the HardwareDevice model.
"""
from launch_control.utils.json import PlainOldData
class HardwareDevice(PlainOldData):
"""
Model representing any HardwareDevice
A device is just a "device_type" attribute with a bag of properties
and a human readable description. Individual device types can be
freely added. For simplicity some common types of devices are
provided as class properties DEVICE_xxx.
Instances will come from a variety of factory classes, each capable
of enumerating devices that it understands. The upside of having a
common class like this is that it's easier to store it in the
database _and_ not have to agree on a common set of properties for,
say, all CPUs.
If you want you can create instances manually, like this:
>>> cpu = HardwareDevice(HardwareDevice.DEVICE_CPU)
>>> cpu.desc = "800MHz OMAP3 Processor"
>>> cpu.attributes['machine'] = 'arm'
>>> cpu.attributes['mhz'] = 800
>>> cpu.attributes['vendor'] = 'Texas Instruments'
"""
DEVICE_CPU = "device.cpu"
DEVICE_MEM = "device.mem"
DEVICE_USB = "device.usb"
DEVICE_PCI = "device.pci"
DEVICE_BOARD = "device.board"
__slots__ = ('device_type', 'desc', 'attributes')
def __init__(self, device_type, description, attributes=None):
self.device_type = device_type
self.description = description
self.attributes = attributes or {}
|
8fff587b9fd7e2cd0ca4d45e869345cbfb248045
|
troposphere/workspaces.py
|
troposphere/workspaces.py
|
# Copyright (c) 2015, Mark Peek <mark@peek.org>
# All rights reserved.
#
# See LICENSE file for full license.
from . import AWSObject
class Workspace(AWSObject):
resource_type = "AWS::WorkSpaces::Workspace"
props = {
'BundleId': (basestring, True),
'DirectoryId': (basestring, True),
'UserName': (basestring, True),
}
|
# Copyright (c) 2015, Mark Peek <mark@peek.org>
# All rights reserved.
#
# See LICENSE file for full license.
from . import AWSObject
from .validators import boolean
class Workspace(AWSObject):
resource_type = "AWS::WorkSpaces::Workspace"
props = {
'BundleId': (basestring, True),
'DirectoryId': (basestring, True),
'UserName': (basestring, True),
'RootVolumeEncryptionEnabled': (boolean, False),
'UserVolumeEncryptionEnabled': (boolean, False),
'VolumeEncryptionKey': (basestring, False),
}
|
Add encryption properties to Workspace
|
Add encryption properties to Workspace
|
Python
|
bsd-2-clause
|
7digital/troposphere,dmm92/troposphere,horacio3/troposphere,ikben/troposphere,alonsodomin/troposphere,pas256/troposphere,cloudtools/troposphere,dmm92/troposphere,ikben/troposphere,Yipit/troposphere,cloudtools/troposphere,johnctitus/troposphere,amosshapira/troposphere,johnctitus/troposphere,pas256/troposphere,alonsodomin/troposphere,7digital/troposphere,craigbruce/troposphere,horacio3/troposphere
|
# Copyright (c) 2015, Mark Peek <mark@peek.org>
# All rights reserved.
#
# See LICENSE file for full license.
from . import AWSObject
class Workspace(AWSObject):
resource_type = "AWS::WorkSpaces::Workspace"
props = {
'BundleId': (basestring, True),
'DirectoryId': (basestring, True),
'UserName': (basestring, True),
}
Add encryption properties to Workspace
|
# Copyright (c) 2015, Mark Peek <mark@peek.org>
# All rights reserved.
#
# See LICENSE file for full license.
from . import AWSObject
from .validators import boolean
class Workspace(AWSObject):
resource_type = "AWS::WorkSpaces::Workspace"
props = {
'BundleId': (basestring, True),
'DirectoryId': (basestring, True),
'UserName': (basestring, True),
'RootVolumeEncryptionEnabled': (boolean, False),
'UserVolumeEncryptionEnabled': (boolean, False),
'VolumeEncryptionKey': (basestring, False),
}
|
<commit_before># Copyright (c) 2015, Mark Peek <mark@peek.org>
# All rights reserved.
#
# See LICENSE file for full license.
from . import AWSObject
class Workspace(AWSObject):
resource_type = "AWS::WorkSpaces::Workspace"
props = {
'BundleId': (basestring, True),
'DirectoryId': (basestring, True),
'UserName': (basestring, True),
}
<commit_msg>Add encryption properties to Workspace<commit_after>
|
# Copyright (c) 2015, Mark Peek <mark@peek.org>
# All rights reserved.
#
# See LICENSE file for full license.
from . import AWSObject
from .validators import boolean
class Workspace(AWSObject):
resource_type = "AWS::WorkSpaces::Workspace"
props = {
'BundleId': (basestring, True),
'DirectoryId': (basestring, True),
'UserName': (basestring, True),
'RootVolumeEncryptionEnabled': (boolean, False),
'UserVolumeEncryptionEnabled': (boolean, False),
'VolumeEncryptionKey': (basestring, False),
}
|
# Copyright (c) 2015, Mark Peek <mark@peek.org>
# All rights reserved.
#
# See LICENSE file for full license.
from . import AWSObject
class Workspace(AWSObject):
resource_type = "AWS::WorkSpaces::Workspace"
props = {
'BundleId': (basestring, True),
'DirectoryId': (basestring, True),
'UserName': (basestring, True),
}
Add encryption properties to Workspace# Copyright (c) 2015, Mark Peek <mark@peek.org>
# All rights reserved.
#
# See LICENSE file for full license.
from . import AWSObject
from .validators import boolean
class Workspace(AWSObject):
resource_type = "AWS::WorkSpaces::Workspace"
props = {
'BundleId': (basestring, True),
'DirectoryId': (basestring, True),
'UserName': (basestring, True),
'RootVolumeEncryptionEnabled': (boolean, False),
'UserVolumeEncryptionEnabled': (boolean, False),
'VolumeEncryptionKey': (basestring, False),
}
|
<commit_before># Copyright (c) 2015, Mark Peek <mark@peek.org>
# All rights reserved.
#
# See LICENSE file for full license.
from . import AWSObject
class Workspace(AWSObject):
resource_type = "AWS::WorkSpaces::Workspace"
props = {
'BundleId': (basestring, True),
'DirectoryId': (basestring, True),
'UserName': (basestring, True),
}
<commit_msg>Add encryption properties to Workspace<commit_after># Copyright (c) 2015, Mark Peek <mark@peek.org>
# All rights reserved.
#
# See LICENSE file for full license.
from . import AWSObject
from .validators import boolean
class Workspace(AWSObject):
resource_type = "AWS::WorkSpaces::Workspace"
props = {
'BundleId': (basestring, True),
'DirectoryId': (basestring, True),
'UserName': (basestring, True),
'RootVolumeEncryptionEnabled': (boolean, False),
'UserVolumeEncryptionEnabled': (boolean, False),
'VolumeEncryptionKey': (basestring, False),
}
|
b4e106271f96b083644b27d313ad80c240fcb0a5
|
gapipy/resources/booking/booking.py
|
gapipy/resources/booking/booking.py
|
# Python 2 and 3
from __future__ import unicode_literals
from gapipy.resources.checkin import Checkin
from ..base import Resource
from .transaction import Payment, Refund
from .document import Invoice, Document
from .override import Override
from .service import Service
class Booking(Resource):
_resource_name = 'bookings'
_is_parent_resource = True
_as_is_fields = ['id', 'href', 'external_id', 'currency']
_price_fields = [
'amount_owing',
'amount_paid',
'amount_pending',
'commission',
'tax_on_commission',
]
_date_fields = [
'date_closed', 'date_of_first_travel', 'date_of_last_travel',
'balance_due_date',
]
_date_time_fields_utc = ['date_created', ]
_resource_fields = [
('agent', 'Agent'),
('agency', 'Agency'),
('associated_agency', 'Agency'),
]
@property
def _resource_collection_fields(self):
return [
('services', Service),
('invoices', Invoice),
('payments', Payment),
('refunds', Refund),
('documents', Document),
('overrides', Override),
('checkins', Checkin),
]
|
# Python 2 and 3
from __future__ import unicode_literals
from gapipy.resources.checkin import Checkin
from ..base import Resource
from .agency_chain import AgencyChain
from .document import Invoice, Document
from .override import Override
from .service import Service
from .transaction import Payment, Refund
class Booking(Resource):
_resource_name = 'bookings'
_is_parent_resource = True
_as_is_fields = ['id', 'href', 'external_id', 'currency']
_price_fields = [
'amount_owing',
'amount_paid',
'amount_pending',
'commission',
'tax_on_commission',
]
_date_fields = [
'date_closed', 'date_of_first_travel', 'date_of_last_travel',
'balance_due_date',
]
_date_time_fields_utc = ['date_created', ]
_resource_fields = [
('agency', 'Agency'),
('agency_chain', AgencyChain),
('agent', 'Agent'),
('associated_agency', 'Agency'),
]
@property
def _resource_collection_fields(self):
return [
('services', Service),
('invoices', Invoice),
('payments', Payment),
('refunds', Refund),
('documents', Document),
('overrides', Override),
('checkins', Checkin),
]
|
Add agency chain to Booking
|
Add agency chain to Booking
|
Python
|
mit
|
gadventures/gapipy
|
# Python 2 and 3
from __future__ import unicode_literals
from gapipy.resources.checkin import Checkin
from ..base import Resource
from .transaction import Payment, Refund
from .document import Invoice, Document
from .override import Override
from .service import Service
class Booking(Resource):
_resource_name = 'bookings'
_is_parent_resource = True
_as_is_fields = ['id', 'href', 'external_id', 'currency']
_price_fields = [
'amount_owing',
'amount_paid',
'amount_pending',
'commission',
'tax_on_commission',
]
_date_fields = [
'date_closed', 'date_of_first_travel', 'date_of_last_travel',
'balance_due_date',
]
_date_time_fields_utc = ['date_created', ]
_resource_fields = [
('agent', 'Agent'),
('agency', 'Agency'),
('associated_agency', 'Agency'),
]
@property
def _resource_collection_fields(self):
return [
('services', Service),
('invoices', Invoice),
('payments', Payment),
('refunds', Refund),
('documents', Document),
('overrides', Override),
('checkins', Checkin),
]
Add agency chain to Booking
|
# Python 2 and 3
from __future__ import unicode_literals
from gapipy.resources.checkin import Checkin
from ..base import Resource
from .agency_chain import AgencyChain
from .document import Invoice, Document
from .override import Override
from .service import Service
from .transaction import Payment, Refund
class Booking(Resource):
_resource_name = 'bookings'
_is_parent_resource = True
_as_is_fields = ['id', 'href', 'external_id', 'currency']
_price_fields = [
'amount_owing',
'amount_paid',
'amount_pending',
'commission',
'tax_on_commission',
]
_date_fields = [
'date_closed', 'date_of_first_travel', 'date_of_last_travel',
'balance_due_date',
]
_date_time_fields_utc = ['date_created', ]
_resource_fields = [
('agency', 'Agency'),
('agency_chain', AgencyChain),
('agent', 'Agent'),
('associated_agency', 'Agency'),
]
@property
def _resource_collection_fields(self):
return [
('services', Service),
('invoices', Invoice),
('payments', Payment),
('refunds', Refund),
('documents', Document),
('overrides', Override),
('checkins', Checkin),
]
|
<commit_before># Python 2 and 3
from __future__ import unicode_literals
from gapipy.resources.checkin import Checkin
from ..base import Resource
from .transaction import Payment, Refund
from .document import Invoice, Document
from .override import Override
from .service import Service
class Booking(Resource):
_resource_name = 'bookings'
_is_parent_resource = True
_as_is_fields = ['id', 'href', 'external_id', 'currency']
_price_fields = [
'amount_owing',
'amount_paid',
'amount_pending',
'commission',
'tax_on_commission',
]
_date_fields = [
'date_closed', 'date_of_first_travel', 'date_of_last_travel',
'balance_due_date',
]
_date_time_fields_utc = ['date_created', ]
_resource_fields = [
('agent', 'Agent'),
('agency', 'Agency'),
('associated_agency', 'Agency'),
]
@property
def _resource_collection_fields(self):
return [
('services', Service),
('invoices', Invoice),
('payments', Payment),
('refunds', Refund),
('documents', Document),
('overrides', Override),
('checkins', Checkin),
]
<commit_msg>Add agency chain to Booking<commit_after>
|
# Python 2 and 3
from __future__ import unicode_literals
from gapipy.resources.checkin import Checkin
from ..base import Resource
from .agency_chain import AgencyChain
from .document import Invoice, Document
from .override import Override
from .service import Service
from .transaction import Payment, Refund
class Booking(Resource):
_resource_name = 'bookings'
_is_parent_resource = True
_as_is_fields = ['id', 'href', 'external_id', 'currency']
_price_fields = [
'amount_owing',
'amount_paid',
'amount_pending',
'commission',
'tax_on_commission',
]
_date_fields = [
'date_closed', 'date_of_first_travel', 'date_of_last_travel',
'balance_due_date',
]
_date_time_fields_utc = ['date_created', ]
_resource_fields = [
('agency', 'Agency'),
('agency_chain', AgencyChain),
('agent', 'Agent'),
('associated_agency', 'Agency'),
]
@property
def _resource_collection_fields(self):
return [
('services', Service),
('invoices', Invoice),
('payments', Payment),
('refunds', Refund),
('documents', Document),
('overrides', Override),
('checkins', Checkin),
]
|
# Python 2 and 3
from __future__ import unicode_literals
from gapipy.resources.checkin import Checkin
from ..base import Resource
from .transaction import Payment, Refund
from .document import Invoice, Document
from .override import Override
from .service import Service
class Booking(Resource):
_resource_name = 'bookings'
_is_parent_resource = True
_as_is_fields = ['id', 'href', 'external_id', 'currency']
_price_fields = [
'amount_owing',
'amount_paid',
'amount_pending',
'commission',
'tax_on_commission',
]
_date_fields = [
'date_closed', 'date_of_first_travel', 'date_of_last_travel',
'balance_due_date',
]
_date_time_fields_utc = ['date_created', ]
_resource_fields = [
('agent', 'Agent'),
('agency', 'Agency'),
('associated_agency', 'Agency'),
]
@property
def _resource_collection_fields(self):
return [
('services', Service),
('invoices', Invoice),
('payments', Payment),
('refunds', Refund),
('documents', Document),
('overrides', Override),
('checkins', Checkin),
]
Add agency chain to Booking# Python 2 and 3
from __future__ import unicode_literals
from gapipy.resources.checkin import Checkin
from ..base import Resource
from .agency_chain import AgencyChain
from .document import Invoice, Document
from .override import Override
from .service import Service
from .transaction import Payment, Refund
class Booking(Resource):
_resource_name = 'bookings'
_is_parent_resource = True
_as_is_fields = ['id', 'href', 'external_id', 'currency']
_price_fields = [
'amount_owing',
'amount_paid',
'amount_pending',
'commission',
'tax_on_commission',
]
_date_fields = [
'date_closed', 'date_of_first_travel', 'date_of_last_travel',
'balance_due_date',
]
_date_time_fields_utc = ['date_created', ]
_resource_fields = [
('agency', 'Agency'),
('agency_chain', AgencyChain),
('agent', 'Agent'),
('associated_agency', 'Agency'),
]
@property
def _resource_collection_fields(self):
return [
('services', Service),
('invoices', Invoice),
('payments', Payment),
('refunds', Refund),
('documents', Document),
('overrides', Override),
('checkins', Checkin),
]
|
<commit_before># Python 2 and 3
from __future__ import unicode_literals
from gapipy.resources.checkin import Checkin
from ..base import Resource
from .transaction import Payment, Refund
from .document import Invoice, Document
from .override import Override
from .service import Service
class Booking(Resource):
_resource_name = 'bookings'
_is_parent_resource = True
_as_is_fields = ['id', 'href', 'external_id', 'currency']
_price_fields = [
'amount_owing',
'amount_paid',
'amount_pending',
'commission',
'tax_on_commission',
]
_date_fields = [
'date_closed', 'date_of_first_travel', 'date_of_last_travel',
'balance_due_date',
]
_date_time_fields_utc = ['date_created', ]
_resource_fields = [
('agent', 'Agent'),
('agency', 'Agency'),
('associated_agency', 'Agency'),
]
@property
def _resource_collection_fields(self):
return [
('services', Service),
('invoices', Invoice),
('payments', Payment),
('refunds', Refund),
('documents', Document),
('overrides', Override),
('checkins', Checkin),
]
<commit_msg>Add agency chain to Booking<commit_after># Python 2 and 3
from __future__ import unicode_literals
from gapipy.resources.checkin import Checkin
from ..base import Resource
from .agency_chain import AgencyChain
from .document import Invoice, Document
from .override import Override
from .service import Service
from .transaction import Payment, Refund
class Booking(Resource):
_resource_name = 'bookings'
_is_parent_resource = True
_as_is_fields = ['id', 'href', 'external_id', 'currency']
_price_fields = [
'amount_owing',
'amount_paid',
'amount_pending',
'commission',
'tax_on_commission',
]
_date_fields = [
'date_closed', 'date_of_first_travel', 'date_of_last_travel',
'balance_due_date',
]
_date_time_fields_utc = ['date_created', ]
_resource_fields = [
('agency', 'Agency'),
('agency_chain', AgencyChain),
('agent', 'Agent'),
('associated_agency', 'Agency'),
]
@property
def _resource_collection_fields(self):
return [
('services', Service),
('invoices', Invoice),
('payments', Payment),
('refunds', Refund),
('documents', Document),
('overrides', Override),
('checkins', Checkin),
]
|
8ba94b216531f249a7097f10eb74f363af6151e2
|
xmlrpc2/client.py
|
xmlrpc2/client.py
|
from __future__ import absolute_import
from __future__ import division
from __future__ import unicode_literals
import urllib.parse
class BaseTransport(object):
@property
def scheme(self):
raise NotImplementedError("Transports must have a scheme")
class HTTPTransport(BaseTransport):
scheme = "http"
class Client(object):
def __init__(self, uri, transports=None):
if transports is None:
transports = [HTTPTransport]
# Initialize transports
self.transports = {}
for transport in transports:
t = transport()
self.transports[t.scheme] = t
parsed = urllib.parse.urlparse(uri)
if parsed.scheme not in self.transports:
raise ValueError("Invalid uri scheme {scheme}. Must be one of {available}.".format(scheme=parsed.scheme, available=",".join(self.transports)))
self.transport = self.transports[parsed.scheme]
# Default to /RPC2 for path as it is a common endpoint
if not parsed.path:
parsed = parsed[:2] + ("/RPC2",) + parsed[3:]
self.uri = urllib.parse.urlunparse(parsed)
|
from __future__ import absolute_import
from __future__ import division
from __future__ import unicode_literals
import urllib.parse
class BaseTransport(object):
@property
def scheme(self):
raise NotImplementedError("Transports must have a scheme")
class HTTPTransport(BaseTransport):
scheme = "http"
class Client(object):
def __init__(self, uri, transports=None):
if transports is None:
transports = [HTTPTransport]
# Initialize transports
self._transports = {}
for transport in transports:
t = transport()
self._transports[t.scheme] = t
parsed = urllib.parse.urlparse(uri)
if parsed.scheme not in self._transports:
raise ValueError("Invalid uri scheme {scheme}. Must be one of {available}.".format(scheme=parsed.scheme, available=",".join(self._transports)))
self._transport = self._transports[parsed.scheme]
# Default to /RPC2 for path as it is a common endpoint
if not parsed.path:
parsed = parsed[:2] + ("/RPC2",) + parsed[3:]
self._uri = urllib.parse.urlunparse(parsed)
|
Rename internal variables to start with a _
|
Rename internal variables to start with a _
|
Python
|
bsd-2-clause
|
dstufft/xmlrpc2
|
from __future__ import absolute_import
from __future__ import division
from __future__ import unicode_literals
import urllib.parse
class BaseTransport(object):
@property
def scheme(self):
raise NotImplementedError("Transports must have a scheme")
class HTTPTransport(BaseTransport):
scheme = "http"
class Client(object):
def __init__(self, uri, transports=None):
if transports is None:
transports = [HTTPTransport]
# Initialize transports
self.transports = {}
for transport in transports:
t = transport()
self.transports[t.scheme] = t
parsed = urllib.parse.urlparse(uri)
if parsed.scheme not in self.transports:
raise ValueError("Invalid uri scheme {scheme}. Must be one of {available}.".format(scheme=parsed.scheme, available=",".join(self.transports)))
self.transport = self.transports[parsed.scheme]
# Default to /RPC2 for path as it is a common endpoint
if not parsed.path:
parsed = parsed[:2] + ("/RPC2",) + parsed[3:]
self.uri = urllib.parse.urlunparse(parsed)
Rename internal variables to start with a _
|
from __future__ import absolute_import
from __future__ import division
from __future__ import unicode_literals
import urllib.parse
class BaseTransport(object):
@property
def scheme(self):
raise NotImplementedError("Transports must have a scheme")
class HTTPTransport(BaseTransport):
scheme = "http"
class Client(object):
def __init__(self, uri, transports=None):
if transports is None:
transports = [HTTPTransport]
# Initialize transports
self._transports = {}
for transport in transports:
t = transport()
self._transports[t.scheme] = t
parsed = urllib.parse.urlparse(uri)
if parsed.scheme not in self._transports:
raise ValueError("Invalid uri scheme {scheme}. Must be one of {available}.".format(scheme=parsed.scheme, available=",".join(self._transports)))
self._transport = self._transports[parsed.scheme]
# Default to /RPC2 for path as it is a common endpoint
if not parsed.path:
parsed = parsed[:2] + ("/RPC2",) + parsed[3:]
self._uri = urllib.parse.urlunparse(parsed)
|
<commit_before>from __future__ import absolute_import
from __future__ import division
from __future__ import unicode_literals
import urllib.parse
class BaseTransport(object):
@property
def scheme(self):
raise NotImplementedError("Transports must have a scheme")
class HTTPTransport(BaseTransport):
scheme = "http"
class Client(object):
def __init__(self, uri, transports=None):
if transports is None:
transports = [HTTPTransport]
# Initialize transports
self.transports = {}
for transport in transports:
t = transport()
self.transports[t.scheme] = t
parsed = urllib.parse.urlparse(uri)
if parsed.scheme not in self.transports:
raise ValueError("Invalid uri scheme {scheme}. Must be one of {available}.".format(scheme=parsed.scheme, available=",".join(self.transports)))
self.transport = self.transports[parsed.scheme]
# Default to /RPC2 for path as it is a common endpoint
if not parsed.path:
parsed = parsed[:2] + ("/RPC2",) + parsed[3:]
self.uri = urllib.parse.urlunparse(parsed)
<commit_msg>Rename internal variables to start with a _<commit_after>
|
from __future__ import absolute_import
from __future__ import division
from __future__ import unicode_literals
import urllib.parse
class BaseTransport(object):
@property
def scheme(self):
raise NotImplementedError("Transports must have a scheme")
class HTTPTransport(BaseTransport):
scheme = "http"
class Client(object):
def __init__(self, uri, transports=None):
if transports is None:
transports = [HTTPTransport]
# Initialize transports
self._transports = {}
for transport in transports:
t = transport()
self._transports[t.scheme] = t
parsed = urllib.parse.urlparse(uri)
if parsed.scheme not in self._transports:
raise ValueError("Invalid uri scheme {scheme}. Must be one of {available}.".format(scheme=parsed.scheme, available=",".join(self._transports)))
self._transport = self._transports[parsed.scheme]
# Default to /RPC2 for path as it is a common endpoint
if not parsed.path:
parsed = parsed[:2] + ("/RPC2",) + parsed[3:]
self._uri = urllib.parse.urlunparse(parsed)
|
from __future__ import absolute_import
from __future__ import division
from __future__ import unicode_literals
import urllib.parse
class BaseTransport(object):
@property
def scheme(self):
raise NotImplementedError("Transports must have a scheme")
class HTTPTransport(BaseTransport):
scheme = "http"
class Client(object):
def __init__(self, uri, transports=None):
if transports is None:
transports = [HTTPTransport]
# Initialize transports
self.transports = {}
for transport in transports:
t = transport()
self.transports[t.scheme] = t
parsed = urllib.parse.urlparse(uri)
if parsed.scheme not in self.transports:
raise ValueError("Invalid uri scheme {scheme}. Must be one of {available}.".format(scheme=parsed.scheme, available=",".join(self.transports)))
self.transport = self.transports[parsed.scheme]
# Default to /RPC2 for path as it is a common endpoint
if not parsed.path:
parsed = parsed[:2] + ("/RPC2",) + parsed[3:]
self.uri = urllib.parse.urlunparse(parsed)
Rename internal variables to start with a _from __future__ import absolute_import
from __future__ import division
from __future__ import unicode_literals
import urllib.parse
class BaseTransport(object):
@property
def scheme(self):
raise NotImplementedError("Transports must have a scheme")
class HTTPTransport(BaseTransport):
scheme = "http"
class Client(object):
def __init__(self, uri, transports=None):
if transports is None:
transports = [HTTPTransport]
# Initialize transports
self._transports = {}
for transport in transports:
t = transport()
self._transports[t.scheme] = t
parsed = urllib.parse.urlparse(uri)
if parsed.scheme not in self._transports:
raise ValueError("Invalid uri scheme {scheme}. Must be one of {available}.".format(scheme=parsed.scheme, available=",".join(self._transports)))
self._transport = self._transports[parsed.scheme]
# Default to /RPC2 for path as it is a common endpoint
if not parsed.path:
parsed = parsed[:2] + ("/RPC2",) + parsed[3:]
self._uri = urllib.parse.urlunparse(parsed)
|
<commit_before>from __future__ import absolute_import
from __future__ import division
from __future__ import unicode_literals
import urllib.parse
class BaseTransport(object):
@property
def scheme(self):
raise NotImplementedError("Transports must have a scheme")
class HTTPTransport(BaseTransport):
scheme = "http"
class Client(object):
def __init__(self, uri, transports=None):
if transports is None:
transports = [HTTPTransport]
# Initialize transports
self.transports = {}
for transport in transports:
t = transport()
self.transports[t.scheme] = t
parsed = urllib.parse.urlparse(uri)
if parsed.scheme not in self.transports:
raise ValueError("Invalid uri scheme {scheme}. Must be one of {available}.".format(scheme=parsed.scheme, available=",".join(self.transports)))
self.transport = self.transports[parsed.scheme]
# Default to /RPC2 for path as it is a common endpoint
if not parsed.path:
parsed = parsed[:2] + ("/RPC2",) + parsed[3:]
self.uri = urllib.parse.urlunparse(parsed)
<commit_msg>Rename internal variables to start with a _<commit_after>from __future__ import absolute_import
from __future__ import division
from __future__ import unicode_literals
import urllib.parse
class BaseTransport(object):
@property
def scheme(self):
raise NotImplementedError("Transports must have a scheme")
class HTTPTransport(BaseTransport):
scheme = "http"
class Client(object):
def __init__(self, uri, transports=None):
if transports is None:
transports = [HTTPTransport]
# Initialize transports
self._transports = {}
for transport in transports:
t = transport()
self._transports[t.scheme] = t
parsed = urllib.parse.urlparse(uri)
if parsed.scheme not in self._transports:
raise ValueError("Invalid uri scheme {scheme}. Must be one of {available}.".format(scheme=parsed.scheme, available=",".join(self._transports)))
self._transport = self._transports[parsed.scheme]
# Default to /RPC2 for path as it is a common endpoint
if not parsed.path:
parsed = parsed[:2] + ("/RPC2",) + parsed[3:]
self._uri = urllib.parse.urlunparse(parsed)
|
e64195a005be583f32754e49e870b198ee7bc396
|
corehq/pillows/mappings/case_search_mapping.py
|
corehq/pillows/mappings/case_search_mapping.py
|
from corehq.pillows.mappings.case_mapping import CASE_ES_TYPE
from corehq.pillows.mappings.utils import mapping_from_json
from corehq.util.elastic import es_index
from pillowtop.es_utils import ElasticsearchIndexInfo
CASE_SEARCH_INDEX = es_index("case_search_2016-03-15")
CASE_SEARCH_ALIAS = "case_search"
CASE_SEARCH_MAX_RESULTS = 10
CASE_SEARCH_MAPPING = mapping_from_json('case_search_mapping.json')
CASE_SEARCH_INDEX_INFO = ElasticsearchIndexInfo(
index=CASE_SEARCH_INDEX,
alias=CASE_SEARCH_ALIAS,
type=CASE_ES_TYPE,
mapping=CASE_SEARCH_MAPPING,
)
|
from corehq.pillows.mappings.case_mapping import CASE_ES_TYPE
from corehq.pillows.mappings.utils import mapping_from_json
from corehq.util.elastic import es_index
from pillowtop.es_utils import ElasticsearchIndexInfo
CASE_SEARCH_INDEX = es_index("case_search_2016-03-15")
CASE_SEARCH_ALIAS = "case_search"
CASE_SEARCH_MAX_RESULTS = 100
CASE_SEARCH_MAPPING = mapping_from_json('case_search_mapping.json')
CASE_SEARCH_INDEX_INFO = ElasticsearchIndexInfo(
index=CASE_SEARCH_INDEX,
alias=CASE_SEARCH_ALIAS,
type=CASE_ES_TYPE,
mapping=CASE_SEARCH_MAPPING,
)
|
Increase case search limit to 100 results
|
Increase case search limit to 100 results
|
Python
|
bsd-3-clause
|
dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq,dimagi/commcare-hq
|
from corehq.pillows.mappings.case_mapping import CASE_ES_TYPE
from corehq.pillows.mappings.utils import mapping_from_json
from corehq.util.elastic import es_index
from pillowtop.es_utils import ElasticsearchIndexInfo
CASE_SEARCH_INDEX = es_index("case_search_2016-03-15")
CASE_SEARCH_ALIAS = "case_search"
CASE_SEARCH_MAX_RESULTS = 10
CASE_SEARCH_MAPPING = mapping_from_json('case_search_mapping.json')
CASE_SEARCH_INDEX_INFO = ElasticsearchIndexInfo(
index=CASE_SEARCH_INDEX,
alias=CASE_SEARCH_ALIAS,
type=CASE_ES_TYPE,
mapping=CASE_SEARCH_MAPPING,
)
Increase case search limit to 100 results
|
from corehq.pillows.mappings.case_mapping import CASE_ES_TYPE
from corehq.pillows.mappings.utils import mapping_from_json
from corehq.util.elastic import es_index
from pillowtop.es_utils import ElasticsearchIndexInfo
CASE_SEARCH_INDEX = es_index("case_search_2016-03-15")
CASE_SEARCH_ALIAS = "case_search"
CASE_SEARCH_MAX_RESULTS = 100
CASE_SEARCH_MAPPING = mapping_from_json('case_search_mapping.json')
CASE_SEARCH_INDEX_INFO = ElasticsearchIndexInfo(
index=CASE_SEARCH_INDEX,
alias=CASE_SEARCH_ALIAS,
type=CASE_ES_TYPE,
mapping=CASE_SEARCH_MAPPING,
)
|
<commit_before>from corehq.pillows.mappings.case_mapping import CASE_ES_TYPE
from corehq.pillows.mappings.utils import mapping_from_json
from corehq.util.elastic import es_index
from pillowtop.es_utils import ElasticsearchIndexInfo
CASE_SEARCH_INDEX = es_index("case_search_2016-03-15")
CASE_SEARCH_ALIAS = "case_search"
CASE_SEARCH_MAX_RESULTS = 10
CASE_SEARCH_MAPPING = mapping_from_json('case_search_mapping.json')
CASE_SEARCH_INDEX_INFO = ElasticsearchIndexInfo(
index=CASE_SEARCH_INDEX,
alias=CASE_SEARCH_ALIAS,
type=CASE_ES_TYPE,
mapping=CASE_SEARCH_MAPPING,
)
<commit_msg>Increase case search limit to 100 results<commit_after>
|
from corehq.pillows.mappings.case_mapping import CASE_ES_TYPE
from corehq.pillows.mappings.utils import mapping_from_json
from corehq.util.elastic import es_index
from pillowtop.es_utils import ElasticsearchIndexInfo
CASE_SEARCH_INDEX = es_index("case_search_2016-03-15")
CASE_SEARCH_ALIAS = "case_search"
CASE_SEARCH_MAX_RESULTS = 100
CASE_SEARCH_MAPPING = mapping_from_json('case_search_mapping.json')
CASE_SEARCH_INDEX_INFO = ElasticsearchIndexInfo(
index=CASE_SEARCH_INDEX,
alias=CASE_SEARCH_ALIAS,
type=CASE_ES_TYPE,
mapping=CASE_SEARCH_MAPPING,
)
|
from corehq.pillows.mappings.case_mapping import CASE_ES_TYPE
from corehq.pillows.mappings.utils import mapping_from_json
from corehq.util.elastic import es_index
from pillowtop.es_utils import ElasticsearchIndexInfo
CASE_SEARCH_INDEX = es_index("case_search_2016-03-15")
CASE_SEARCH_ALIAS = "case_search"
CASE_SEARCH_MAX_RESULTS = 10
CASE_SEARCH_MAPPING = mapping_from_json('case_search_mapping.json')
CASE_SEARCH_INDEX_INFO = ElasticsearchIndexInfo(
index=CASE_SEARCH_INDEX,
alias=CASE_SEARCH_ALIAS,
type=CASE_ES_TYPE,
mapping=CASE_SEARCH_MAPPING,
)
Increase case search limit to 100 resultsfrom corehq.pillows.mappings.case_mapping import CASE_ES_TYPE
from corehq.pillows.mappings.utils import mapping_from_json
from corehq.util.elastic import es_index
from pillowtop.es_utils import ElasticsearchIndexInfo
CASE_SEARCH_INDEX = es_index("case_search_2016-03-15")
CASE_SEARCH_ALIAS = "case_search"
CASE_SEARCH_MAX_RESULTS = 100
CASE_SEARCH_MAPPING = mapping_from_json('case_search_mapping.json')
CASE_SEARCH_INDEX_INFO = ElasticsearchIndexInfo(
index=CASE_SEARCH_INDEX,
alias=CASE_SEARCH_ALIAS,
type=CASE_ES_TYPE,
mapping=CASE_SEARCH_MAPPING,
)
|
<commit_before>from corehq.pillows.mappings.case_mapping import CASE_ES_TYPE
from corehq.pillows.mappings.utils import mapping_from_json
from corehq.util.elastic import es_index
from pillowtop.es_utils import ElasticsearchIndexInfo
CASE_SEARCH_INDEX = es_index("case_search_2016-03-15")
CASE_SEARCH_ALIAS = "case_search"
CASE_SEARCH_MAX_RESULTS = 10
CASE_SEARCH_MAPPING = mapping_from_json('case_search_mapping.json')
CASE_SEARCH_INDEX_INFO = ElasticsearchIndexInfo(
index=CASE_SEARCH_INDEX,
alias=CASE_SEARCH_ALIAS,
type=CASE_ES_TYPE,
mapping=CASE_SEARCH_MAPPING,
)
<commit_msg>Increase case search limit to 100 results<commit_after>from corehq.pillows.mappings.case_mapping import CASE_ES_TYPE
from corehq.pillows.mappings.utils import mapping_from_json
from corehq.util.elastic import es_index
from pillowtop.es_utils import ElasticsearchIndexInfo
CASE_SEARCH_INDEX = es_index("case_search_2016-03-15")
CASE_SEARCH_ALIAS = "case_search"
CASE_SEARCH_MAX_RESULTS = 100
CASE_SEARCH_MAPPING = mapping_from_json('case_search_mapping.json')
CASE_SEARCH_INDEX_INFO = ElasticsearchIndexInfo(
index=CASE_SEARCH_INDEX,
alias=CASE_SEARCH_ALIAS,
type=CASE_ES_TYPE,
mapping=CASE_SEARCH_MAPPING,
)
|
db20fc6b7a21efbd7de0f5b0d1aa754c19c1a21f
|
race/management/commands/update_leaderboard.py
|
race/management/commands/update_leaderboard.py
|
from django.core.management.base import BaseCommand
from django.conf import settings
from ...models import OverallDriverPrediction, OverallConstructorPrediction
class Command(BaseCommand):
can_import_settings = True
def handle(self, *args, **kwargs):
conn = settings.REDIS_CONN
num_ranks = conn.zcard("ranks")
conn.zremrangebyscore("ranks", 0, num_ranks)
for driver_prediction in OverallDriverPrediction.objects.all():
conn.zadd("ranks",
driver_prediction.user.username,
driver_prediction.score)
for constructor_prediction in OverallConstructorPrediction.objects.all():
score = conn.zscore("ranks", constructor_prediction.user.username)
if not score:
score = 0
conn.zadd("ranks",
constructor_prediction.user.username,
constructor_prediction.score + score)
|
from django.core.management.base import BaseCommand
from django.conf import settings
from ...models import OverallDriverPrediction, OverallConstructorPrediction
class Command(BaseCommand):
can_import_settings = True
def handle(self, *args, **kwargs):
conn = settings.REDIS_CONN
num_ranks = conn.zcard("ranks")
conn.zremrangebyscore("ranks", 0, num_ranks + 1)
for driver_prediction in OverallDriverPrediction.objects.all():
conn.zadd("ranks",
driver_prediction.user.username,
driver_prediction.score)
for constructor_prediction in OverallConstructorPrediction.objects.all():
score = conn.zscore("ranks", constructor_prediction.user.username)
if not score:
score = 0
conn.zadd("ranks",
constructor_prediction.user.username,
constructor_prediction.score + score)
|
Remove all scores before populating the sorted set.
|
Remove all scores before populating the sorted set.
|
Python
|
bsd-3-clause
|
theju/f1oracle,theju/f1oracle
|
from django.core.management.base import BaseCommand
from django.conf import settings
from ...models import OverallDriverPrediction, OverallConstructorPrediction
class Command(BaseCommand):
can_import_settings = True
def handle(self, *args, **kwargs):
conn = settings.REDIS_CONN
num_ranks = conn.zcard("ranks")
conn.zremrangebyscore("ranks", 0, num_ranks)
for driver_prediction in OverallDriverPrediction.objects.all():
conn.zadd("ranks",
driver_prediction.user.username,
driver_prediction.score)
for constructor_prediction in OverallConstructorPrediction.objects.all():
score = conn.zscore("ranks", constructor_prediction.user.username)
if not score:
score = 0
conn.zadd("ranks",
constructor_prediction.user.username,
constructor_prediction.score + score)
Remove all scores before populating the sorted set.
|
from django.core.management.base import BaseCommand
from django.conf import settings
from ...models import OverallDriverPrediction, OverallConstructorPrediction
class Command(BaseCommand):
can_import_settings = True
def handle(self, *args, **kwargs):
conn = settings.REDIS_CONN
num_ranks = conn.zcard("ranks")
conn.zremrangebyscore("ranks", 0, num_ranks + 1)
for driver_prediction in OverallDriverPrediction.objects.all():
conn.zadd("ranks",
driver_prediction.user.username,
driver_prediction.score)
for constructor_prediction in OverallConstructorPrediction.objects.all():
score = conn.zscore("ranks", constructor_prediction.user.username)
if not score:
score = 0
conn.zadd("ranks",
constructor_prediction.user.username,
constructor_prediction.score + score)
|
<commit_before>from django.core.management.base import BaseCommand
from django.conf import settings
from ...models import OverallDriverPrediction, OverallConstructorPrediction
class Command(BaseCommand):
can_import_settings = True
def handle(self, *args, **kwargs):
conn = settings.REDIS_CONN
num_ranks = conn.zcard("ranks")
conn.zremrangebyscore("ranks", 0, num_ranks)
for driver_prediction in OverallDriverPrediction.objects.all():
conn.zadd("ranks",
driver_prediction.user.username,
driver_prediction.score)
for constructor_prediction in OverallConstructorPrediction.objects.all():
score = conn.zscore("ranks", constructor_prediction.user.username)
if not score:
score = 0
conn.zadd("ranks",
constructor_prediction.user.username,
constructor_prediction.score + score)
<commit_msg>Remove all scores before populating the sorted set.<commit_after>
|
from django.core.management.base import BaseCommand
from django.conf import settings
from ...models import OverallDriverPrediction, OverallConstructorPrediction
class Command(BaseCommand):
can_import_settings = True
def handle(self, *args, **kwargs):
conn = settings.REDIS_CONN
num_ranks = conn.zcard("ranks")
conn.zremrangebyscore("ranks", 0, num_ranks + 1)
for driver_prediction in OverallDriverPrediction.objects.all():
conn.zadd("ranks",
driver_prediction.user.username,
driver_prediction.score)
for constructor_prediction in OverallConstructorPrediction.objects.all():
score = conn.zscore("ranks", constructor_prediction.user.username)
if not score:
score = 0
conn.zadd("ranks",
constructor_prediction.user.username,
constructor_prediction.score + score)
|
from django.core.management.base import BaseCommand
from django.conf import settings
from ...models import OverallDriverPrediction, OverallConstructorPrediction
class Command(BaseCommand):
can_import_settings = True
def handle(self, *args, **kwargs):
conn = settings.REDIS_CONN
num_ranks = conn.zcard("ranks")
conn.zremrangebyscore("ranks", 0, num_ranks)
for driver_prediction in OverallDriverPrediction.objects.all():
conn.zadd("ranks",
driver_prediction.user.username,
driver_prediction.score)
for constructor_prediction in OverallConstructorPrediction.objects.all():
score = conn.zscore("ranks", constructor_prediction.user.username)
if not score:
score = 0
conn.zadd("ranks",
constructor_prediction.user.username,
constructor_prediction.score + score)
Remove all scores before populating the sorted set.from django.core.management.base import BaseCommand
from django.conf import settings
from ...models import OverallDriverPrediction, OverallConstructorPrediction
class Command(BaseCommand):
can_import_settings = True
def handle(self, *args, **kwargs):
conn = settings.REDIS_CONN
num_ranks = conn.zcard("ranks")
conn.zremrangebyscore("ranks", 0, num_ranks + 1)
for driver_prediction in OverallDriverPrediction.objects.all():
conn.zadd("ranks",
driver_prediction.user.username,
driver_prediction.score)
for constructor_prediction in OverallConstructorPrediction.objects.all():
score = conn.zscore("ranks", constructor_prediction.user.username)
if not score:
score = 0
conn.zadd("ranks",
constructor_prediction.user.username,
constructor_prediction.score + score)
|
<commit_before>from django.core.management.base import BaseCommand
from django.conf import settings
from ...models import OverallDriverPrediction, OverallConstructorPrediction
class Command(BaseCommand):
can_import_settings = True
def handle(self, *args, **kwargs):
conn = settings.REDIS_CONN
num_ranks = conn.zcard("ranks")
conn.zremrangebyscore("ranks", 0, num_ranks)
for driver_prediction in OverallDriverPrediction.objects.all():
conn.zadd("ranks",
driver_prediction.user.username,
driver_prediction.score)
for constructor_prediction in OverallConstructorPrediction.objects.all():
score = conn.zscore("ranks", constructor_prediction.user.username)
if not score:
score = 0
conn.zadd("ranks",
constructor_prediction.user.username,
constructor_prediction.score + score)
<commit_msg>Remove all scores before populating the sorted set.<commit_after>from django.core.management.base import BaseCommand
from django.conf import settings
from ...models import OverallDriverPrediction, OverallConstructorPrediction
class Command(BaseCommand):
can_import_settings = True
def handle(self, *args, **kwargs):
conn = settings.REDIS_CONN
num_ranks = conn.zcard("ranks")
conn.zremrangebyscore("ranks", 0, num_ranks + 1)
for driver_prediction in OverallDriverPrediction.objects.all():
conn.zadd("ranks",
driver_prediction.user.username,
driver_prediction.score)
for constructor_prediction in OverallConstructorPrediction.objects.all():
score = conn.zscore("ranks", constructor_prediction.user.username)
if not score:
score = 0
conn.zadd("ranks",
constructor_prediction.user.username,
constructor_prediction.score + score)
|
0f0e0e91db679f18ad9dc7568047b76e447ac589
|
stock_inventory_chatter/__openerp__.py
|
stock_inventory_chatter/__openerp__.py
|
# -*- coding: utf-8 -*-
# Copyright 2017 Eficent Business and IT Consulting Services S.L.
# (http://www.eficent.com)
# License AGPL-3.0 or later (https://www.gnu.org/licenses/agpl.html).
{
'name': 'Stock Inventory Chatter',
'version': '9.0.1.0.0',
'author': "Eficent, "
"Odoo Community Association (OCA)",
"website": "https://github.com/OCA/stock-logistics-warehouse",
'category': 'Warehouse',
'summary': "Log changes being done in Inventory Adjustments",
'depends': ['stock'],
"data": [
'data/stock_data.xml',
'views/stock_inventory_view.xml',
],
'license': 'AGPL-3',
'installable': True,
'application': False,
}
|
# -*- coding: utf-8 -*-
# Copyright 2017 Eficent Business and IT Consulting Services S.L.
# Copyright 2018 initOS GmbH
# (http://www.eficent.com)
# License AGPL-3.0 or later (https://www.gnu.org/licenses/agpl.html).
{
'name': 'Stock Inventory Chatter',
'version': '8.0.1.0.0',
'author': "Eficent, "
"initOS GmbH, "
"Odoo Community Association (OCA)",
"website": "https://github.com/OCA/stock-logistics-warehouse",
'category': 'Warehouse',
'summary': "Log changes being done in Inventory Adjustments",
'depends': ['stock'],
"data": [
'data/stock_data.xml',
'views/stock_inventory_view.xml',
],
'license': 'AGPL-3',
'installable': True,
'application': False,
}
|
Change of the module version
|
Change of the module version
|
Python
|
agpl-3.0
|
kmee/stock-logistics-warehouse,acsone/stock-logistics-warehouse,open-synergy/stock-logistics-warehouse
|
# -*- coding: utf-8 -*-
# Copyright 2017 Eficent Business and IT Consulting Services S.L.
# (http://www.eficent.com)
# License AGPL-3.0 or later (https://www.gnu.org/licenses/agpl.html).
{
'name': 'Stock Inventory Chatter',
'version': '9.0.1.0.0',
'author': "Eficent, "
"Odoo Community Association (OCA)",
"website": "https://github.com/OCA/stock-logistics-warehouse",
'category': 'Warehouse',
'summary': "Log changes being done in Inventory Adjustments",
'depends': ['stock'],
"data": [
'data/stock_data.xml',
'views/stock_inventory_view.xml',
],
'license': 'AGPL-3',
'installable': True,
'application': False,
}
Change of the module version
|
# -*- coding: utf-8 -*-
# Copyright 2017 Eficent Business and IT Consulting Services S.L.
# Copyright 2018 initOS GmbH
# (http://www.eficent.com)
# License AGPL-3.0 or later (https://www.gnu.org/licenses/agpl.html).
{
'name': 'Stock Inventory Chatter',
'version': '8.0.1.0.0',
'author': "Eficent, "
"initOS GmbH, "
"Odoo Community Association (OCA)",
"website": "https://github.com/OCA/stock-logistics-warehouse",
'category': 'Warehouse',
'summary': "Log changes being done in Inventory Adjustments",
'depends': ['stock'],
"data": [
'data/stock_data.xml',
'views/stock_inventory_view.xml',
],
'license': 'AGPL-3',
'installable': True,
'application': False,
}
|
<commit_before># -*- coding: utf-8 -*-
# Copyright 2017 Eficent Business and IT Consulting Services S.L.
# (http://www.eficent.com)
# License AGPL-3.0 or later (https://www.gnu.org/licenses/agpl.html).
{
'name': 'Stock Inventory Chatter',
'version': '9.0.1.0.0',
'author': "Eficent, "
"Odoo Community Association (OCA)",
"website": "https://github.com/OCA/stock-logistics-warehouse",
'category': 'Warehouse',
'summary': "Log changes being done in Inventory Adjustments",
'depends': ['stock'],
"data": [
'data/stock_data.xml',
'views/stock_inventory_view.xml',
],
'license': 'AGPL-3',
'installable': True,
'application': False,
}
<commit_msg>Change of the module version<commit_after>
|
# -*- coding: utf-8 -*-
# Copyright 2017 Eficent Business and IT Consulting Services S.L.
# Copyright 2018 initOS GmbH
# (http://www.eficent.com)
# License AGPL-3.0 or later (https://www.gnu.org/licenses/agpl.html).
{
'name': 'Stock Inventory Chatter',
'version': '8.0.1.0.0',
'author': "Eficent, "
"initOS GmbH, "
"Odoo Community Association (OCA)",
"website": "https://github.com/OCA/stock-logistics-warehouse",
'category': 'Warehouse',
'summary': "Log changes being done in Inventory Adjustments",
'depends': ['stock'],
"data": [
'data/stock_data.xml',
'views/stock_inventory_view.xml',
],
'license': 'AGPL-3',
'installable': True,
'application': False,
}
|
# -*- coding: utf-8 -*-
# Copyright 2017 Eficent Business and IT Consulting Services S.L.
# (http://www.eficent.com)
# License AGPL-3.0 or later (https://www.gnu.org/licenses/agpl.html).
{
'name': 'Stock Inventory Chatter',
'version': '9.0.1.0.0',
'author': "Eficent, "
"Odoo Community Association (OCA)",
"website": "https://github.com/OCA/stock-logistics-warehouse",
'category': 'Warehouse',
'summary': "Log changes being done in Inventory Adjustments",
'depends': ['stock'],
"data": [
'data/stock_data.xml',
'views/stock_inventory_view.xml',
],
'license': 'AGPL-3',
'installable': True,
'application': False,
}
Change of the module version# -*- coding: utf-8 -*-
# Copyright 2017 Eficent Business and IT Consulting Services S.L.
# Copyright 2018 initOS GmbH
# (http://www.eficent.com)
# License AGPL-3.0 or later (https://www.gnu.org/licenses/agpl.html).
{
'name': 'Stock Inventory Chatter',
'version': '8.0.1.0.0',
'author': "Eficent, "
"initOS GmbH, "
"Odoo Community Association (OCA)",
"website": "https://github.com/OCA/stock-logistics-warehouse",
'category': 'Warehouse',
'summary': "Log changes being done in Inventory Adjustments",
'depends': ['stock'],
"data": [
'data/stock_data.xml',
'views/stock_inventory_view.xml',
],
'license': 'AGPL-3',
'installable': True,
'application': False,
}
|
<commit_before># -*- coding: utf-8 -*-
# Copyright 2017 Eficent Business and IT Consulting Services S.L.
# (http://www.eficent.com)
# License AGPL-3.0 or later (https://www.gnu.org/licenses/agpl.html).
{
'name': 'Stock Inventory Chatter',
'version': '9.0.1.0.0',
'author': "Eficent, "
"Odoo Community Association (OCA)",
"website": "https://github.com/OCA/stock-logistics-warehouse",
'category': 'Warehouse',
'summary': "Log changes being done in Inventory Adjustments",
'depends': ['stock'],
"data": [
'data/stock_data.xml',
'views/stock_inventory_view.xml',
],
'license': 'AGPL-3',
'installable': True,
'application': False,
}
<commit_msg>Change of the module version<commit_after># -*- coding: utf-8 -*-
# Copyright 2017 Eficent Business and IT Consulting Services S.L.
# Copyright 2018 initOS GmbH
# (http://www.eficent.com)
# License AGPL-3.0 or later (https://www.gnu.org/licenses/agpl.html).
{
'name': 'Stock Inventory Chatter',
'version': '8.0.1.0.0',
'author': "Eficent, "
"initOS GmbH, "
"Odoo Community Association (OCA)",
"website": "https://github.com/OCA/stock-logistics-warehouse",
'category': 'Warehouse',
'summary': "Log changes being done in Inventory Adjustments",
'depends': ['stock'],
"data": [
'data/stock_data.xml',
'views/stock_inventory_view.xml',
],
'license': 'AGPL-3',
'installable': True,
'application': False,
}
|
d3227e87b658b4ee634dd273a97d1a8fba4c96c9
|
lc461_hamming_distance.py
|
lc461_hamming_distance.py
|
"""Leetcode 461. Hamming Distance
Medium
URL: https://leetcode.com/problems/hamming-distance/description/
The Hamming distance between two integers is the number of positions at which
the corresponding bits are different.
Given two integers x and y, calculate the Hamming distance.
Note:
0 β€ x, y < 231.
Example:
Input: x = 1, y = 4
Output: 2
Explanation:
1 (0 0 0 1)
4 (0 1 0 0)
β β
The above arrows point to positions where the corresponding bits are different.
"""
class Solution(object):
def hammingDistance(self, x, y):
"""
:type x: int
:type y: int
:rtype: int
Time complexity: O(1).
Space complexity: O(1).
"""
return bin(x ^ y).count('1')
def main():
print Solution().hammingDistance(1, 4)
if __name__ == '__main__':
main()
|
"""Leetcode 461. Hamming Distance
Medium
URL: https://leetcode.com/problems/hamming-distance/description/
The Hamming distance between two integers is the number of positions at which
the corresponding bits are different.
Given two integers x and y, calculate the Hamming distance.
Note:
0 β€ x, y < 231.
Example:
Input: x = 1, y = 4
Output: 2
Explanation:
1 (0 0 0 1)
4 (0 1 0 0)
β β
The above arrows point to positions where the corresponding bits
are different.
"""
class Solution(object):
def hammingDistance(self, x, y):
"""
:type x: int
:type y: int
:rtype: int
Time complexity: O(1).
Space complexity: O(1).
"""
return bin(x ^ y).count('1')
def main():
print Solution().hammingDistance(1, 4)
if __name__ == '__main__':
main()
|
Revise docstring and add space line
|
Revise docstring and add space line
|
Python
|
bsd-2-clause
|
bowen0701/algorithms_data_structures
|
"""Leetcode 461. Hamming Distance
Medium
URL: https://leetcode.com/problems/hamming-distance/description/
The Hamming distance between two integers is the number of positions at which
the corresponding bits are different.
Given two integers x and y, calculate the Hamming distance.
Note:
0 β€ x, y < 231.
Example:
Input: x = 1, y = 4
Output: 2
Explanation:
1 (0 0 0 1)
4 (0 1 0 0)
β β
The above arrows point to positions where the corresponding bits are different.
"""
class Solution(object):
def hammingDistance(self, x, y):
"""
:type x: int
:type y: int
:rtype: int
Time complexity: O(1).
Space complexity: O(1).
"""
return bin(x ^ y).count('1')
def main():
print Solution().hammingDistance(1, 4)
if __name__ == '__main__':
main()
Revise docstring and add space line
|
"""Leetcode 461. Hamming Distance
Medium
URL: https://leetcode.com/problems/hamming-distance/description/
The Hamming distance between two integers is the number of positions at which
the corresponding bits are different.
Given two integers x and y, calculate the Hamming distance.
Note:
0 β€ x, y < 231.
Example:
Input: x = 1, y = 4
Output: 2
Explanation:
1 (0 0 0 1)
4 (0 1 0 0)
β β
The above arrows point to positions where the corresponding bits
are different.
"""
class Solution(object):
def hammingDistance(self, x, y):
"""
:type x: int
:type y: int
:rtype: int
Time complexity: O(1).
Space complexity: O(1).
"""
return bin(x ^ y).count('1')
def main():
print Solution().hammingDistance(1, 4)
if __name__ == '__main__':
main()
|
<commit_before>"""Leetcode 461. Hamming Distance
Medium
URL: https://leetcode.com/problems/hamming-distance/description/
The Hamming distance between two integers is the number of positions at which
the corresponding bits are different.
Given two integers x and y, calculate the Hamming distance.
Note:
0 β€ x, y < 231.
Example:
Input: x = 1, y = 4
Output: 2
Explanation:
1 (0 0 0 1)
4 (0 1 0 0)
β β
The above arrows point to positions where the corresponding bits are different.
"""
class Solution(object):
def hammingDistance(self, x, y):
"""
:type x: int
:type y: int
:rtype: int
Time complexity: O(1).
Space complexity: O(1).
"""
return bin(x ^ y).count('1')
def main():
print Solution().hammingDistance(1, 4)
if __name__ == '__main__':
main()
<commit_msg>Revise docstring and add space line<commit_after>
|
"""Leetcode 461. Hamming Distance
Medium
URL: https://leetcode.com/problems/hamming-distance/description/
The Hamming distance between two integers is the number of positions at which
the corresponding bits are different.
Given two integers x and y, calculate the Hamming distance.
Note:
0 β€ x, y < 231.
Example:
Input: x = 1, y = 4
Output: 2
Explanation:
1 (0 0 0 1)
4 (0 1 0 0)
β β
The above arrows point to positions where the corresponding bits
are different.
"""
class Solution(object):
def hammingDistance(self, x, y):
"""
:type x: int
:type y: int
:rtype: int
Time complexity: O(1).
Space complexity: O(1).
"""
return bin(x ^ y).count('1')
def main():
print Solution().hammingDistance(1, 4)
if __name__ == '__main__':
main()
|
"""Leetcode 461. Hamming Distance
Medium
URL: https://leetcode.com/problems/hamming-distance/description/
The Hamming distance between two integers is the number of positions at which
the corresponding bits are different.
Given two integers x and y, calculate the Hamming distance.
Note:
0 β€ x, y < 231.
Example:
Input: x = 1, y = 4
Output: 2
Explanation:
1 (0 0 0 1)
4 (0 1 0 0)
β β
The above arrows point to positions where the corresponding bits are different.
"""
class Solution(object):
def hammingDistance(self, x, y):
"""
:type x: int
:type y: int
:rtype: int
Time complexity: O(1).
Space complexity: O(1).
"""
return bin(x ^ y).count('1')
def main():
print Solution().hammingDistance(1, 4)
if __name__ == '__main__':
main()
Revise docstring and add space line"""Leetcode 461. Hamming Distance
Medium
URL: https://leetcode.com/problems/hamming-distance/description/
The Hamming distance between two integers is the number of positions at which
the corresponding bits are different.
Given two integers x and y, calculate the Hamming distance.
Note:
0 β€ x, y < 231.
Example:
Input: x = 1, y = 4
Output: 2
Explanation:
1 (0 0 0 1)
4 (0 1 0 0)
β β
The above arrows point to positions where the corresponding bits
are different.
"""
class Solution(object):
def hammingDistance(self, x, y):
"""
:type x: int
:type y: int
:rtype: int
Time complexity: O(1).
Space complexity: O(1).
"""
return bin(x ^ y).count('1')
def main():
print Solution().hammingDistance(1, 4)
if __name__ == '__main__':
main()
|
<commit_before>"""Leetcode 461. Hamming Distance
Medium
URL: https://leetcode.com/problems/hamming-distance/description/
The Hamming distance between two integers is the number of positions at which
the corresponding bits are different.
Given two integers x and y, calculate the Hamming distance.
Note:
0 β€ x, y < 231.
Example:
Input: x = 1, y = 4
Output: 2
Explanation:
1 (0 0 0 1)
4 (0 1 0 0)
β β
The above arrows point to positions where the corresponding bits are different.
"""
class Solution(object):
def hammingDistance(self, x, y):
"""
:type x: int
:type y: int
:rtype: int
Time complexity: O(1).
Space complexity: O(1).
"""
return bin(x ^ y).count('1')
def main():
print Solution().hammingDistance(1, 4)
if __name__ == '__main__':
main()
<commit_msg>Revise docstring and add space line<commit_after>"""Leetcode 461. Hamming Distance
Medium
URL: https://leetcode.com/problems/hamming-distance/description/
The Hamming distance between two integers is the number of positions at which
the corresponding bits are different.
Given two integers x and y, calculate the Hamming distance.
Note:
0 β€ x, y < 231.
Example:
Input: x = 1, y = 4
Output: 2
Explanation:
1 (0 0 0 1)
4 (0 1 0 0)
β β
The above arrows point to positions where the corresponding bits
are different.
"""
class Solution(object):
def hammingDistance(self, x, y):
"""
:type x: int
:type y: int
:rtype: int
Time complexity: O(1).
Space complexity: O(1).
"""
return bin(x ^ y).count('1')
def main():
print Solution().hammingDistance(1, 4)
if __name__ == '__main__':
main()
|
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.