Dataset Viewer
Auto-converted to Parquet Duplicate
commit
stringlengths
40
40
old_file
stringlengths
4
106
new_file
stringlengths
4
106
old_contents
stringlengths
10
2.94k
new_contents
stringlengths
21
2.95k
subject
stringlengths
16
444
message
stringlengths
17
2.63k
lang
stringclasses
1 value
license
stringclasses
13 values
repos
stringlengths
7
43k
ndiff
stringlengths
52
3.31k
instruction
stringlengths
16
444
content
stringlengths
133
4.32k
diff
stringlengths
49
3.61k
bc5abf988956235b48aeb1234d9944fe70be619a
pytest_hidecaptured.py
pytest_hidecaptured.py
def pytest_runtest_logreport(report): """Overwrite report by removing any captured stderr.""" # print("PLUGIN SAYS -> report -> {0}".format(report)) # print("PLUGIN SAYS -> report.sections -> {0}".format(report.sections)) # print("PLUGIN SAYS -> dir(report) -> {0}".format(dir(report))) # print("PLUG...
import pytest @pytest.mark.tryfirst def pytest_runtest_logreport(report): """Overwrite report by removing any captured stderr.""" # print("PLUGIN SAYS -> report -> {0}".format(report)) # print("PLUGIN SAYS -> report.sections -> {0}".format(report.sections)) # print("PLUGIN SAYS -> dir(report) -> {0}".f...
Fix interop issues with pytest-instafail
Fix interop issues with pytest-instafail
Python
mit
hamzasheikh/pytest-hidecaptured
+ import pytest + + @pytest.mark.tryfirst def pytest_runtest_logreport(report): """Overwrite report by removing any captured stderr.""" # print("PLUGIN SAYS -> report -> {0}".format(report)) # print("PLUGIN SAYS -> report.sections -> {0}".format(report.sections)) # print("PLUGIN SAYS -> dir(r...
Fix interop issues with pytest-instafail
## Code Before: def pytest_runtest_logreport(report): """Overwrite report by removing any captured stderr.""" # print("PLUGIN SAYS -> report -> {0}".format(report)) # print("PLUGIN SAYS -> report.sections -> {0}".format(report.sections)) # print("PLUGIN SAYS -> dir(report) -> {0}".format(dir(report))) ...
+ import pytest + + @pytest.mark.tryfirst def pytest_runtest_logreport(report): """Overwrite report by removing any captured stderr.""" # print("PLUGIN SAYS -> report -> {0}".format(report)) # print("PLUGIN SAYS -> report.sections -> {0}".format(report.sections)) # print("PLUGIN SAYS -> dir(r...
cc93d6b9ade1d15236904978f012f91b0a9d567d
examples/manage.py
examples/manage.py
import logging from aio_manager import Manager from aioapp.app import build_application logging.basicConfig(level=logging.WARNING) app = build_application() manager = Manager(app) # To support SQLAlchemy commands, use this # # from aio_manager.commands.ext import sqlalchemy # sqlalchemy.configure_manager(manager, ap...
import logging from aio_manager import Manager from aioapp.app import build_application logging.basicConfig(level=logging.WARNING) app = build_application() manager = Manager(app) # To support SQLAlchemy commands, use this # # from aio_manager.commands.ext import sqlalchemy # [from aiopg.sa import create_engine] # s...
Update sqlalchemy command configuration example
Update sqlalchemy command configuration example
Python
bsd-3-clause
rrader/aio_manager
import logging from aio_manager import Manager from aioapp.app import build_application logging.basicConfig(level=logging.WARNING) app = build_application() manager = Manager(app) # To support SQLAlchemy commands, use this # # from aio_manager.commands.ext import sqlalchemy + # [from aiopg.sa...
Update sqlalchemy command configuration example
## Code Before: import logging from aio_manager import Manager from aioapp.app import build_application logging.basicConfig(level=logging.WARNING) app = build_application() manager = Manager(app) # To support SQLAlchemy commands, use this # # from aio_manager.commands.ext import sqlalchemy # sqlalchemy.configure_man...
import logging from aio_manager import Manager from aioapp.app import build_application logging.basicConfig(level=logging.WARNING) app = build_application() manager = Manager(app) # To support SQLAlchemy commands, use this # # from aio_manager.commands.ext import sqlalchemy + # [from aiopg.sa...
47d9a8df136e235f49921d4782c5e392b0101107
migrations/versions/147_add_cleaned_subject.py
migrations/versions/147_add_cleaned_subject.py
# revision identifiers, used by Alembic. revision = '486c7fa5b533' down_revision = 'c77a90d524' from alembic import op import sqlalchemy as sa from sqlalchemy.sql import text def upgrade(): conn = op.get_bind() conn.execute(text("set @@lock_wait_timeout = 20;")) op.add_column('thread', sa.Column('_clea...
# revision identifiers, used by Alembic. revision = '486c7fa5b533' down_revision = 'c77a90d524' from alembic import op import sqlalchemy as sa from sqlalchemy.sql import text def upgrade(): conn = op.get_bind() conn.execute(text("set @@lock_wait_timeout = 20;")) op.add_column('thread', sa.Column('_clea...
Make _cleaned_subject migration match declared schema.
Make _cleaned_subject migration match declared schema. Test Plan: Upgrade old database to head. Reviewers: kav-ya Reviewed By: kav-ya Differential Revision: https://review.inboxapp.com/D1394
Python
agpl-3.0
Eagles2F/sync-engine,Eagles2F/sync-engine,EthanBlackburn/sync-engine,PriviPK/privipk-sync-engine,PriviPK/privipk-sync-engine,nylas/sync-engine,closeio/nylas,jobscore/sync-engine,jobscore/sync-engine,jobscore/sync-engine,wakermahmud/sync-engine,PriviPK/privipk-sync-engine,wakermahmud/sync-engine,gale320/sync-engine,Eagl...
# revision identifiers, used by Alembic. revision = '486c7fa5b533' down_revision = 'c77a90d524' from alembic import op import sqlalchemy as sa from sqlalchemy.sql import text def upgrade(): conn = op.get_bind() conn.execute(text("set @@lock_wait_timeout = 20;")) op.add_col...
Make _cleaned_subject migration match declared schema.
## Code Before: # revision identifiers, used by Alembic. revision = '486c7fa5b533' down_revision = 'c77a90d524' from alembic import op import sqlalchemy as sa from sqlalchemy.sql import text def upgrade(): conn = op.get_bind() conn.execute(text("set @@lock_wait_timeout = 20;")) op.add_column('thread', ...
# revision identifiers, used by Alembic. revision = '486c7fa5b533' down_revision = 'c77a90d524' from alembic import op import sqlalchemy as sa from sqlalchemy.sql import text def upgrade(): conn = op.get_bind() conn.execute(text("set @@lock_wait_timeout = 20;")) op.add_col...
1ef76b4f4395c9b5e3c2338822947999d5581013
labs/lab-3/ex-3-2.events.py
labs/lab-3/ex-3-2.events.py
import tspapi api = tspapi.API() source = tspapi.Source(ref='myhost') api.event_create(title="bar", fingerprint_fields=['@title'], source=source)
import tspapi api = tspapi.API() source = tspapi.Source(ref='myhost', _type='host') api.event_create(title="bar", fingerprint_fields=['@title'], source=source)
Add type field to source
Add type field to source
Python
apache-2.0
jdgwartney/tsi-lab,boundary/tsi-lab,jdgwartney/tsi-lab,boundary/tsi-lab,boundary/tsi-lab,boundary/tsi-lab,jdgwartney/tsi-lab,jdgwartney/tsi-lab
import tspapi api = tspapi.API() - source = tspapi.Source(ref='myhost') + source = tspapi.Source(ref='myhost', _type='host') api.event_create(title="bar", fingerprint_fields=['@title'], source=source)
Add type field to source
## Code Before: import tspapi api = tspapi.API() source = tspapi.Source(ref='myhost') api.event_create(title="bar", fingerprint_fields=['@title'], source=source) ## Instruction: Add type field to source ## Code After: import tspapi api = tspapi.API() source = tspapi.Source(ref='myhost', _type='host') api.event_cre...
import tspapi api = tspapi.API() - source = tspapi.Source(ref='myhost') + source = tspapi.Source(ref='myhost', _type='host') ? ++++++++++++++ api.event_create(title="bar", fingerprint_fields=['@title'], source=source)
6ea9d0c4b4e2a117e3e74c34cc77f83d262e62d8
sendgrid_events/models.py
sendgrid_events/models.py
import json from django.db import models from django.utils import timezone from jsonfield import JSONField from sendgrid_events.signals import batch_processed class Event(models.Model): kind = models.CharField(max_length=75) email = models.CharField(max_length=150) data = JSONField(blank=True) crea...
import json from django.db import models from django.utils import timezone from jsonfield import JSONField from sendgrid_events.signals import batch_processed class Event(models.Model): kind = models.CharField(max_length=75) email = models.CharField(max_length=150) data = JSONField(blank=True) crea...
Update for latest Sendgrid webhook format
Update for latest Sendgrid webhook format
Python
bsd-3-clause
digital-eskimo/django-sendgrid-events,kronok/django-sendgrid-events,eldarion/django-sendgrid-events,rorito/django-sendgrid-events
import json from django.db import models from django.utils import timezone from jsonfield import JSONField from sendgrid_events.signals import batch_processed class Event(models.Model): kind = models.CharField(max_length=75) email = models.CharField(max_length=150) data = JS...
Update for latest Sendgrid webhook format
## Code Before: import json from django.db import models from django.utils import timezone from jsonfield import JSONField from sendgrid_events.signals import batch_processed class Event(models.Model): kind = models.CharField(max_length=75) email = models.CharField(max_length=150) data = JSONField(blan...
import json from django.db import models from django.utils import timezone from jsonfield import JSONField from sendgrid_events.signals import batch_processed class Event(models.Model): kind = models.CharField(max_length=75) email = models.CharField(max_length=150) data = JS...
e14b3fad26dce8dad3ca97c06e624f1d6b0764f9
mqueue/__init__.py
mqueue/__init__.py
__version__ = '0.5.5' default_app_config = 'mqueue.apps.MqueueConfig'
__version__ = '0.5.5' default_app_config = 'mqueue.apps.MqueueConfig' import sys reload(sys) sys.setdefaultencoding("utf-8")
Set default encoding to fix unicode errors
Set default encoding to fix unicode errors
Python
mit
synw/django-mqueue,synw/django-mqueue,synw/django-mqueue
__version__ = '0.5.5' default_app_config = 'mqueue.apps.MqueueConfig' + + import sys + reload(sys) + sys.setdefaultencoding("utf-8")
Set default encoding to fix unicode errors
## Code Before: __version__ = '0.5.5' default_app_config = 'mqueue.apps.MqueueConfig' ## Instruction: Set default encoding to fix unicode errors ## Code After: __version__ = '0.5.5' default_app_config = 'mqueue.apps.MqueueConfig' import sys reload(sys) sys.setdefaultencoding("utf-8")
__version__ = '0.5.5' default_app_config = 'mqueue.apps.MqueueConfig' + import sys + reload(sys) + sys.setdefaultencoding("utf-8") +
49e95022577eb40bcf9e1d1c9f95be7269fd0e3b
scripts/update_acq_stats.py
scripts/update_acq_stats.py
from mica.stats import update_acq_stats update_acq_stats.main() import os table_file = mica.stats.acq_stats.table_file file_stat = os.stat(table_file) if file_stat.st_size > 50e6: print(""" Warning: {tfile} is larger than 50MB and may need Warning: to be manually repacked (i.e.): Warning: Warning: ptrepack --chun...
import os from mica.stats import update_acq_stats import mica.stats.acq_stats update_acq_stats.main() table_file = mica.stats.acq_stats.TABLE_FILE file_stat = os.stat(table_file) if file_stat.st_size > 50e6: print(""" Warning: {tfile} is larger than 50MB and may need Warning: to be manually repacked (i.e.): Warni...
Fix reference to acq table file in script
Fix reference to acq table file in script
Python
bsd-3-clause
sot/mica,sot/mica
- + import os from mica.stats import update_acq_stats + import mica.stats.acq_stats update_acq_stats.main() - import os + - table_file = mica.stats.acq_stats.table_file + table_file = mica.stats.acq_stats.TABLE_FILE file_stat = os.stat(table_file) if file_stat.st_size > 50e6: print(""" Warning: {tf...
Fix reference to acq table file in script
## Code Before: from mica.stats import update_acq_stats update_acq_stats.main() import os table_file = mica.stats.acq_stats.table_file file_stat = os.stat(table_file) if file_stat.st_size > 50e6: print(""" Warning: {tfile} is larger than 50MB and may need Warning: to be manually repacked (i.e.): Warning: Warning:...
- + import os from mica.stats import update_acq_stats + import mica.stats.acq_stats update_acq_stats.main() - import os + - table_file = mica.stats.acq_stats.table_file ? ^^^^^ ^^^^ + table_file = mica.stats.acq_stats.TABLE_FILE ? ^^^^^ ^^^^ ...
9fb89f885dd26b530b4cc95427373f06ddc7d13d
emptiness.py
emptiness.py
import argparse import requests import timetable if __name__ == '__main__': parser = argparse.ArgumentParser() parser.add_argument("-d", "--day", default='', required=True, help="Day to check the timetable on. eg: Thursday") parser.add_argument("-t", "--time", default='', required=True, help="The time the block ...
import argparse import requests import timetable import datetime import time if __name__ == '__main__': parser = argparse.ArgumentParser() parser.add_argument("-d", "--day", default='', required=False, help="Day to check the timetable on. eg: Thursday") parser.add_argument("-t", "--time", default='', required=Fa...
Use current time if no arguments given
Use current time if no arguments given
Python
mit
egeldenhuys/emptiness,egeldenhuys/emptiness,egeldenhuys/emptiness
import argparse import requests import timetable + import datetime + import time if __name__ == '__main__': parser = argparse.ArgumentParser() - parser.add_argument("-d", "--day", default='', required=True, help="Day to check the timetable on. eg: Thursday") + parser.add_argument("-d", "--day",...
Use current time if no arguments given
## Code Before: import argparse import requests import timetable if __name__ == '__main__': parser = argparse.ArgumentParser() parser.add_argument("-d", "--day", default='', required=True, help="Day to check the timetable on. eg: Thursday") parser.add_argument("-t", "--time", default='', required=True, help="The...
import argparse import requests import timetable + import datetime + import time if __name__ == '__main__': parser = argparse.ArgumentParser() - parser.add_argument("-d", "--day", default='', required=True, help="Day to check the timetable on. eg: Thursday") ? ...
592ffbcd7fbbc29bfd377b5abadb39aa29f1c88d
foyer/tests/conftest.py
foyer/tests/conftest.py
import pytest @pytest.fixture(scope="session") def initdir(tmpdir): tmpdir.chdir()
import pytest @pytest.fixture(autouse=True) def initdir(tmpdir): tmpdir.chdir()
Switch from scope="session" to autouse=True
Switch from scope="session" to autouse=True
Python
mit
iModels/foyer,mosdef-hub/foyer,mosdef-hub/foyer,iModels/foyer
import pytest - @pytest.fixture(scope="session") + @pytest.fixture(autouse=True) def initdir(tmpdir): tmpdir.chdir()
Switch from scope="session" to autouse=True
## Code Before: import pytest @pytest.fixture(scope="session") def initdir(tmpdir): tmpdir.chdir() ## Instruction: Switch from scope="session" to autouse=True ## Code After: import pytest @pytest.fixture(autouse=True) def initdir(tmpdir): tmpdir.chdir()
import pytest - @pytest.fixture(scope="session") + @pytest.fixture(autouse=True) def initdir(tmpdir): tmpdir.chdir()
1e90db8de39bd8c4b1a4d58148b991af8b5c32dd
storage/models/fighter.py
storage/models/fighter.py
from storage.models.base import * class Fighter(Base): __tablename__ = 'fighters' id = Column(Integer, primary_key=True) ref = Column(String(STR_SIZE), unique=True, nullable=False) name = Column(String(STR_SIZE), nullable=False) country = Column(String(STR_SIZE)) city = Column(String(STR_SIZE...
from storage.models.base import * class Fighter(Base): __tablename__ = 'fighters' id = Column(Integer, primary_key=True) ref = Column(String(STR_SIZE), unique=True, nullable=False) name = Column(String(STR_SIZE), nullable=False) country = Column(String(STR_SIZE)) city = Column(String(STR_SIZE...
Add restriction for specialization string in db
Add restriction for specialization string in db
Python
apache-2.0
Some1Nebo/ufcpy
from storage.models.base import * class Fighter(Base): __tablename__ = 'fighters' id = Column(Integer, primary_key=True) ref = Column(String(STR_SIZE), unique=True, nullable=False) name = Column(String(STR_SIZE), nullable=False) country = Column(String(STR_SIZE)) city = ...
Add restriction for specialization string in db
## Code Before: from storage.models.base import * class Fighter(Base): __tablename__ = 'fighters' id = Column(Integer, primary_key=True) ref = Column(String(STR_SIZE), unique=True, nullable=False) name = Column(String(STR_SIZE), nullable=False) country = Column(String(STR_SIZE)) city = Column...
from storage.models.base import * class Fighter(Base): __tablename__ = 'fighters' id = Column(Integer, primary_key=True) ref = Column(String(STR_SIZE), unique=True, nullable=False) name = Column(String(STR_SIZE), nullable=False) country = Column(String(STR_SIZE)) city = ...
d70ccd856bb4ddb061ff608716ef15f778380d62
gnsq/stream/defalte.py
gnsq/stream/defalte.py
from __future__ import absolute_import import zlib from .compression import CompressionSocket class DefalteSocket(CompressionSocket): def __init__(self, socket, level): self._decompressor = zlib.decompressobj(level) self._compressor = zlib.compressobj(level) super(DefalteSocket, self).__i...
from __future__ import absolute_import import zlib from .compression import CompressionSocket class DefalteSocket(CompressionSocket): def __init__(self, socket, level): wbits = -zlib.MAX_WBITS self._decompressor = zlib.decompressobj(wbits) self._compressor = zlib.compressobj(level, zlib.D...
Set correct waits for deflate.
Set correct waits for deflate.
Python
bsd-3-clause
wtolson/gnsq,hiringsolved/gnsq,wtolson/gnsq
from __future__ import absolute_import import zlib from .compression import CompressionSocket class DefalteSocket(CompressionSocket): def __init__(self, socket, level): + wbits = -zlib.MAX_WBITS - self._decompressor = zlib.decompressobj(level) + self._decompressor = zlib.d...
Set correct waits for deflate.
## Code Before: from __future__ import absolute_import import zlib from .compression import CompressionSocket class DefalteSocket(CompressionSocket): def __init__(self, socket, level): self._decompressor = zlib.decompressobj(level) self._compressor = zlib.compressobj(level) super(DefalteS...
from __future__ import absolute_import import zlib from .compression import CompressionSocket class DefalteSocket(CompressionSocket): def __init__(self, socket, level): + wbits = -zlib.MAX_WBITS - self._decompressor = zlib.decompressobj(level) ? ...
eaa13f9005a8aaf8c748a98de697b03eee9e675b
salt/client/netapi.py
salt/client/netapi.py
''' The main entry point for salt-api ''' from __future__ import absolute_import # Import python libs import logging # Import salt-api libs import salt.loader import salt.utils.process logger = logging.getLogger(__name__) class NetapiClient(object): ''' Start each netapi module that is configured to run ...
''' The main entry point for salt-api ''' from __future__ import absolute_import # Import python libs import logging # Import salt-api libs import salt.loader import salt.utils.process logger = logging.getLogger(__name__) class NetapiClient(object): ''' Start each netapi module that is configured to run ...
Add log error if we run salt-api w/ no config
Add log error if we run salt-api w/ no config Currently, the salt-api script will exit with no error or hint of why it failed if there is no netapi module configured. Added a short line if we find no api modules to start, warning the user that the config may be missing. Fixes #28240
Python
apache-2.0
saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt,saltstack/salt
''' The main entry point for salt-api ''' from __future__ import absolute_import # Import python libs import logging # Import salt-api libs import salt.loader import salt.utils.process logger = logging.getLogger(__name__) class NetapiClient(object): ''' Start each netapi mo...
Add log error if we run salt-api w/ no config
## Code Before: ''' The main entry point for salt-api ''' from __future__ import absolute_import # Import python libs import logging # Import salt-api libs import salt.loader import salt.utils.process logger = logging.getLogger(__name__) class NetapiClient(object): ''' Start each netapi module that is confi...
''' The main entry point for salt-api ''' from __future__ import absolute_import # Import python libs import logging # Import salt-api libs import salt.loader import salt.utils.process logger = logging.getLogger(__name__) class NetapiClient(object): ''' Start each netapi mo...
888584a49e697551c4f680cc8651be2fe80fc65d
configgen/generators/ppsspp/ppssppGenerator.py
configgen/generators/ppsspp/ppssppGenerator.py
import Command #~ import reicastControllers import recalboxFiles from generators.Generator import Generator import ppssppControllers import shutil import os.path import ConfigParser class PPSSPPGenerator(Generator): # Main entry of the module # Configure fba and return a command def generate(self, system,...
import Command #~ import reicastControllers import recalboxFiles from generators.Generator import Generator import ppssppControllers import shutil import os.path import ConfigParser class PPSSPPGenerator(Generator): # Main entry of the module # Configure fba and return a command def generate(self, system,...
Remove a bad typo from reicast
Remove a bad typo from reicast
Python
mit
nadenislamarre/recalbox-configgen,recalbox/recalbox-configgen,digitalLumberjack/recalbox-configgen
import Command #~ import reicastControllers import recalboxFiles from generators.Generator import Generator import ppssppControllers import shutil import os.path import ConfigParser class PPSSPPGenerator(Generator): # Main entry of the module # Configure fba and return a command ...
Remove a bad typo from reicast
## Code Before: import Command #~ import reicastControllers import recalboxFiles from generators.Generator import Generator import ppssppControllers import shutil import os.path import ConfigParser class PPSSPPGenerator(Generator): # Main entry of the module # Configure fba and return a command def genera...
import Command #~ import reicastControllers import recalboxFiles from generators.Generator import Generator import ppssppControllers import shutil import os.path import ConfigParser class PPSSPPGenerator(Generator): # Main entry of the module # Configure fba and return a command ...
cc09da295d61965af1552b35b7ece0caf4e5a399
accountant/interface/forms.py
accountant/interface/forms.py
from django import forms from django.core.exceptions import NON_FIELD_ERRORS from django.utils.translation import ugettext_lazy as _ from core import models DUPLICATE_PLAYER_ERROR = \ _('There is already a player with this name in your game') class CreateGameForm(forms.Form): bank_cash = forms.IntegerField(re...
from django import forms from django.core.exceptions import NON_FIELD_ERRORS from django.utils.translation import ugettext_lazy as _ from core import models DUPLICATE_PLAYER_ERROR = \ _('There is already a player with this name in your game') class CreateGameForm(forms.Form): bank_cash = forms.IntegerField(re...
Hide Game ID input since it is automatically set
Hide Game ID input since it is automatically set
Python
mit
XeryusTC/18xx-accountant,XeryusTC/18xx-accountant,XeryusTC/18xx-accountant,XeryusTC/18xx-accountant,XeryusTC/18xx-accountant
from django import forms from django.core.exceptions import NON_FIELD_ERRORS from django.utils.translation import ugettext_lazy as _ from core import models DUPLICATE_PLAYER_ERROR = \ _('There is already a player with this name in your game') class CreateGameForm(forms.Form): bank_cash = f...
Hide Game ID input since it is automatically set
## Code Before: from django import forms from django.core.exceptions import NON_FIELD_ERRORS from django.utils.translation import ugettext_lazy as _ from core import models DUPLICATE_PLAYER_ERROR = \ _('There is already a player with this name in your game') class CreateGameForm(forms.Form): bank_cash = forms...
from django import forms from django.core.exceptions import NON_FIELD_ERRORS from django.utils.translation import ugettext_lazy as _ from core import models DUPLICATE_PLAYER_ERROR = \ _('There is already a player with this name in your game') class CreateGameForm(forms.Form): bank_cash = f...
2f635e890414f777fbe3ddde1aea74ab13558313
llvmlite/tests/test_dylib.py
llvmlite/tests/test_dylib.py
import unittest from . import TestCase from llvmlite import binding as llvm from llvmlite.binding import dylib import platform class TestDylib(TestCase): def setUp(self): llvm.initialize() llvm.initialize_native_target() llvm.initialize_native_asmprinter() def test_bad_library(self)...
from . import TestCase from llvmlite import binding as llvm from llvmlite.binding import dylib import platform from ctypes.util import find_library import unittest @unittest.skipUnless(platform.system() in {"Linux", "Darwin"}, "Unsupport test for current OS") class TestDylib(TestCase): def setUp(self): ll...
Add tests to check loading library.
Add tests to check loading library.
Python
bsd-2-clause
m-labs/llvmlite,pitrou/llvmlite,ssarangi/llvmlite,m-labs/llvmlite,markdewing/llvmlite,pitrou/llvmlite,numba/llvmlite,markdewing/llvmlite,sklam/llvmlite,sklam/llvmlite,pitrou/llvmlite,numba/llvmlite,ssarangi/llvmlite,markdewing/llvmlite,squisher/llvmlite,ssarangi/llvmlite,m-labs/llvmlite,numba/llvmlite,numba/llvmlite,sq...
- import unittest from . import TestCase from llvmlite import binding as llvm from llvmlite.binding import dylib import platform + from ctypes.util import find_library + import unittest - + @unittest.skipUnless(platform.system() in {"Linux", "Darwin"}, "Unsupport test for current OS") class TestDylib(T...
Add tests to check loading library.
## Code Before: import unittest from . import TestCase from llvmlite import binding as llvm from llvmlite.binding import dylib import platform class TestDylib(TestCase): def setUp(self): llvm.initialize() llvm.initialize_native_target() llvm.initialize_native_asmprinter() def test_b...
- import unittest from . import TestCase from llvmlite import binding as llvm from llvmlite.binding import dylib import platform + from ctypes.util import find_library + import unittest - + @unittest.skipUnless(platform.system() in {"Linux", "Darwin"}, "Unsupport test for current OS") class TestDylib(T...
77a5ecc7c406e4a6acf814a2f0381dc605e0d14c
leds/led_dance.py
leds/led_dance.py
import pyb def led_dance(delay): dots = {} control = pyb.Switch(1) while True: if not control.value(): dots[pyb.millis() % 25] = 16 for d in dots: pyb.pixel(d, dots[d]) if dots[d] == 0: del(dots[d]) else: dots[...
import microbit def led_dance(delay): dots = [ [0]*5, [0]*5, [0]*5, [0]*5, [0]*5 ] microbit.display.set_display_mode(1) while True: dots[microbit.random(5)][microbit.random(5)] = 128 for i in range(5): for j in range(5): microbit.display.image.set_pixel_value(i,...
Update for new version of micropython for microbit
Update for new version of micropython for microbit
Python
mit
jrmhaig/microbit_playground
- import pyb + import microbit def led_dance(delay): - dots = {} - control = pyb.Switch(1) + dots = [ [0]*5, [0]*5, [0]*5, [0]*5, [0]*5 ] + microbit.display.set_display_mode(1) while True: + dots[microbit.random(5)][microbit.random(5)] = 128 + for i in range(5): + ...
Update for new version of micropython for microbit
## Code Before: import pyb def led_dance(delay): dots = {} control = pyb.Switch(1) while True: if not control.value(): dots[pyb.millis() % 25] = 16 for d in dots: pyb.pixel(d, dots[d]) if dots[d] == 0: del(dots[d]) else: ...
- import pyb + import microbit def led_dance(delay): - dots = {} - control = pyb.Switch(1) + dots = [ [0]*5, [0]*5, [0]*5, [0]*5, [0]*5 ] + microbit.display.set_display_mode(1) while True: + dots[microbit.random(5)][microbit.random(5)] = 128 + for i in range(5): + ...
606b2b6c84e9f9f67606a4d7e521cf4805855a98
migrations/versions/0311_populate_returned_letters.py
migrations/versions/0311_populate_returned_letters.py
from alembic import op from app.dao.returned_letters_dao import insert_or_update_returned_letters revision = '0311_populate_returned_letters' down_revision = '0310_returned_letters_table' def upgrade(): conn = op.get_bind() sql = """ select id, service_id, reference from notification_histor...
from alembic import op revision = '0311_populate_returned_letters' down_revision = '0310_returned_letters_table' def upgrade(): conn = op.get_bind() sql = """ select id, service_id, reference, updated_at from notification_history where notification_type = 'letter' and notific...
Change the insert to use updated_at as the reported_at date
Change the insert to use updated_at as the reported_at date
Python
mit
alphagov/notifications-api,alphagov/notifications-api
from alembic import op - - from app.dao.returned_letters_dao import insert_or_update_returned_letters revision = '0311_populate_returned_letters' down_revision = '0310_returned_letters_table' def upgrade(): conn = op.get_bind() sql = """ - select id, service_id, reference + ...
Change the insert to use updated_at as the reported_at date
## Code Before: from alembic import op from app.dao.returned_letters_dao import insert_or_update_returned_letters revision = '0311_populate_returned_letters' down_revision = '0310_returned_letters_table' def upgrade(): conn = op.get_bind() sql = """ select id, service_id, reference from not...
from alembic import op - - from app.dao.returned_letters_dao import insert_or_update_returned_letters revision = '0311_populate_returned_letters' down_revision = '0310_returned_letters_table' def upgrade(): conn = op.get_bind() sql = """ - select id, service_id, reference + ...
7a21d2bccbcff2eb6a8b7cfd00c38a28553c0bcd
gratipay/models/country.py
gratipay/models/country.py
from __future__ import absolute_import, division, print_function, unicode_literals from postgres.orm import Model class Country(Model): """Represent country records from our database (read-only). :var int id: the record's primary key in our ``countries`` table :var unicode code: the country's `ISO 3166-...
from __future__ import absolute_import, division, print_function, unicode_literals from postgres.orm import Model class Country(Model): """Represent country records from our database (read-only). :var int id: the record's primary key in our ``countries`` table :var unicode code: the country's `ISO 3166-...
Add a helper to Country; should go upstream prolly
Add a helper to Country; should go upstream prolly
Python
mit
gratipay/gratipay.com,gratipay/gratipay.com,gratipay/gratipay.com,gratipay/gratipay.com
from __future__ import absolute_import, division, print_function, unicode_literals from postgres.orm import Model class Country(Model): """Represent country records from our database (read-only). :var int id: the record's primary key in our ``countries`` table :var unicode code: the ...
Add a helper to Country; should go upstream prolly
## Code Before: from __future__ import absolute_import, division, print_function, unicode_literals from postgres.orm import Model class Country(Model): """Represent country records from our database (read-only). :var int id: the record's primary key in our ``countries`` table :var unicode code: the coun...
from __future__ import absolute_import, division, print_function, unicode_literals from postgres.orm import Model class Country(Model): """Represent country records from our database (read-only). :var int id: the record's primary key in our ``countries`` table :var unicode code: the ...
0eb20c8025a838d93a5854442640550d5bf05b0b
settings.py
settings.py
# Replace the following lines with client IDs obtained from the APIs # Console or Cloud Console. WEB_CLIENT_ID = '757224007118-0lblpo8abqeantp8mvckmabupik9edk4.apps.googleusercontent.com' ANDROID_CLIENT_ID = 'replace with Android client ID' IOS_CLIENT_ID = 'replace with iOS client ID' ANDROID_AUDIENCE = WEB_CLIENT_ID
# Replace the following lines with client IDs obtained from the APIs # Console or Cloud Console. WEB_CLIENT_ID = '757224007118-0lblpo8abqeantp8mvckmabupik9edk4.apps.googleusercontent.com' ANDROID_CLIENT_ID = '757224007118-dpqfa375ra8rgbpslig7beh4jb6qd03s.apps.googleusercontent.com' IOS_CLIENT_ID = '757224007118-nfgr65...
Add android and ios client IDs
Add android and ios client IDs
Python
apache-2.0
elbernante/conference-central,elbernante/conference-central,elbernante/conference-central
# Replace the following lines with client IDs obtained from the APIs # Console or Cloud Console. WEB_CLIENT_ID = '757224007118-0lblpo8abqeantp8mvckmabupik9edk4.apps.googleusercontent.com' - ANDROID_CLIENT_ID = 'replace with Android client ID' - IOS_CLIENT_ID = 'replace with iOS client ID' + ANDROID_CLIENT_ID =...
Add android and ios client IDs
## Code Before: # Replace the following lines with client IDs obtained from the APIs # Console or Cloud Console. WEB_CLIENT_ID = '757224007118-0lblpo8abqeantp8mvckmabupik9edk4.apps.googleusercontent.com' ANDROID_CLIENT_ID = 'replace with Android client ID' IOS_CLIENT_ID = 'replace with iOS client ID' ANDROID_AUDIENCE ...
# Replace the following lines with client IDs obtained from the APIs # Console or Cloud Console. WEB_CLIENT_ID = '757224007118-0lblpo8abqeantp8mvckmabupik9edk4.apps.googleusercontent.com' - ANDROID_CLIENT_ID = 'replace with Android client ID' - IOS_CLIENT_ID = 'replace with iOS client ID' + ANDROID_CLIENT_ID =...
2eb07ae9b98c36dc94e143003a7c44c7fbfb54f7
stronghold/middleware.py
stronghold/middleware.py
from django.contrib.auth.decorators import login_required from stronghold import conf, utils class LoginRequiredMiddleware(object): """ Force all views to use login required View is deemed to be public if the @public decorator is applied to the view View is also deemed to be Public if listed in in d...
from django.contrib.auth.decorators import login_required from stronghold import conf, utils class LoginRequiredMiddleware(object): """ Force all views to use login required View is deemed to be public if the @public decorator is applied to the view View is also deemed to be Public if listed in in d...
Refactor away unnecessary multiple return None
Refactor away unnecessary multiple return None
Python
mit
SunilMohanAdapa/django-stronghold,SunilMohanAdapa/django-stronghold,mgrouchy/django-stronghold
from django.contrib.auth.decorators import login_required from stronghold import conf, utils class LoginRequiredMiddleware(object): """ Force all views to use login required View is deemed to be public if the @public decorator is applied to the view View is also deemed to be Pu...
Refactor away unnecessary multiple return None
## Code Before: from django.contrib.auth.decorators import login_required from stronghold import conf, utils class LoginRequiredMiddleware(object): """ Force all views to use login required View is deemed to be public if the @public decorator is applied to the view View is also deemed to be Public i...
from django.contrib.auth.decorators import login_required from stronghold import conf, utils class LoginRequiredMiddleware(object): """ Force all views to use login required View is deemed to be public if the @public decorator is applied to the view View is also deemed to be Pu...
39d0c335759781de8cf1644cdf499588441b434d
tviserrys/urls.py
tviserrys/urls.py
from django.contrib.auth import views as auth_views from django.conf.urls import patterns, include, url from django.conf.urls import url from django.contrib import admin from . import views from tviserrys.settings import MEDIA_ROOT urlpatterns = [ url(r'^$', views.IndexView.as_view(), name='index'), url(r'^adm...
from django.contrib.auth import views as auth_views from django.conf.urls import patterns, include, url from django.conf.urls import url from django.contrib import admin from . import views from tviserrys.settings import MEDIA_ROOT urlpatterns = [ url(r'^$', views.IndexView.as_view(), name='index'), url(r'^tvi...
Add Tviit and profile url patterns
Add Tviit and profile url patterns
Python
mit
DeWaster/Tviserrys,DeWaster/Tviserrys
from django.contrib.auth import views as auth_views from django.conf.urls import patterns, include, url from django.conf.urls import url from django.contrib import admin from . import views from tviserrys.settings import MEDIA_ROOT urlpatterns = [ url(r'^$', views.IndexView.as_view(), name='index...
Add Tviit and profile url patterns
## Code Before: from django.contrib.auth import views as auth_views from django.conf.urls import patterns, include, url from django.conf.urls import url from django.contrib import admin from . import views from tviserrys.settings import MEDIA_ROOT urlpatterns = [ url(r'^$', views.IndexView.as_view(), name='index')...
from django.contrib.auth import views as auth_views from django.conf.urls import patterns, include, url from django.conf.urls import url from django.contrib import admin from . import views from tviserrys.settings import MEDIA_ROOT urlpatterns = [ url(r'^$', views.IndexView.as_view(), name='index...
df8ae0415f9bf10c04472fb3009e91d7c3d7e24f
teuthology/sentry.py
teuthology/sentry.py
from raven import Client client = None def get_client(ctx): if client: return client dsn = ctx.teuthology_config.get('sentry_dsn') if dsn: client = Client(dsn=dsn) return client
from raven import Client client = None def get_client(ctx): global client if client: return client dsn = ctx.teuthology_config.get('sentry_dsn') if dsn: client = Client(dsn=dsn) return client
Make client a global variable
Make client a global variable
Python
mit
robbat2/teuthology,ceph/teuthology,tchaikov/teuthology,zhouyuan/teuthology,dmick/teuthology,michaelsevilla/teuthology,dreamhost/teuthology,SUSE/teuthology,t-miyamae/teuthology,caibo2014/teuthology,yghannam/teuthology,SUSE/teuthology,SUSE/teuthology,tchaikov/teuthology,michaelsevilla/teuthology,dmick/teuthology,ktdreyer...
from raven import Client client = None def get_client(ctx): + global client if client: return client dsn = ctx.teuthology_config.get('sentry_dsn') if dsn: client = Client(dsn=dsn) return client
Make client a global variable
## Code Before: from raven import Client client = None def get_client(ctx): if client: return client dsn = ctx.teuthology_config.get('sentry_dsn') if dsn: client = Client(dsn=dsn) return client ## Instruction: Make client a global variable ## Code After: from raven import Client ...
from raven import Client client = None def get_client(ctx): + global client if client: return client dsn = ctx.teuthology_config.get('sentry_dsn') if dsn: client = Client(dsn=dsn) return client
76166f243b9f5f21582c95a843ddfa174ded8602
PyFVCOM/__init__.py
PyFVCOM/__init__.py
__version__ = '1.6.2' __author__ = 'Pierre Cazenave' __credits__ = ['Pierre Cazenave'] __license__ = 'MIT' __maintainer__ = 'Pierre Cazenave' __email__ = 'pica@pml.ac.uk' import inspect from warnings import warn # Import everything! from PyFVCOM import buoy_tools from PyFVCOM import cst_tools from PyFVCOM import ctd...
__version__ = '1.6.2' __author__ = 'Pierre Cazenave' __credits__ = ['Pierre Cazenave'] __license__ = 'MIT' __maintainer__ = 'Pierre Cazenave' __email__ = 'pica@pml.ac.uk' import inspect from warnings import warn # Import everything! from PyFVCOM import buoy_tools from PyFVCOM import cst_tools from PyFVCOM import ctd...
Put things in in alphabetical order.
Put things in in alphabetical order.
Python
mit
pwcazenave/PyFVCOM
__version__ = '1.6.2' __author__ = 'Pierre Cazenave' __credits__ = ['Pierre Cazenave'] __license__ = 'MIT' __maintainer__ = 'Pierre Cazenave' __email__ = 'pica@pml.ac.uk' import inspect from warnings import warn # Import everything! from PyFVCOM import buoy_tools from PyFVCOM import cst_t...
Put things in in alphabetical order.
## Code Before: __version__ = '1.6.2' __author__ = 'Pierre Cazenave' __credits__ = ['Pierre Cazenave'] __license__ = 'MIT' __maintainer__ = 'Pierre Cazenave' __email__ = 'pica@pml.ac.uk' import inspect from warnings import warn # Import everything! from PyFVCOM import buoy_tools from PyFVCOM import cst_tools from Py...
__version__ = '1.6.2' __author__ = 'Pierre Cazenave' __credits__ = ['Pierre Cazenave'] __license__ = 'MIT' __maintainer__ = 'Pierre Cazenave' __email__ = 'pica@pml.ac.uk' import inspect from warnings import warn # Import everything! from PyFVCOM import buoy_tools from PyFVCOM import cst_t...
c97e5cf11fc21e2ef4ee04779a424e4d6a2b96ae
tools/perf/metrics/__init__.py
tools/perf/metrics/__init__.py
class Metric(object): """Base class for all the metrics that are used by telemetry measurements. The Metric class represents a way of measuring something. Metrics are helper classes used by PageMeasurements. Each PageMeasurement may use multiple metrics; each metric should be focussed on collecting data abo...
class Metric(object): """Base class for all the metrics that are used by telemetry measurements. The Metric class represents a way of measuring something. Metrics are helper classes used by PageMeasurements. Each PageMeasurement may use multiple metrics; each metric should be focussed on collecting data abo...
Add CustomizeBrowserOptions method to Metric base class
Add CustomizeBrowserOptions method to Metric base class BUG=271177 Review URL: https://chromiumcodereview.appspot.com/22938004 git-svn-id: de016e52bd170d2d4f2344f9bf92d50478b649e0@217198 0039d316-1c4b-4281-b951-d872f2087c98
Python
bsd-3-clause
mogoweb/chromium-crosswalk,Just-D/chromium-1,M4sse/chromium.src,dushu1203/chromium.src,Jonekee/chromium.src,Just-D/chromium-1,ChromiumWebApps/chromium,axinging/chromium-crosswalk,bright-sparks/chromium-spacewalk,PeterWangIntel/chromium-crosswalk,crosswalk-project/chromium-crosswalk-efl,markYoungH/chromium.src,M4sse/chr...
class Metric(object): """Base class for all the metrics that are used by telemetry measurements. The Metric class represents a way of measuring something. Metrics are helper classes used by PageMeasurements. Each PageMeasurement may use multiple metrics; each metric should be focussed on collect...
Add CustomizeBrowserOptions method to Metric base class
## Code Before: class Metric(object): """Base class for all the metrics that are used by telemetry measurements. The Metric class represents a way of measuring something. Metrics are helper classes used by PageMeasurements. Each PageMeasurement may use multiple metrics; each metric should be focussed on colle...
class Metric(object): """Base class for all the metrics that are used by telemetry measurements. The Metric class represents a way of measuring something. Metrics are helper classes used by PageMeasurements. Each PageMeasurement may use multiple metrics; each metric should be focussed on collect...
a8601d8a17c9ba8e87b8336870e0d52f79e0ffa2
indra/tests/test_omnipath.py
indra/tests/test_omnipath.py
from __future__ import unicode_literals from builtins import dict, str from indra.statements import Phosphorylation from indra.databases import omnipath as op def test_query_ptms(): stmts = op.get_ptms(['Q13873']) assert len(stmts) == 1 assert isinstance(stmts[0], Phosphorylation) assert stmts[0].enz.n...
import requests from indra.sources.omnipath import OmniPathModificationProcessor,\ OmniPathLiganReceptorProcessor from indra.sources.omnipath.api import op_url from indra.statements import Agent, Phosphorylation from indra.preassembler.grounding_mapper import GroundingMapper BRAF_UPID = 'P15056' JAK2_UPID = 'O6067...
Update imports, test general web api
Update imports, test general web api
Python
bsd-2-clause
johnbachman/indra,johnbachman/indra,johnbachman/belpy,johnbachman/indra,johnbachman/belpy,sorgerlab/belpy,sorgerlab/indra,sorgerlab/indra,bgyori/indra,johnbachman/belpy,bgyori/indra,sorgerlab/belpy,bgyori/indra,sorgerlab/indra,sorgerlab/belpy
- from __future__ import unicode_literals - from builtins import dict, str + import requests + from indra.sources.omnipath import OmniPathModificationProcessor,\ + OmniPathLiganReceptorProcessor + from indra.sources.omnipath.api import op_url - from indra.statements import Phosphorylation + from indra.statements im...
Update imports, test general web api
## Code Before: from __future__ import unicode_literals from builtins import dict, str from indra.statements import Phosphorylation from indra.databases import omnipath as op def test_query_ptms(): stmts = op.get_ptms(['Q13873']) assert len(stmts) == 1 assert isinstance(stmts[0], Phosphorylation) asser...
- from __future__ import unicode_literals - from builtins import dict, str + import requests + from indra.sources.omnipath import OmniPathModificationProcessor,\ + OmniPathLiganReceptorProcessor + from indra.sources.omnipath.api import op_url - from indra.statements import Phosphorylation + from indra.statements im...
ba23f58f7359b943d8d8ae7f05e15419c6918c6f
test/blacklist.py
test/blacklist.py
blacklist = {}
blacklist = {'BasicExprCommandsTestCase.test_evaluate_expression_python': 'Crashed while running the entire test suite with CC=clang' # To reproduce the crash: CC=clang ./dotest.py -v -w 2> ~/Developer/Log/lldbtest.log # The clang version used is clang-126. # Two radars filed for...
Add an entry for test case BasicExprCommandsTestCase.test_evaluate_expression_python, due to crashes while running the entire test suite with clang-126.
Add an entry for test case BasicExprCommandsTestCase.test_evaluate_expression_python, due to crashes while running the entire test suite with clang-126. To reproduce: CC=clang ./dotest.py -v -w 2> ~/Developer/Log/lldbtest.log To skip this test case: CC=clang ./dotest.py -b blacklist.py -v -w 2> ~/Developer/Log/lldb...
Python
apache-2.0
llvm-mirror/lldb,apple/swift-lldb,apple/swift-lldb,apple/swift-lldb,apple/swift-lldb,apple/swift-lldb,llvm-mirror/lldb,apple/swift-lldb,llvm-mirror/lldb,llvm-mirror/lldb,llvm-mirror/lldb
- blacklist = {} + blacklist = {'BasicExprCommandsTestCase.test_evaluate_expression_python': 'Crashed while running the entire test suite with CC=clang' + # To reproduce the crash: CC=clang ./dotest.py -v -w 2> ~/Developer/Log/lldbtest.log + # The clang version used is clang-126. + ...
Add an entry for test case BasicExprCommandsTestCase.test_evaluate_expression_python, due to crashes while running the entire test suite with clang-126.
## Code Before: blacklist = {} ## Instruction: Add an entry for test case BasicExprCommandsTestCase.test_evaluate_expression_python, due to crashes while running the entire test suite with clang-126. ## Code After: blacklist = {'BasicExprCommandsTestCase.test_evaluate_expression_python': 'Crashed while running the e...
- blacklist = {} + blacklist = {'BasicExprCommandsTestCase.test_evaluate_expression_python': 'Crashed while running the entire test suite with CC=clang' + # To reproduce the crash: CC=clang ./dotest.py -v -w 2> ~/Developer/Log/lldbtest.log + # The clang version used is clang-126. + ...
3d1521892ba17120ca4461335713b9d2254311fe
marble/tests/test_clustering.py
marble/tests/test_clustering.py
""" Tests for the clustering computation """ from nose.tools import * import marble as mb # Test c = 0 in the checkerboard case # Test c = 1 in the fully clustered case # Test an intermediate situation with known result
""" Tests for the clustering computation """ from nose.tools import * import itertools from shapely.geometry import Polygon import marble as mb # # Synthetic data for tests # def grid(): """ Areal units arranged in a grid """ au = [i*3+j for i,j in itertools.product(range(3), repeat=2)] units = {a:Polygo...
Add tests for the clustering of cities
Add tests for the clustering of cities
Python
bsd-3-clause
walkerke/marble,scities/marble
""" Tests for the clustering computation """ from nose.tools import * + import itertools + from shapely.geometry import Polygon import marble as mb - # Test c = 0 in the checkerboard case - # Test c = 1 in the fully clustered case - # Test an intermediate situation with known result + # + # Synthetic data f...
Add tests for the clustering of cities
## Code Before: """ Tests for the clustering computation """ from nose.tools import * import marble as mb # Test c = 0 in the checkerboard case # Test c = 1 in the fully clustered case # Test an intermediate situation with known result ## Instruction: Add tests for the clustering of cities ## Code After: """ Tests ...
""" Tests for the clustering computation """ from nose.tools import * + import itertools + from shapely.geometry import Polygon import marble as mb - # Test c = 0 in the checkerboard case - # Test c = 1 in the fully clustered case - # Test an intermediate situation with known result + # + # Synthetic data f...
6dfed291a253174672d7003700ab770aabcacae4
backend/breach/models/__init__.py
backend/breach/models/__init__.py
from .victim import Victim from .target import Target from .round import Round from .sampleset import SampleSet
__all__ = ['victim', 'target', 'round', 'sampleset'] from .victim import Victim from .target import Target from .round import Round from .sampleset import SampleSet
Add __all__ to models init file
Add __all__ to models init file
Python
mit
dimriou/rupture,esarafianou/rupture,dimriou/rupture,dimkarakostas/rupture,dionyziz/rupture,dimkarakostas/rupture,esarafianou/rupture,esarafianou/rupture,dimkarakostas/rupture,dimriou/rupture,dionyziz/rupture,esarafianou/rupture,dionyziz/rupture,dimkarakostas/rupture,dionyziz/rupture,dimriou/rupture,dimriou/rupture,dion...
+ __all__ = ['victim', 'target', 'round', 'sampleset'] from .victim import Victim from .target import Target from .round import Round from .sampleset import SampleSet
Add __all__ to models init file
## Code Before: from .victim import Victim from .target import Target from .round import Round from .sampleset import SampleSet ## Instruction: Add __all__ to models init file ## Code After: __all__ = ['victim', 'target', 'round', 'sampleset'] from .victim import Victim from .target import Target from .round import Ro...
+ __all__ = ['victim', 'target', 'round', 'sampleset'] from .victim import Victim from .target import Target from .round import Round from .sampleset import SampleSet
895d51105cd51387e3ac5db595333ff794f3e2a7
yotta/lib/ordered_json.py
yotta/lib/ordered_json.py
import json import os import stat from collections import OrderedDict # provide read & write methods for json files that maintain the order of # dictionary keys, and indent consistently # Internals def load(path): with open(path, 'r') as f: # using an ordered dictionary for objects so that we preserve the...
import json import os import stat from collections import OrderedDict # provide read & write methods for json files that maintain the order of # dictionary keys, and indent consistently # Internals def load(path): with open(path, 'r') as f: # using an ordered dictionary for objects so that we preserve the...
Add a newline at the end of json files when writing them.
Add a newline at the end of json files when writing them. This fixes the really irritating ping-pong of newline/nonewline when editing json files with an editor, and with `yotta version` commands.
Python
apache-2.0
BlackstoneEngineering/yotta,autopulated/yotta,ARMmbed/yotta,stevenewey/yotta,ARMmbed/yotta,autopulated/yotta,ntoll/yotta,BlackstoneEngineering/yotta,stevenewey/yotta,eyeye/yotta,ntoll/yotta,eyeye/yotta
import json import os import stat from collections import OrderedDict # provide read & write methods for json files that maintain the order of # dictionary keys, and indent consistently # Internals def load(path): with open(path, 'r') as f: # using an ordered dictionary for objects...
Add a newline at the end of json files when writing them.
## Code Before: import json import os import stat from collections import OrderedDict # provide read & write methods for json files that maintain the order of # dictionary keys, and indent consistently # Internals def load(path): with open(path, 'r') as f: # using an ordered dictionary for objects so that...
import json import os import stat from collections import OrderedDict # provide read & write methods for json files that maintain the order of # dictionary keys, and indent consistently # Internals def load(path): with open(path, 'r') as f: # using an ordered dictionary for objects...
f86c925604356b25a8c5c0c71644f0df6f1b48f8
setup_directory.py
setup_directory.py
from __future__ import division, print_function, absolute_import import argparse import os import subprocess as sp from contextlib import contextmanager import tempfile try: import urllib.request as urllib2 except ImportError: import urllib2 MINICONDA_URL = 'https://repo.continuum.io/miniconda/Miniconda-lates...
from __future__ import division, print_function, absolute_import import argparse import os import subprocess as sp from contextlib import contextmanager import tempfile try: import urllib.request as urllib2 except ImportError: import urllib2 MINICONDA_URL = 'https://repo.continuum.io/miniconda/Miniconda-lates...
Add function to install miniconda
Add function to install miniconda
Python
mit
NGTS/pipeline-output-analysis-setup-script
from __future__ import division, print_function, absolute_import import argparse import os import subprocess as sp from contextlib import contextmanager import tempfile try: import urllib.request as urllib2 except ImportError: import urllib2 MINICONDA_URL = 'https://repo.continuum.io...
Add function to install miniconda
## Code Before: from __future__ import division, print_function, absolute_import import argparse import os import subprocess as sp from contextlib import contextmanager import tempfile try: import urllib.request as urllib2 except ImportError: import urllib2 MINICONDA_URL = 'https://repo.continuum.io/miniconda...
from __future__ import division, print_function, absolute_import import argparse import os import subprocess as sp from contextlib import contextmanager import tempfile try: import urllib.request as urllib2 except ImportError: import urllib2 MINICONDA_URL = 'https://repo.continuum.io...
ebd6d12ca16003e771a7015505be1b42d96483a3
roles/gvl.commandline-utilities/templates/jupyterhub_config.py
roles/gvl.commandline-utilities/templates/jupyterhub_config.py
c.JupyterHub.ip = '127.0.0.1' # The ip for the proxy API handlers c.JupyterHub.proxy_api_ip = '127.0.0.1' # The public facing port of the proxy c.JupyterHub.port = 9510 # The base URL of the entire application c.JupyterHub.base_url = '/jupyterhub' # The ip for this process c.JupyterHub.hub_ip = '127.0.0.1' # put t...
c.JupyterHub.ip = '127.0.0.1' # The ip for the proxy API handlers c.JupyterHub.proxy_api_ip = '127.0.0.1' # The public facing port of the proxy c.JupyterHub.port = 9510 # The base URL of the entire application c.JupyterHub.base_url = '/jupyterhub' # The ip for this process c.JupyterHub.hub_ip = '127.0.0.1' # put t...
Set log level to 'WARN'
Set log level to 'WARN'
Python
mit
gvlproject/gvl_commandline_utilities,nuwang/gvl_commandline_utilities,claresloggett/gvl_commandline_utilities,nuwang/gvl_commandline_utilities,claresloggett/gvl_commandline_utilities,gvlproject/gvl_commandline_utilities
c.JupyterHub.ip = '127.0.0.1' # The ip for the proxy API handlers c.JupyterHub.proxy_api_ip = '127.0.0.1' # The public facing port of the proxy c.JupyterHub.port = 9510 # The base URL of the entire application c.JupyterHub.base_url = '/jupyterhub' # The ip for this process c.JupyterHub.hub...
Set log level to 'WARN'
## Code Before: c.JupyterHub.ip = '127.0.0.1' # The ip for the proxy API handlers c.JupyterHub.proxy_api_ip = '127.0.0.1' # The public facing port of the proxy c.JupyterHub.port = 9510 # The base URL of the entire application c.JupyterHub.base_url = '/jupyterhub' # The ip for this process c.JupyterHub.hub_ip = '127...
c.JupyterHub.ip = '127.0.0.1' # The ip for the proxy API handlers c.JupyterHub.proxy_api_ip = '127.0.0.1' # The public facing port of the proxy c.JupyterHub.port = 9510 # The base URL of the entire application c.JupyterHub.base_url = '/jupyterhub' # The ip for this process c.JupyterHub.hub...
e7b853c667b5785355214380954c83b843c46f05
tests/modules/contrib/test_publicip.py
tests/modules/contrib/test_publicip.py
import pytest from unittest import TestCase, mock import core.config import core.widget import modules.contrib.publicip def build_module(): config = core.config.Config([]) return modules.contrib.publicip.Module(config=config, theme=None) def widget(module): return module.widgets()[0] class PublicIPTest...
import pytest from unittest import TestCase, mock import core.config import core.widget import modules.contrib.publicip def build_module(): config = core.config.Config([]) return modules.contrib.publicip.Module(config=config, theme=None) def widget(module): return module.widgets()[0] class PublicIPTest...
Remove useless mock side effect
Remove useless mock side effect
Python
mit
tobi-wan-kenobi/bumblebee-status,tobi-wan-kenobi/bumblebee-status
import pytest from unittest import TestCase, mock import core.config import core.widget import modules.contrib.publicip def build_module(): config = core.config.Config([]) return modules.contrib.publicip.Module(config=config, theme=None) def widget(module): return module.widg...
Remove useless mock side effect
## Code Before: import pytest from unittest import TestCase, mock import core.config import core.widget import modules.contrib.publicip def build_module(): config = core.config.Config([]) return modules.contrib.publicip.Module(config=config, theme=None) def widget(module): return module.widgets()[0] cl...
import pytest from unittest import TestCase, mock import core.config import core.widget import modules.contrib.publicip def build_module(): config = core.config.Config([]) return modules.contrib.publicip.Module(config=config, theme=None) def widget(module): return module.widg...
a1c87c491bf936d441ef7fd79b531384fa174138
simpleubjson/version.py
simpleubjson/version.py
__version_info__ = (0, 6, 0, 'dev', 0) __version__ = '{version}{tag}{build}'.format( version='.'.join(map(str, __version_info__[:3])), tag='-' + __version_info__[3] if __version_info__[3] else '', build='.' + str(__version_info__[4]) if __version_info__[4] else '' )
__version_info__ = (0, 6, 0, 'dev', 0) __version__ = '%(version)s%(tag)s%(build)s' % { 'version': '.'.join(map(str, __version_info__[:3])), 'tag': '-' + __version_info__[3] if __version_info__[3] else '', 'build': '.' + str(__version_info__[4]) if __version_info__[4] else '' }
Fix compatibility with Python 2.5
Fix compatibility with Python 2.5
Python
bsd-2-clause
kxepal/simpleubjson,brainwater/simpleubjson,samipshah/simpleubjson,498888197/simpleubjson
__version_info__ = (0, 6, 0, 'dev', 0) - __version__ = '{version}{tag}{build}'.format( + __version__ = '%(version)s%(tag)s%(build)s' % { - version='.'.join(map(str, __version_info__[:3])), + 'version': '.'.join(map(str, __version_info__[:3])), - tag='-' + __version_info__[3] if __version_info__[3] else...
Fix compatibility with Python 2.5
## Code Before: __version_info__ = (0, 6, 0, 'dev', 0) __version__ = '{version}{tag}{build}'.format( version='.'.join(map(str, __version_info__[:3])), tag='-' + __version_info__[3] if __version_info__[3] else '', build='.' + str(__version_info__[4]) if __version_info__[4] else '' ) ## Instruction: Fix com...
__version_info__ = (0, 6, 0, 'dev', 0) - __version__ = '{version}{tag}{build}'.format( + __version__ = '%(version)s%(tag)s%(build)s' % { - version='.'.join(map(str, __version_info__[:3])), ? ^ + 'version': '.'.join(map(str, __version_info__[:3])), ? + ^^^ - tag='-' + __version_in...
5547e59360126baa20e1684a22e7f88fdacb530a
s2v2.py
s2v2.py
from s2v1 import * def number_of_records(data_sample): return len(data_sample) number_of_ties = number_of_records(data_from_csv) - 1 # minus header row # print(number_of_ties, "ties in our data sample") def number_of_records2(data_sample): return data_sample.size number_of_ties_my_csv = number_of_records2(my_csv...
from s2v1 import * def number_of_records(data_sample): return len(data_sample) def number_of_records_ignore_header(data_sample, header=True): if header: return len(data_sample) - 1 else: return len(data_sample) number_of_ties = number_of_records(data_from_csv) - 1 # minus header row # print(number_of_ties,...
Create new function for number of records and do a header check
Create new function for number of records and do a header check
Python
mit
alexmilesyounger/ds_basics
from s2v1 import * def number_of_records(data_sample): return len(data_sample) + def number_of_records_ignore_header(data_sample, header=True): + if header: + return len(data_sample) - 1 + else: + return len(data_sample) + number_of_ties = number_of_records(data_from_csv) - 1 # minus header row ...
Create new function for number of records and do a header check
## Code Before: from s2v1 import * def number_of_records(data_sample): return len(data_sample) number_of_ties = number_of_records(data_from_csv) - 1 # minus header row # print(number_of_ties, "ties in our data sample") def number_of_records2(data_sample): return data_sample.size number_of_ties_my_csv = number_of...
from s2v1 import * def number_of_records(data_sample): return len(data_sample) + def number_of_records_ignore_header(data_sample, header=True): + if header: + return len(data_sample) - 1 + else: + return len(data_sample) + number_of_ties = number_of_records(data_from_csv) - 1 # minus header row ...
9bb312c505c2749862372c0ff56ba47e087a9edc
searx/engines/semantic_scholar.py
searx/engines/semantic_scholar.py
from json import dumps, loads search_url = 'https://www.semanticscholar.org/api/1/search' def request(query, params): params['url'] = search_url params['method'] = 'POST' params['headers']['content-type'] = 'application/json' params['data'] = dumps({ "queryString": query, "page": pa...
from json import dumps, loads search_url = 'https://www.semanticscholar.org/api/1/search' def request(query, params): params['url'] = search_url params['method'] = 'POST' params['headers']['content-type'] = 'application/json' params['data'] = dumps({ "queryString": query, "page": pa...
Remove duplicated key from dict in Semantic Scholar
Remove duplicated key from dict in Semantic Scholar
Python
agpl-3.0
dalf/searx,dalf/searx,dalf/searx,dalf/searx
from json import dumps, loads search_url = 'https://www.semanticscholar.org/api/1/search' def request(query, params): params['url'] = search_url params['method'] = 'POST' params['headers']['content-type'] = 'application/json' params['data'] = dumps({ "queryString":...
Remove duplicated key from dict in Semantic Scholar
## Code Before: from json import dumps, loads search_url = 'https://www.semanticscholar.org/api/1/search' def request(query, params): params['url'] = search_url params['method'] = 'POST' params['headers']['content-type'] = 'application/json' params['data'] = dumps({ "queryString": query, ...
from json import dumps, loads search_url = 'https://www.semanticscholar.org/api/1/search' def request(query, params): params['url'] = search_url params['method'] = 'POST' params['headers']['content-type'] = 'application/json' params['data'] = dumps({ "queryString":...
End of preview. Expand in Data Studio
README.md exists but content is empty.
Downloads last month
14