prompt
large_stringlengths
72
9.34k
completion
large_stringlengths
0
7.61k
<|file_name|>test_DataSeriesLoader.py<|end_file_name|><|fim▁begin|>import unittest from datetime import datetime import numpy as np import pandas as pd from excel_helper.helper import DataSeriesLoader class TestDataFrameWithCAGRCalculation(unittest.TestCase): def test_simple_CAGR(self): """ Basi...
test_simple_CAGR_from_pandas
<|file_name|>test_DataSeriesLoader.py<|end_file_name|><|fim▁begin|>import unittest from datetime import datetime import numpy as np import pandas as pd from excel_helper.helper import DataSeriesLoader class TestDataFrameWithCAGRCalculation(unittest.TestCase): def test_simple_CAGR(self): """ Basi...
test_simple_CAGR_mm
<|file_name|>conf.py<|end_file_name|><|fim▁begin|>#!/usr/bin/env python3 # -*- coding: utf-8 -*- # # Backend.AI Library documentation build configuration file, created by # sphinx-quickstart on Tue Mar 1 21:26:20 2016. # # This file is execfile()d with the current directory set to its # containing dir. # # Note that n...
<|file_name|>models.py<|end_file_name|><|fim▁begin|>from django.db import models import warnings from django.utils import timezone import requests from image_cropping import ImageRatioField class CompMember(models.Model): """A member of compsoc""" class Meta: verbose_name = 'CompSoc Member' v...
<|file_name|>models.py<|end_file_name|><|fim▁begin|>from django.db import models import warnings from django.utils import timezone import requests from image_cropping import ImageRatioField class CompMember(models.Model): <|fim_middle|> class Variable(models.Model): ##NOTE: This should not be used anymore ...
"""A member of compsoc""" class Meta: verbose_name = 'CompSoc Member' verbose_name_plural = 'CompSoc Members' index = models.IntegerField(blank=False, help_text="This field is present just for ordering members based on their posts. President = 2, VPs = 1, Gen. Sec. = 0, Everyone else = -1"...
<|file_name|>models.py<|end_file_name|><|fim▁begin|>from django.db import models import warnings from django.utils import timezone import requests from image_cropping import ImageRatioField class CompMember(models.Model): """A member of compsoc""" class Meta: <|fim_middle|> index = models.Integ...
verbose_name = 'CompSoc Member' verbose_name_plural = 'CompSoc Members'
<|file_name|>models.py<|end_file_name|><|fim▁begin|>from django.db import models import warnings from django.utils import timezone import requests from image_cropping import ImageRatioField class CompMember(models.Model): """A member of compsoc""" class Meta: verbose_name = 'CompSoc Member' v...
''' Returns the social_link if present. Otherwise, sends javascript:void(0) ''' if self.social_link == '': return 'javascript:void(0)' else: return self.social_link
<|file_name|>models.py<|end_file_name|><|fim▁begin|>from django.db import models import warnings from django.utils import timezone import requests from image_cropping import ImageRatioField class CompMember(models.Model): """A member of compsoc""" class Meta: verbose_name = 'CompSoc Member' v...
return self.name
<|file_name|>models.py<|end_file_name|><|fim▁begin|>from django.db import models import warnings from django.utils import timezone import requests from image_cropping import ImageRatioField class CompMember(models.Model): """A member of compsoc""" class Meta: verbose_name = 'CompSoc Member' v...
def __str__(self): warnings.warn('''You are using a "General Variable". Stop doing that. This is bad design on Arjoonn's part so don't fall into the same trap. If you are using this for Orfik, that has already been fixed. If you are using this for logos, same thing. Over a f...
<|file_name|>models.py<|end_file_name|><|fim▁begin|>from django.db import models import warnings from django.utils import timezone import requests from image_cropping import ImageRatioField class CompMember(models.Model): """A member of compsoc""" class Meta: verbose_name = 'CompSoc Member' v...
warnings.warn('''You are using a "General Variable". Stop doing that. This is bad design on Arjoonn's part so don't fall into the same trap. If you are using this for Orfik, that has already been fixed. If you are using this for logos, same thing. Over a few cycles this entire t...
<|file_name|>models.py<|end_file_name|><|fim▁begin|>from django.db import models import warnings from django.utils import timezone import requests from image_cropping import ImageRatioField class CompMember(models.Model): """A member of compsoc""" class Meta: verbose_name = 'CompSoc Member' v...
instance.image.delete(False)
<|file_name|>models.py<|end_file_name|><|fim▁begin|>from django.db import models import warnings from django.utils import timezone import requests from image_cropping import ImageRatioField class CompMember(models.Model): """A member of compsoc""" class Meta: verbose_name = 'CompSoc Member' v...
return 'javascript:void(0)'
<|file_name|>models.py<|end_file_name|><|fim▁begin|>from django.db import models import warnings from django.utils import timezone import requests from image_cropping import ImageRatioField class CompMember(models.Model): """A member of compsoc""" class Meta: verbose_name = 'CompSoc Member' v...
return self.social_link
<|file_name|>models.py<|end_file_name|><|fim▁begin|>from django.db import models import warnings from django.utils import timezone import requests from image_cropping import ImageRatioField class CompMember(models.Model): """A member of compsoc""" class Meta: verbose_name = 'CompSoc Member' v...
get_social_link
<|file_name|>models.py<|end_file_name|><|fim▁begin|>from django.db import models import warnings from django.utils import timezone import requests from image_cropping import ImageRatioField class CompMember(models.Model): """A member of compsoc""" class Meta: verbose_name = 'CompSoc Member' v...
__str__
<|file_name|>models.py<|end_file_name|><|fim▁begin|>from django.db import models import warnings from django.utils import timezone import requests from image_cropping import ImageRatioField class CompMember(models.Model): """A member of compsoc""" class Meta: verbose_name = 'CompSoc Member' v...
__str__
<|file_name|>models.py<|end_file_name|><|fim▁begin|>from django.db import models import warnings from django.utils import timezone import requests from image_cropping import ImageRatioField class CompMember(models.Model): """A member of compsoc""" class Meta: verbose_name = 'CompSoc Member' v...
compsoc_member_delete
<|file_name|>test_multiple_dictation.py<|end_file_name|><|fim▁begin|>""" Multiple dictation constructs =============================================================================== This file is a showcase investigating the use and functionality of multiple dictation elements within Dragonfly speech recognition gramm...
extras = [Dictation("dictation1"), Dictation("dictation2"), Dictation("dictation3")]
<|file_name|>test_multiple_dictation.py<|end_file_name|><|fim▁begin|>""" Multiple dictation constructs =============================================================================== This file is a showcase investigating the use and functionality of multiple dictation elements within Dragonfly speech recognition gramm...
""" Base class that implements reporting in human-readable format details about the recognized phrase. It is used by the actual testing rules below, and allows the doctests above to be easily readable and informative. """ def _process_recognition(self, node, extras): ...
<|file_name|>test_multiple_dictation.py<|end_file_name|><|fim▁begin|>""" Multiple dictation constructs =============================================================================== This file is a showcase investigating the use and functionality of multiple dictation elements within Dragonfly speech recognition gramm...
Paste(text).execute()
<|file_name|>test_multiple_dictation.py<|end_file_name|><|fim▁begin|>""" Multiple dictation constructs =============================================================================== This file is a showcase investigating the use and functionality of multiple dictation elements within Dragonfly speech recognition gramm...
return self.get_recognition_info(node)
<|file_name|>test_multiple_dictation.py<|end_file_name|><|fim▁begin|>""" Multiple dictation constructs =============================================================================== This file is a showcase investigating the use and functionality of multiple dictation elements within Dragonfly speech recognition gramm...
output = [] output.append('Recognition: "{0}"'.format(" ".join(node.words()))) output.append('Word and rule pairs: ("1000000" is "dgndictation")') for word, rule in node.full_results(): output.append(" - {0} ({1})".format(word, rule)) output.append("Extras:") ...
<|file_name|>test_multiple_dictation.py<|end_file_name|><|fim▁begin|>""" Multiple dictation constructs =============================================================================== This file is a showcase investigating the use and functionality of multiple dictation elements within Dragonfly speech recognition gramm...
spec = "consecutive <dictation1> <dictation2> <dictation3>" extras = [Dictation("dictation1"), Dictation("dictation2"), Dictation("dictation3")]
<|file_name|>test_multiple_dictation.py<|end_file_name|><|fim▁begin|>""" Multiple dictation constructs =============================================================================== This file is a showcase investigating the use and functionality of multiple dictation elements within Dragonfly speech recognition gramm...
spec = "mixed [<dictation1>] <dictation2> command <dictation3>" extras = [Dictation("dictation1"), Dictation("dictation2"), Dictation("dictation3")]
<|file_name|>test_multiple_dictation.py<|end_file_name|><|fim▁begin|>""" Multiple dictation constructs =============================================================================== This file is a showcase investigating the use and functionality of multiple dictation elements within Dragonfly speech recognition gramm...
spec = "<repetition>" extras = [Repetition(name="repetition", child=Sequence([Literal("command"), Dictation()]))]
<|file_name|>test_multiple_dictation.py<|end_file_name|><|fim▁begin|>""" Multiple dictation constructs =============================================================================== This file is a showcase investigating the use and functionality of multiple dictation elements within Dragonfly speech recognition gramm...
engine = get_engine() engine.connect() try: doctest.testmod(verbose=True) finally: engine.disconnect()
<|file_name|>test_multiple_dictation.py<|end_file_name|><|fim▁begin|>""" Multiple dictation constructs =============================================================================== This file is a showcase investigating the use and functionality of multiple dictation elements within Dragonfly speech recognition gramm...
output.append(" - {0}: {1}".format(key, extra_node.value()))
<|file_name|>test_multiple_dictation.py<|end_file_name|><|fim▁begin|>""" Multiple dictation constructs =============================================================================== This file is a showcase investigating the use and functionality of multiple dictation elements within Dragonfly speech recognition gramm...
main()
<|file_name|>test_multiple_dictation.py<|end_file_name|><|fim▁begin|>""" Multiple dictation constructs =============================================================================== This file is a showcase investigating the use and functionality of multiple dictation elements within Dragonfly speech recognition gramm...
_process_recognition
<|file_name|>test_multiple_dictation.py<|end_file_name|><|fim▁begin|>""" Multiple dictation constructs =============================================================================== This file is a showcase investigating the use and functionality of multiple dictation elements within Dragonfly speech recognition gramm...
value
<|file_name|>test_multiple_dictation.py<|end_file_name|><|fim▁begin|>""" Multiple dictation constructs =============================================================================== This file is a showcase investigating the use and functionality of multiple dictation elements within Dragonfly speech recognition gramm...
get_recognition_info
<|file_name|>test_multiple_dictation.py<|end_file_name|><|fim▁begin|>""" Multiple dictation constructs =============================================================================== This file is a showcase investigating the use and functionality of multiple dictation elements within Dragonfly speech recognition gramm...
main
<|file_name|>test_grids.py<|end_file_name|><|fim▁begin|>#!/usr/bin/env python3 from __future__ import print_function, division import numpy as np from sht.grids import standard_grid, get_cartesian_grid def test_grids(): L = 10 thetas, phis = standard_grid(L) # Can't really test much here assert thet...
grid = get_cartesian_grid(thetas, phis) assert grid.shape == (L**2, 3)
<|file_name|>test_grids.py<|end_file_name|><|fim▁begin|>#!/usr/bin/env python3 from __future__ import print_function, division import numpy as np from sht.grids import standard_grid, get_cartesian_grid def test_grids(): <|fim_middle|> <|fim▁end|>
L = 10 thetas, phis = standard_grid(L) # Can't really test much here assert thetas.size == L assert phis.size == L**2 grid = get_cartesian_grid(thetas, phis) assert grid.shape == (L**2, 3)
<|file_name|>test_grids.py<|end_file_name|><|fim▁begin|>#!/usr/bin/env python3 from __future__ import print_function, division import numpy as np from sht.grids import standard_grid, get_cartesian_grid def <|fim_middle|>(): L = 10 thetas, phis = standard_grid(L) # Can't really test much here assert ...
test_grids
<|file_name|>squareplot.py<|end_file_name|><|fim▁begin|># Simple plotter for Gut books. # All this does is draw a sqare for every stat in the input # and color it based on the score. # # Another fine example of how Viz lies to you. You will # need to fudge the range by adjusting the clipping. import numpy as np imp...
print cmap(0.1)
<|file_name|>squareplot.py<|end_file_name|><|fim▁begin|># Simple plotter for Gut books. # All this does is draw a sqare for every stat in the input # and color it based on the score. # # Another fine example of how Viz lies to you. You will # need to fudge the range by adjusting the clipping. import numpy as np imp...
print "Need an input file with many rows of 'id score'\n" sys.exit(1)
<|file_name|>wrapper_head_tail.py<|end_file_name|><|fim▁begin|>import sys import petsc4py petsc4py.init(sys.argv) from ecoli_in_pipe import head_tail # import numpy as np # from scipy.interpolate import interp1d # from petsc4py import PETSc # from ecoli_in_pipe import single_ecoli, ecoliInPipe, head_tail, ecoli_U # f...
# main_kwargs = {'fileHandle': fileHandle} # # head_U, tail_U, ref_U = ecoli_common.ecoli_restart(**main_kwargs) # # ecoli_common.ecoli_restart(**main_kwargs) # head_U = np.array([0, 0, 1, 0, 0, 1])
<|file_name|>sort.py<|end_file_name|><|fim▁begin|>#!/usr/bin/env python import sys, argparse def main(): parser = argparse.ArgumentParser() parser.add_argument('-i', '--input', type=str, action='store', dest='input', default=None, help="Input file") args = parser.parse_args() stats = dict() if arg...
start_time, prev_tx = time, tx_bytes else: print (time-start_time), (tx_bytes-prev_tx)
<|file_name|>sort.py<|end_file_name|><|fim▁begin|>#!/usr/bin/env python import sys, argparse def main(): <|fim_middle|> if __name__ == "__main__": main() <|fim▁end|>
parser = argparse.ArgumentParser() parser.add_argument('-i', '--input', type=str, action='store', dest='input', default=None, help="Input file") args = parser.parse_args() stats = dict() if args.input is None: print "Error: No input file" with open(args.input) as in_file: for lin...
<|file_name|>sort.py<|end_file_name|><|fim▁begin|>#!/usr/bin/env python import sys, argparse def main(): parser = argparse.ArgumentParser() parser.add_argument('-i', '--input', type=str, action='store', dest='input', default=None, help="Input file") args = parser.parse_args() stats = dict() if arg...
print "Error: No input file"
<|file_name|>sort.py<|end_file_name|><|fim▁begin|>#!/usr/bin/env python import sys, argparse def main(): parser = argparse.ArgumentParser() parser.add_argument('-i', '--input', type=str, action='store', dest='input', default=None, help="Input file") args = parser.parse_args() stats = dict() if arg...
if tx_bytes > (prev_tx+100000): no_traffic_flag = False start_time, prev_tx = time, tx_bytes
<|file_name|>sort.py<|end_file_name|><|fim▁begin|>#!/usr/bin/env python import sys, argparse def main(): parser = argparse.ArgumentParser() parser.add_argument('-i', '--input', type=str, action='store', dest='input', default=None, help="Input file") args = parser.parse_args() stats = dict() if arg...
no_traffic_flag = False start_time, prev_tx = time, tx_bytes
<|file_name|>sort.py<|end_file_name|><|fim▁begin|>#!/usr/bin/env python import sys, argparse def main(): parser = argparse.ArgumentParser() parser.add_argument('-i', '--input', type=str, action='store', dest='input', default=None, help="Input file") args = parser.parse_args() stats = dict() if arg...
print (time-start_time), (tx_bytes-prev_tx) prev_tx = tx_bytes
<|file_name|>sort.py<|end_file_name|><|fim▁begin|>#!/usr/bin/env python import sys, argparse def main(): parser = argparse.ArgumentParser() parser.add_argument('-i', '--input', type=str, action='store', dest='input', default=None, help="Input file") args = parser.parse_args() stats = dict() if arg...
main()
<|file_name|>sort.py<|end_file_name|><|fim▁begin|>#!/usr/bin/env python import sys, argparse def <|fim_middle|>(): parser = argparse.ArgumentParser() parser.add_argument('-i', '--input', type=str, action='store', dest='input', default=None, help="Input file") args = parser.parse_args() stats = dict() ...
main
<|file_name|>signpdf.py<|end_file_name|><|fim▁begin|>#!/usr/bin/env python import os import time import argparse import tempfile import PyPDF2 import datetime from reportlab.pdfgen import canvas parser = argparse.ArgumentParser("Add signatures to PDF files") parser.add_argument("pdf", help="The pdf file to annotate")...
c.drawString(x1 + width, y1, datetime.datetime.now().strftime("%Y-%m-%d"))
<|file_name|>signpdf.py<|end_file_name|><|fim▁begin|>#!/usr/bin/env python import os import time import argparse import tempfile import PyPDF2 import datetime from reportlab.pdfgen import canvas parser = argparse.ArgumentParser("Add signatures to PDF files") parser.add_argument("pdf", help="The pdf file to annotate")...
with tempfile.NamedTemporaryFile(suffix=".pdf") as fh: return fh.name
<|file_name|>signpdf.py<|end_file_name|><|fim▁begin|>#!/usr/bin/env python import os import time import argparse import tempfile import PyPDF2 import datetime from reportlab.pdfgen import canvas parser = argparse.ArgumentParser("Add signatures to PDF files") parser.add_argument("pdf", help="The pdf file to annotate")...
page_num, x1, y1, width, height = [int(a) for a in args.coords.split("x")] page_num -= 1 output_filename = args.output or "{}_signed{}".format( *os.path.splitext(args.pdf) ) pdf_fh = open(args.pdf, 'rb') sig_tmp_fh = None pdf = PyPDF2.PdfFileReader(pdf_fh) writer = PyPDF2.PdfF...
<|file_name|>signpdf.py<|end_file_name|><|fim▁begin|>#!/usr/bin/env python import os import time import argparse import tempfile import PyPDF2 import datetime from reportlab.pdfgen import canvas parser = argparse.ArgumentParser("Add signatures to PDF files") parser.add_argument("pdf", help="The pdf file to annotate")...
sign_pdf(parser.parse_args())
<|file_name|>signpdf.py<|end_file_name|><|fim▁begin|>#!/usr/bin/env python import os import time import argparse import tempfile import PyPDF2 import datetime from reportlab.pdfgen import canvas parser = argparse.ArgumentParser("Add signatures to PDF files") parser.add_argument("pdf", help="The pdf file to annotate")...
sig_tmp_filename = _get_tmp_filename() c = canvas.Canvas(sig_tmp_filename, pagesize=page.cropBox) c.drawImage(args.signature, x1, y1, width, height, mask='auto') if args.date: c.drawString(x1 + width, y1, datetime.datetime.now().strftime("%Y-%m-%d")) ...
<|file_name|>signpdf.py<|end_file_name|><|fim▁begin|>#!/usr/bin/env python import os import time import argparse import tempfile import PyPDF2 import datetime from reportlab.pdfgen import canvas parser = argparse.ArgumentParser("Add signatures to PDF files") parser.add_argument("pdf", help="The pdf file to annotate")...
c.drawString(x1 + width, y1, datetime.datetime.now().strftime("%Y-%m-%d"))
<|file_name|>signpdf.py<|end_file_name|><|fim▁begin|>#!/usr/bin/env python import os import time import argparse import tempfile import PyPDF2 import datetime from reportlab.pdfgen import canvas parser = argparse.ArgumentParser("Add signatures to PDF files") parser.add_argument("pdf", help="The pdf file to annotate")...
handle.close()
<|file_name|>signpdf.py<|end_file_name|><|fim▁begin|>#!/usr/bin/env python import os import time import argparse import tempfile import PyPDF2 import datetime from reportlab.pdfgen import canvas parser = argparse.ArgumentParser("Add signatures to PDF files") parser.add_argument("pdf", help="The pdf file to annotate")...
os.remove(sig_tmp_filename)
<|file_name|>signpdf.py<|end_file_name|><|fim▁begin|>#!/usr/bin/env python import os import time import argparse import tempfile import PyPDF2 import datetime from reportlab.pdfgen import canvas parser = argparse.ArgumentParser("Add signatures to PDF files") parser.add_argument("pdf", help="The pdf file to annotate")...
main()
<|file_name|>signpdf.py<|end_file_name|><|fim▁begin|>#!/usr/bin/env python import os import time import argparse import tempfile import PyPDF2 import datetime from reportlab.pdfgen import canvas parser = argparse.ArgumentParser("Add signatures to PDF files") parser.add_argument("pdf", help="The pdf file to annotate")...
_get_tmp_filename
<|file_name|>signpdf.py<|end_file_name|><|fim▁begin|>#!/usr/bin/env python import os import time import argparse import tempfile import PyPDF2 import datetime from reportlab.pdfgen import canvas parser = argparse.ArgumentParser("Add signatures to PDF files") parser.add_argument("pdf", help="The pdf file to annotate")...
sign_pdf
<|file_name|>signpdf.py<|end_file_name|><|fim▁begin|>#!/usr/bin/env python import os import time import argparse import tempfile import PyPDF2 import datetime from reportlab.pdfgen import canvas parser = argparse.ArgumentParser("Add signatures to PDF files") parser.add_argument("pdf", help="The pdf file to annotate")...
main
<|file_name|>urls.py<|end_file_name|><|fim▁begin|>from django.conf.urls import patterns, include, url from django.views.generic import TemplateView home = TemplateView.as_view(template_name='home.html') urlpatterns = patterns( '',<|fim▁hole|> # An informative homepage. url(r'', home, name='home') )<|fim▁...
url(r'^filter/', include('demoproject.filter.urls')),
<|file_name|>filecheckpoints.py<|end_file_name|><|fim▁begin|>""" File-based Checkpoints implementations. """ import os import shutil from tornado.web import HTTPError from .checkpoints import ( Checkpoints, GenericCheckpointsMixin, ) from .fileio import FileManagerMixin from IPython.utils import tz from IPyt...
content, format = self._read_file(os_checkpoint_path, format=None)
<|file_name|>filecheckpoints.py<|end_file_name|><|fim▁begin|>""" File-based Checkpoints implementations. """ import os import shutil from tornado.web import HTTPError from .checkpoints import ( Checkpoints, GenericCheckpointsMixin, ) from .fileio import FileManagerMixin from IPython.utils import tz from IPyt...
""" A Checkpoints that caches checkpoints for files in adjacent directories. Only works with FileContentsManager. Use GenericFileCheckpoints if you want file-based checkpoints with another ContentsManager. """ checkpoint_dir = Unicode( '.ipynb_checkpoints', config=True, ...
<|file_name|>filecheckpoints.py<|end_file_name|><|fim▁begin|>""" File-based Checkpoints implementations. """ import os import shutil from tornado.web import HTTPError from .checkpoints import ( Checkpoints, GenericCheckpointsMixin, ) from .fileio import FileManagerMixin from IPython.utils import tz from IPyt...
try: return self.parent.root_dir except AttributeError: return getcwd()
<|file_name|>filecheckpoints.py<|end_file_name|><|fim▁begin|>""" File-based Checkpoints implementations. """ import os import shutil from tornado.web import HTTPError from .checkpoints import ( Checkpoints, GenericCheckpointsMixin, ) from .fileio import FileManagerMixin from IPython.utils import tz from IPyt...
"""Create a checkpoint.""" checkpoint_id = u'checkpoint' src_path = contents_mgr._get_os_path(path) dest_path = self.checkpoint_path(checkpoint_id, path) self._copy(src_path, dest_path) return self.checkpoint_model(checkpoint_id, dest_path)
<|file_name|>filecheckpoints.py<|end_file_name|><|fim▁begin|>""" File-based Checkpoints implementations. """ import os import shutil from tornado.web import HTTPError from .checkpoints import ( Checkpoints, GenericCheckpointsMixin, ) from .fileio import FileManagerMixin from IPython.utils import tz from IPyt...
"""Restore a checkpoint.""" src_path = self.checkpoint_path(checkpoint_id, path) dest_path = contents_mgr._get_os_path(path) self._copy(src_path, dest_path)
<|file_name|>filecheckpoints.py<|end_file_name|><|fim▁begin|>""" File-based Checkpoints implementations. """ import os import shutil from tornado.web import HTTPError from .checkpoints import ( Checkpoints, GenericCheckpointsMixin, ) from .fileio import FileManagerMixin from IPython.utils import tz from IPyt...
"""Rename a checkpoint from old_path to new_path.""" old_cp_path = self.checkpoint_path(checkpoint_id, old_path) new_cp_path = self.checkpoint_path(checkpoint_id, new_path) if os.path.isfile(old_cp_path): self.log.debug( "Renaming checkpoint %s -> %s", ...
<|file_name|>filecheckpoints.py<|end_file_name|><|fim▁begin|>""" File-based Checkpoints implementations. """ import os import shutil from tornado.web import HTTPError from .checkpoints import ( Checkpoints, GenericCheckpointsMixin, ) from .fileio import FileManagerMixin from IPython.utils import tz from IPyt...
"""delete a file's checkpoint""" path = path.strip('/') cp_path = self.checkpoint_path(checkpoint_id, path) if not os.path.isfile(cp_path): self.no_such_checkpoint(path, checkpoint_id) self.log.debug("unlinking %s", cp_path) with self.perm_to_403(): ...
<|file_name|>filecheckpoints.py<|end_file_name|><|fim▁begin|>""" File-based Checkpoints implementations. """ import os import shutil from tornado.web import HTTPError from .checkpoints import ( Checkpoints, GenericCheckpointsMixin, ) from .fileio import FileManagerMixin from IPython.utils import tz from IPyt...
"""list the checkpoints for a given file This contents manager currently only supports one checkpoint per file. """ path = path.strip('/') checkpoint_id = "checkpoint" os_path = self.checkpoint_path(checkpoint_id, path) if not os.path.isfile(os_path): ...
<|file_name|>filecheckpoints.py<|end_file_name|><|fim▁begin|>""" File-based Checkpoints implementations. """ import os import shutil from tornado.web import HTTPError from .checkpoints import ( Checkpoints, GenericCheckpointsMixin, ) from .fileio import FileManagerMixin from IPython.utils import tz from IPyt...
"""find the path to a checkpoint""" path = path.strip('/') parent, name = ('/' + path).rsplit('/', 1) parent = parent.strip('/') basename, ext = os.path.splitext(name) filename = u"{name}-{checkpoint_id}{ext}".format( name=basename, checkpoint_id=c...
<|file_name|>filecheckpoints.py<|end_file_name|><|fim▁begin|>""" File-based Checkpoints implementations. """ import os import shutil from tornado.web import HTTPError from .checkpoints import ( Checkpoints, GenericCheckpointsMixin, ) from .fileio import FileManagerMixin from IPython.utils import tz from IPyt...
"""construct the info dict for a given checkpoint""" stats = os.stat(os_path) last_modified = tz.utcfromtimestamp(stats.st_mtime) info = dict( id=checkpoint_id, last_modified=last_modified, ) return info
<|file_name|>filecheckpoints.py<|end_file_name|><|fim▁begin|>""" File-based Checkpoints implementations. """ import os import shutil from tornado.web import HTTPError from .checkpoints import ( Checkpoints, GenericCheckpointsMixin, ) from .fileio import FileManagerMixin from IPython.utils import tz from IPyt...
raise HTTPError( 404, u'Checkpoint does not exist: %s@%s' % (path, checkpoint_id) )
<|file_name|>filecheckpoints.py<|end_file_name|><|fim▁begin|>""" File-based Checkpoints implementations. """ import os import shutil from tornado.web import HTTPError from .checkpoints import ( Checkpoints, GenericCheckpointsMixin, ) from .fileio import FileManagerMixin from IPython.utils import tz from IPyt...
""" Local filesystem Checkpoints that works with any conforming ContentsManager. """ def create_file_checkpoint(self, content, format, path): """Create a checkpoint from the current content of a file.""" path = path.strip('/') # only the one checkpoint ID: checkpoint_...
<|file_name|>filecheckpoints.py<|end_file_name|><|fim▁begin|>""" File-based Checkpoints implementations. """ import os import shutil from tornado.web import HTTPError from .checkpoints import ( Checkpoints, GenericCheckpointsMixin, ) from .fileio import FileManagerMixin from IPython.utils import tz from IPyt...
"""Create a checkpoint from the current content of a file.""" path = path.strip('/') # only the one checkpoint ID: checkpoint_id = u"checkpoint" os_checkpoint_path = self.checkpoint_path(checkpoint_id, path) self.log.debug("creating checkpoint for %s", path) with ...
<|file_name|>filecheckpoints.py<|end_file_name|><|fim▁begin|>""" File-based Checkpoints implementations. """ import os import shutil from tornado.web import HTTPError from .checkpoints import ( Checkpoints, GenericCheckpointsMixin, ) from .fileio import FileManagerMixin from IPython.utils import tz from IPyt...
"""Create a checkpoint from the current content of a notebook.""" path = path.strip('/') # only the one checkpoint ID: checkpoint_id = u"checkpoint" os_checkpoint_path = self.checkpoint_path(checkpoint_id, path) self.log.debug("creating checkpoint for %s", path) w...
<|file_name|>filecheckpoints.py<|end_file_name|><|fim▁begin|>""" File-based Checkpoints implementations. """ import os import shutil from tornado.web import HTTPError from .checkpoints import ( Checkpoints, GenericCheckpointsMixin, ) from .fileio import FileManagerMixin from IPython.utils import tz from IPyt...
"""Get a checkpoint for a notebook.""" path = path.strip('/') self.log.info("restoring %s from checkpoint %s", path, checkpoint_id) os_checkpoint_path = self.checkpoint_path(checkpoint_id, path) if not os.path.isfile(os_checkpoint_path): self.no_such_checkpoint(path,...
<|file_name|>filecheckpoints.py<|end_file_name|><|fim▁begin|>""" File-based Checkpoints implementations. """ import os import shutil from tornado.web import HTTPError from .checkpoints import ( Checkpoints, GenericCheckpointsMixin, ) from .fileio import FileManagerMixin from IPython.utils import tz from IPyt...
"""Get a checkpoint for a file.""" path = path.strip('/') self.log.info("restoring %s from checkpoint %s", path, checkpoint_id) os_checkpoint_path = self.checkpoint_path(checkpoint_id, path) if not os.path.isfile(os_checkpoint_path): self.no_such_checkpoint(path, che...
<|file_name|>filecheckpoints.py<|end_file_name|><|fim▁begin|>""" File-based Checkpoints implementations. """ import os import shutil from tornado.web import HTTPError from .checkpoints import ( Checkpoints, GenericCheckpointsMixin, ) from .fileio import FileManagerMixin from IPython.utils import tz from IPyt...
self.log.debug( "Renaming checkpoint %s -> %s", old_cp_path, new_cp_path, ) with self.perm_to_403(): shutil.move(old_cp_path, new_cp_path)
<|file_name|>filecheckpoints.py<|end_file_name|><|fim▁begin|>""" File-based Checkpoints implementations. """ import os import shutil from tornado.web import HTTPError from .checkpoints import ( Checkpoints, GenericCheckpointsMixin, ) from .fileio import FileManagerMixin from IPython.utils import tz from IPyt...
self.no_such_checkpoint(path, checkpoint_id)
<|file_name|>filecheckpoints.py<|end_file_name|><|fim▁begin|>""" File-based Checkpoints implementations. """ import os import shutil from tornado.web import HTTPError from .checkpoints import ( Checkpoints, GenericCheckpointsMixin, ) from .fileio import FileManagerMixin from IPython.utils import tz from IPyt...
return []
<|file_name|>filecheckpoints.py<|end_file_name|><|fim▁begin|>""" File-based Checkpoints implementations. """ import os import shutil from tornado.web import HTTPError from .checkpoints import ( Checkpoints, GenericCheckpointsMixin, ) from .fileio import FileManagerMixin from IPython.utils import tz from IPyt...
return [self.checkpoint_model(checkpoint_id, os_path)]
<|file_name|>filecheckpoints.py<|end_file_name|><|fim▁begin|>""" File-based Checkpoints implementations. """ import os import shutil from tornado.web import HTTPError from .checkpoints import ( Checkpoints, GenericCheckpointsMixin, ) from .fileio import FileManagerMixin from IPython.utils import tz from IPyt...
self.no_such_checkpoint(path, checkpoint_id)
<|file_name|>filecheckpoints.py<|end_file_name|><|fim▁begin|>""" File-based Checkpoints implementations. """ import os import shutil from tornado.web import HTTPError from .checkpoints import ( Checkpoints, GenericCheckpointsMixin, ) from .fileio import FileManagerMixin from IPython.utils import tz from IPyt...
self.no_such_checkpoint(path, checkpoint_id)
<|file_name|>filecheckpoints.py<|end_file_name|><|fim▁begin|>""" File-based Checkpoints implementations. """ import os import shutil from tornado.web import HTTPError from .checkpoints import ( Checkpoints, GenericCheckpointsMixin, ) from .fileio import FileManagerMixin from IPython.utils import tz from IPyt...
_root_dir_default
<|file_name|>filecheckpoints.py<|end_file_name|><|fim▁begin|>""" File-based Checkpoints implementations. """ import os import shutil from tornado.web import HTTPError from .checkpoints import ( Checkpoints, GenericCheckpointsMixin, ) from .fileio import FileManagerMixin from IPython.utils import tz from IPyt...
create_checkpoint
<|file_name|>filecheckpoints.py<|end_file_name|><|fim▁begin|>""" File-based Checkpoints implementations. """ import os import shutil from tornado.web import HTTPError from .checkpoints import ( Checkpoints, GenericCheckpointsMixin, ) from .fileio import FileManagerMixin from IPython.utils import tz from IPyt...
restore_checkpoint
<|file_name|>filecheckpoints.py<|end_file_name|><|fim▁begin|>""" File-based Checkpoints implementations. """ import os import shutil from tornado.web import HTTPError from .checkpoints import ( Checkpoints, GenericCheckpointsMixin, ) from .fileio import FileManagerMixin from IPython.utils import tz from IPyt...
rename_checkpoint
<|file_name|>filecheckpoints.py<|end_file_name|><|fim▁begin|>""" File-based Checkpoints implementations. """ import os import shutil from tornado.web import HTTPError from .checkpoints import ( Checkpoints, GenericCheckpointsMixin, ) from .fileio import FileManagerMixin from IPython.utils import tz from IPyt...
delete_checkpoint
<|file_name|>filecheckpoints.py<|end_file_name|><|fim▁begin|>""" File-based Checkpoints implementations. """ import os import shutil from tornado.web import HTTPError from .checkpoints import ( Checkpoints, GenericCheckpointsMixin, ) from .fileio import FileManagerMixin from IPython.utils import tz from IPyt...
list_checkpoints
<|file_name|>filecheckpoints.py<|end_file_name|><|fim▁begin|>""" File-based Checkpoints implementations. """ import os import shutil from tornado.web import HTTPError from .checkpoints import ( Checkpoints, GenericCheckpointsMixin, ) from .fileio import FileManagerMixin from IPython.utils import tz from IPyt...
checkpoint_path
<|file_name|>filecheckpoints.py<|end_file_name|><|fim▁begin|>""" File-based Checkpoints implementations. """ import os import shutil from tornado.web import HTTPError from .checkpoints import ( Checkpoints, GenericCheckpointsMixin, ) from .fileio import FileManagerMixin from IPython.utils import tz from IPyt...
checkpoint_model
<|file_name|>filecheckpoints.py<|end_file_name|><|fim▁begin|>""" File-based Checkpoints implementations. """ import os import shutil from tornado.web import HTTPError from .checkpoints import ( Checkpoints, GenericCheckpointsMixin, ) from .fileio import FileManagerMixin from IPython.utils import tz from IPyt...
no_such_checkpoint
<|file_name|>filecheckpoints.py<|end_file_name|><|fim▁begin|>""" File-based Checkpoints implementations. """ import os import shutil from tornado.web import HTTPError from .checkpoints import ( Checkpoints, GenericCheckpointsMixin, ) from .fileio import FileManagerMixin from IPython.utils import tz from IPyt...
create_file_checkpoint
<|file_name|>filecheckpoints.py<|end_file_name|><|fim▁begin|>""" File-based Checkpoints implementations. """ import os import shutil from tornado.web import HTTPError from .checkpoints import ( Checkpoints, GenericCheckpointsMixin, ) from .fileio import FileManagerMixin from IPython.utils import tz from IPyt...
create_notebook_checkpoint
<|file_name|>filecheckpoints.py<|end_file_name|><|fim▁begin|>""" File-based Checkpoints implementations. """ import os import shutil from tornado.web import HTTPError from .checkpoints import ( Checkpoints, GenericCheckpointsMixin, ) from .fileio import FileManagerMixin from IPython.utils import tz from IPyt...
get_notebook_checkpoint
<|file_name|>filecheckpoints.py<|end_file_name|><|fim▁begin|>""" File-based Checkpoints implementations. """ import os import shutil from tornado.web import HTTPError from .checkpoints import ( Checkpoints, GenericCheckpointsMixin, ) from .fileio import FileManagerMixin from IPython.utils import tz from IPyt...
get_file_checkpoint
<|file_name|>HELP.py<|end_file_name|><|fim▁begin|><|fim▁hole|><|fim▁end|>
print("hello!!!!")
<|file_name|>test_iamgen.py<|end_file_name|><|fim▁begin|># Copyright 2020 Amazon.com, Inc. or its affiliates. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://w...
<|file_name|>test_iamgen.py<|end_file_name|><|fim▁begin|># Copyright 2020 Amazon.com, Inc. or its affiliates. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://w...
def check_permissions(self, perm_db, perm_set, path): invalid = [] for p in perm_set: if ':' not in p: invalid.append(p) continue s, a = p.split(':', 1) if s not in perm_db: invalid.append(p) continue...
<|file_name|>test_iamgen.py<|end_file_name|><|fim▁begin|># Copyright 2020 Amazon.com, Inc. or its affiliates. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://w...
invalid = [] for p in perm_set: if ':' not in p: invalid.append(p) continue s, a = p.split(':', 1) if s not in perm_db: invalid.append(p) continue if '*' in a: if not fnmatch.f...
<|file_name|>test_iamgen.py<|end_file_name|><|fim▁begin|># Copyright 2020 Amazon.com, Inc. or its affiliates. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://w...
cfg = Config.empty() missing = set() all_invalid = [] perms = load_data('iam-actions.json') for k, v in manager.resources.items(): p = Bag({'name': 'permcheck', 'resource': k, 'provider_name': 'aws'}) ctx = self.get_context(config=cfg, policy=p) ...
<|file_name|>test_iamgen.py<|end_file_name|><|fim▁begin|># Copyright 2020 Amazon.com, Inc. or its affiliates. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://w...
invalid.append(p) continue