Search is not available for this dataset
text
stringlengths
75
104k
def info_formatter(info): """Produce a sequence of formatted lines from info. `info` is a sequence of pairs (label, data). The produced lines are nicely formatted, ready to print. """ label_len = max([len(l) for l, _d in info]) for label, data in info: if data == []: data = "-none-" if isinstance(data, (list, tuple)): prefix = "%*s:" % (label_len, label) for e in data: yield "%*s %s" % (label_len+1, prefix, e) prefix = "" else: yield "%*s: %s" % (label_len, label, data)
def write(self, msg): """Write a line of debug output.""" if self.should('pid'): msg = "pid %5d: %s" % (os.getpid(), msg) self.output.write(msg+"\n") self.output.flush()
def _config_changed(self, name, old, new): """Update all the class traits having ``config=True`` as metadata. For any class trait with a ``config`` metadata attribute that is ``True``, we update the trait with the value of the corresponding config entry. """ # Get all traits with a config metadata entry that is True traits = self.traits(config=True) # We auto-load config section for this class as well as any parent # classes that are Configurable subclasses. This starts with Configurable # and works down the mro loading the config for each section. section_names = [cls.__name__ for cls in \ reversed(self.__class__.__mro__) if issubclass(cls, Configurable) and issubclass(self.__class__, cls)] for sname in section_names: # Don't do a blind getattr as that would cause the config to # dynamically create the section with name self.__class__.__name__. if new._has_section(sname): my_config = new[sname] for k, v in traits.iteritems(): # Don't allow traitlets with config=True to start with # uppercase. Otherwise, they are confused with Config # subsections. But, developers shouldn't have uppercase # attributes anyways! (PEP 6) if k[0].upper()==k[0] and not k.startswith('_'): raise ConfigurableError('Configurable traitlets with ' 'config=True must start with a lowercase so they are ' 'not confused with Config subsections: %s.%s' % \ (self.__class__.__name__, k)) try: # Here we grab the value from the config # If k has the naming convention of a config # section, it will be auto created. config_value = my_config[k] except KeyError: pass else: # print "Setting %s.%s from %s.%s=%r" % \ # (self.__class__.__name__,k,sname,k,config_value) # We have to do a deepcopy here if we don't deepcopy the entire # config object. If we don't, a mutable config_value will be # shared by all instances, effectively making it a class attribute. setattr(self, k, deepcopy(config_value))
def class_get_help(cls, inst=None): """Get the help string for this class in ReST format. If `inst` is given, it's current trait values will be used in place of class defaults. """ assert inst is None or isinstance(inst, cls) cls_traits = cls.class_traits(config=True) final_help = [] final_help.append(u'%s options' % cls.__name__) final_help.append(len(final_help[0])*u'-') for k,v in sorted(cls.class_traits(config=True).iteritems()): help = cls.class_get_trait_help(v, inst) final_help.append(help) return '\n'.join(final_help)
def class_get_trait_help(cls, trait, inst=None): """Get the help string for a single trait. If `inst` is given, it's current trait values will be used in place of the class default. """ assert inst is None or isinstance(inst, cls) lines = [] header = "--%s.%s=<%s>" % (cls.__name__, trait.name, trait.__class__.__name__) lines.append(header) if inst is not None: lines.append(indent('Current: %r' % getattr(inst, trait.name), 4)) else: try: dvr = repr(trait.get_default_value()) except Exception: dvr = None # ignore defaults we can't construct if dvr is not None: if len(dvr) > 64: dvr = dvr[:61]+'...' lines.append(indent('Default: %s' % dvr, 4)) if 'Enum' in trait.__class__.__name__: # include Enum choices lines.append(indent('Choices: %r' % (trait.values,))) help = trait.get_metadata('help') if help is not None: help = '\n'.join(wrap_paragraphs(help, 76)) lines.append(indent(help, 4)) return '\n'.join(lines)
def class_config_section(cls): """Get the config class config section""" def c(s): """return a commented, wrapped block.""" s = '\n\n'.join(wrap_paragraphs(s, 78)) return '# ' + s.replace('\n', '\n# ') # section header breaker = '#' + '-'*78 s = "# %s configuration"%cls.__name__ lines = [breaker, s, breaker, ''] # get the description trait desc = cls.class_traits().get('description') if desc: desc = desc.default_value else: # no description trait, use __doc__ desc = getattr(cls, '__doc__', '') if desc: lines.append(c(desc)) lines.append('') parents = [] for parent in cls.mro(): # only include parents that are not base classes # and are not the class itself # and have some configurable traits to inherit if parent is not cls and issubclass(parent, Configurable) and \ parent.class_traits(config=True): parents.append(parent) if parents: pstr = ', '.join([ p.__name__ for p in parents ]) lines.append(c('%s will inherit config from: %s'%(cls.__name__, pstr))) lines.append('') for name,trait in cls.class_traits(config=True).iteritems(): help = trait.get_metadata('help') or '' lines.append(c(help)) lines.append('# c.%s.%s = %r'%(cls.__name__, name, trait.get_default_value())) lines.append('') return '\n'.join(lines)
def _walk_mro(cls): """Walk the cls.mro() for parent classes that are also singletons For use in instance() """ for subclass in cls.mro(): if issubclass(cls, subclass) and \ issubclass(subclass, SingletonConfigurable) and \ subclass != SingletonConfigurable: yield subclass
def clear_instance(cls): """unset _instance for this class and singleton parents. """ if not cls.initialized(): return for subclass in cls._walk_mro(): if isinstance(subclass._instance, cls): # only clear instances that are instances # of the calling class subclass._instance = None
def instance(cls, *args, **kwargs): """Returns a global instance of this class. This method create a new instance if none have previously been created and returns a previously created instance is one already exists. The arguments and keyword arguments passed to this method are passed on to the :meth:`__init__` method of the class upon instantiation. Examples -------- Create a singleton class using instance, and retrieve it:: >>> from IPython.config.configurable import SingletonConfigurable >>> class Foo(SingletonConfigurable): pass >>> foo = Foo.instance() >>> foo == Foo.instance() True Create a subclass that is retrived using the base class instance:: >>> class Bar(SingletonConfigurable): pass >>> class Bam(Bar): pass >>> bam = Bam.instance() >>> bam == Bar.instance() True """ # Create and save the instance if cls._instance is None: inst = cls(*args, **kwargs) # Now make sure that the instance will also be returned by # parent classes' _instance attribute. for subclass in cls._walk_mro(): subclass._instance = inst if isinstance(cls._instance, cls): return cls._instance else: raise MultipleInstanceError( 'Multiple incompatible subclass instances of ' '%s are being created.' % cls.__name__ )
def configure(self, options, conf): """Configure plugin. """ if not self.can_configure: return self.enabled = options.detailedErrors self.conf = conf
def formatFailure(self, test, err): """Add detail from traceback inspection to error message of a failure. """ ec, ev, tb = err tbinfo = inspect_traceback(tb) test.tbinfo = tbinfo return (ec, '\n'.join([str(ev), tbinfo]), tb)
def crash_handler_lite(etype, evalue, tb): """a light excepthook, adding a small message to the usual traceback""" traceback.print_exception(etype, evalue, tb) from IPython.core.interactiveshell import InteractiveShell if InteractiveShell.initialized(): # we are in a Shell environment, give %magic example config = "%config " else: # we are not in a shell, show generic config config = "c." print >> sys.stderr, _lite_message_template.format(email=author_email, config=config)
def make_report(self,traceback): """Return a string containing a crash report.""" sec_sep = self.section_sep report = ['*'*75+'\n\n'+'IPython post-mortem report\n\n'] rpt_add = report.append rpt_add(sys_info()) try: config = pformat(self.app.config) rpt_add(sec_sep) rpt_add('Application name: %s\n\n' % self.app_name) rpt_add('Current user configuration structure:\n\n') rpt_add(config) except: pass rpt_add(sec_sep+'Crash traceback:\n\n' + traceback) return ''.join(report)
def setvar(parser, token): """ {% setvar <var_name> to <var_value> %} """ try: setvar, var_name, to_, var_value = token.split_contents() except ValueError: raise template.TemplateSyntaxError('Invalid arguments for %r' % token.split_contents()[0]) return SetVarNode(var_name, var_value)
def call_handlers(self, msg): """ Reimplemented to emit signals instead of making callbacks. """ # Emit the generic signal. self.message_received.emit(msg) # Emit signals for specialized message types. msg_type = msg['header']['msg_type'] signal = getattr(self, msg_type, None) if signal: signal.emit(msg) if not self._handlers_called: self.first_reply.emit() self._handlers_called = True
def call_handlers(self, msg): """ Reimplemented to emit signals instead of making callbacks. """ # Emit the generic signal. self.message_received.emit(msg) # Emit signals for specialized message types. msg_type = msg['header']['msg_type'] signal = getattr(self, msg_type + '_received', None) if signal: signal.emit(msg) elif msg_type in ('stdout', 'stderr'): self.stream_received.emit(msg)
def flush(self): """ Reimplemented to ensure that signals are dispatched immediately. """ super(QtSubSocketChannel, self).flush() QtCore.QCoreApplication.instance().processEvents()
def call_handlers(self, msg): """ Reimplemented to emit signals instead of making callbacks. """ # Emit the generic signal. self.message_received.emit(msg) # Emit signals for specialized message types. msg_type = msg['header']['msg_type'] if msg_type == 'input_request': self.input_requested.emit(msg)
def start_kernel(self, *args, **kw): """ Reimplemented for proper heartbeat management. """ if self._shell_channel is not None: self._shell_channel.reset_first_reply() super(QtKernelManager, self).start_kernel(*args, **kw) self.started_kernel.emit()
def start_channels(self, *args, **kw): """ Reimplemented to emit signal. """ super(QtKernelManager, self).start_channels(*args, **kw) self.started_channels.emit()
def shell_channel(self): """ Reimplemented for proper heartbeat management. """ if self._shell_channel is None: self._shell_channel = super(QtKernelManager, self).shell_channel self._shell_channel.first_reply.connect(self._first_reply) return self._shell_channel
def restore_bytes(nb): """Restore bytes of image data from unicode-only formats. Base64 encoding is handled elsewhere. Bytes objects in the notebook are always b64-encoded. We DO NOT encode/decode around file formats. """ for ws in nb.worksheets: for cell in ws.cells: if cell.cell_type == 'code': for output in cell.outputs: if 'png' in output: output.png = str_to_bytes(output.png, 'ascii') if 'jpeg' in output: output.jpeg = str_to_bytes(output.jpeg, 'ascii') return nb
def _join_lines(lines): """join lines that have been written by splitlines() Has logic to protect against `splitlines()`, which should have been `splitlines(True)` """ if lines and lines[0].endswith(('\n', '\r')): # created by splitlines(True) return u''.join(lines) else: # created by splitlines() return u'\n'.join(lines)
def rejoin_lines(nb): """rejoin multiline text into strings For reversing effects of ``split_lines(nb)``. This only rejoins lines that have been split, so if text objects were not split they will pass through unchanged. Used when reading JSON files that may have been passed through split_lines. """ for ws in nb.worksheets: for cell in ws.cells: if cell.cell_type == 'code': if 'input' in cell and isinstance(cell.input, list): cell.input = _join_lines(cell.input) for output in cell.outputs: for key in _multiline_outputs: item = output.get(key, None) if isinstance(item, list): output[key] = _join_lines(item) else: # text, heading cell for key in ['source', 'rendered']: item = cell.get(key, None) if isinstance(item, list): cell[key] = _join_lines(item) return nb
def base64_decode(nb): """Restore all bytes objects in the notebook from base64-encoded strings. Note: This is never used """ for ws in nb.worksheets: for cell in ws.cells: if cell.cell_type == 'code': for output in cell.outputs: if 'png' in output: if isinstance(output.png, unicode): output.png = output.png.encode('ascii') output.png = decodestring(output.png) if 'jpeg' in output: if isinstance(output.jpeg, unicode): output.jpeg = output.jpeg.encode('ascii') output.jpeg = decodestring(output.jpeg) return nb
def base64_encode(nb): """Base64 encode all bytes objects in the notebook. These will be b64-encoded unicode strings Note: This is never used """ for ws in nb.worksheets: for cell in ws.cells: if cell.cell_type == 'code': for output in cell.outputs: if 'png' in output: output.png = encodestring(output.png).decode('ascii') if 'jpeg' in output: output.jpeg = encodestring(output.jpeg).decode('ascii') return nb
def read(self, fp, **kwargs): """Read a notebook from a file like object""" nbs = fp.read() if not py3compat.PY3 and not isinstance(nbs, unicode): nbs = py3compat.str_to_unicode(nbs) return self.reads(nbs, **kwargs)
def write(self, nb, fp, **kwargs): """Write a notebook to a file like object""" nbs = self.writes(nb,**kwargs) if not py3compat.PY3 and not isinstance(nbs, unicode): # this branch is likely only taken for JSON on Python 2 nbs = py3compat.str_to_unicode(nbs) return fp.write(nbs)
def get_mirrors(hostname=None): """Return the list of mirrors from the last record found on the DNS entry:: >>> from pip.index import get_mirrors >>> get_mirrors() ['a.pypi.python.org', 'b.pypi.python.org', 'c.pypi.python.org', 'd.pypi.python.org'] Originally written for the distutils2 project by Alexis Metaireau. """ if hostname is None: hostname = DEFAULT_MIRROR_HOSTNAME # return the last mirror registered on PyPI. last_mirror_hostname = None try: last_mirror_hostname = socket.gethostbyname_ex(hostname)[0] except socket.gaierror: return [] if not last_mirror_hostname or last_mirror_hostname == DEFAULT_MIRROR_HOSTNAME: last_mirror_hostname = "z.pypi.python.org" end_letter = last_mirror_hostname.split(".", 1) # determine the list from the last one. return ["%s.%s" % (s, end_letter[1]) for s in string_range(end_letter[0])]
def read_no_interrupt(p): """Read from a pipe ignoring EINTR errors. This is necessary because when reading from pipes with GUI event loops running in the background, often interrupts are raised that stop the command from completing.""" import errno try: return p.read() except IOError, err: if err.errno != errno.EINTR: raise
def process_handler(cmd, callback, stderr=subprocess.PIPE): """Open a command in a shell subprocess and execute a callback. This function provides common scaffolding for creating subprocess.Popen() calls. It creates a Popen object and then calls the callback with it. Parameters ---------- cmd : str A string to be executed with the underlying system shell (by calling :func:`Popen` with ``shell=True``. callback : callable A one-argument function that will be called with the Popen object. stderr : file descriptor number, optional By default this is set to ``subprocess.PIPE``, but you can also pass the value ``subprocess.STDOUT`` to force the subprocess' stderr to go into the same file descriptor as its stdout. This is useful to read stdout and stderr combined in the order they are generated. Returns ------- The return value of the provided callback is returned. """ sys.stdout.flush() sys.stderr.flush() # On win32, close_fds can't be true when using pipes for stdin/out/err close_fds = sys.platform != 'win32' p = subprocess.Popen(cmd, shell=True, stdin=subprocess.PIPE, stdout=subprocess.PIPE, stderr=stderr, close_fds=close_fds) try: out = callback(p) except KeyboardInterrupt: print('^C') sys.stdout.flush() sys.stderr.flush() out = None finally: # Make really sure that we don't leave processes behind, in case the # call above raises an exception # We start by assuming the subprocess finished (to avoid NameErrors # later depending on the path taken) if p.returncode is None: try: p.terminate() p.poll() except OSError: pass # One last try on our way out if p.returncode is None: try: p.kill() except OSError: pass return out
def getoutput(cmd): """Return standard output of executing cmd in a shell. Accepts the same arguments as os.system(). Parameters ---------- cmd : str A command to be executed in the system shell. Returns ------- stdout : str """ out = process_handler(cmd, lambda p: p.communicate()[0], subprocess.STDOUT) if out is None: return '' return py3compat.bytes_to_str(out)
def getoutputerror(cmd): """Return (standard output, standard error) of executing cmd in a shell. Accepts the same arguments as os.system(). Parameters ---------- cmd : str A command to be executed in the system shell. Returns ------- stdout : str stderr : str """ out_err = process_handler(cmd, lambda p: p.communicate()) if out_err is None: return '', '' out, err = out_err return py3compat.bytes_to_str(out), py3compat.bytes_to_str(err)
def arg_split(s, posix=False, strict=True): """Split a command line's arguments in a shell-like manner. This is a modified version of the standard library's shlex.split() function, but with a default of posix=False for splitting, so that quotes in inputs are respected. if strict=False, then any errors shlex.split would raise will result in the unparsed remainder being the last element of the list, rather than raising. This is because we sometimes use arg_split to parse things other than command-line args. """ # Unfortunately, python's shlex module is buggy with unicode input: # http://bugs.python.org/issue1170 # At least encoding the input when it's unicode seems to help, but there # may be more problems lurking. Apparently this is fixed in python3. is_unicode = False if (not py3compat.PY3) and isinstance(s, unicode): is_unicode = True s = s.encode('utf-8') lex = shlex.shlex(s, posix=posix) lex.whitespace_split = True # Extract tokens, ensuring that things like leaving open quotes # does not cause this to raise. This is important, because we # sometimes pass Python source through this (e.g. %timeit f(" ")), # and it shouldn't raise an exception. # It may be a bad idea to parse things that are not command-line args # through this function, but we do, so let's be safe about it. lex.commenters='' #fix for GH-1269 tokens = [] while True: try: tokens.append(lex.next()) except StopIteration: break except ValueError: if strict: raise # couldn't parse, get remaining blob as last token tokens.append(lex.token) break if is_unicode: # Convert the tokens back to unicode. tokens = [x.decode('utf-8') for x in tokens] return tokens
def collector(): """TestSuite replacement entry point. Use anywhere you might use a unittest.TestSuite. The collector will, by default, load options from all config files and execute loader.loadTestsFromNames() on the configured testNames, or '.' if no testNames are configured. """ # plugins that implement any of these methods are disabled, since # we don't control the test runner and won't be able to run them # finalize() is also not called, but plugins that use it aren't disabled, # because capture needs it. setuptools_incompat = ('report', 'prepareTest', 'prepareTestLoader', 'prepareTestRunner', 'setOutputStream') plugins = RestrictedPluginManager(exclude=setuptools_incompat) conf = Config(files=all_config_files(), plugins=plugins) conf.configure(argv=['collector']) loader = defaultTestLoader(conf) if conf.testNames: suite = loader.loadTestsFromNames(conf.testNames) else: suite = loader.loadTestsFromNames(('.',)) return FinalizingSuiteWrapper(suite, plugins.finalize)
def compress_dhist(dh): """Compress a directory history into a new one with at most 20 entries. Return a new list made from the first and last 10 elements of dhist after removal of duplicates. """ head, tail = dh[:-10], dh[-10:] newhead = [] done = set() for h in head: if h in done: continue newhead.append(h) done.add(h) return newhead + tail
def magics_class(cls): """Class decorator for all subclasses of the main Magics class. Any class that subclasses Magics *must* also apply this decorator, to ensure that all the methods that have been decorated as line/cell magics get correctly registered in the class instance. This is necessary because when method decorators run, the class does not exist yet, so they temporarily store their information into a module global. Application of this class decorator copies that global data to the class instance and clears the global. Obviously, this mechanism is not thread-safe, which means that the *creation* of subclasses of Magic should only be done in a single-thread context. Instantiation of the classes has no restrictions. Given that these classes are typically created at IPython startup time and before user application code becomes active, in practice this should not pose any problems. """ cls.registered = True cls.magics = dict(line = magics['line'], cell = magics['cell']) magics['line'] = {} magics['cell'] = {} return cls
def record_magic(dct, magic_kind, magic_name, func): """Utility function to store a function as a magic of a specific kind. Parameters ---------- dct : dict A dictionary with 'line' and 'cell' subdicts. magic_kind : str Kind of magic to be stored. magic_name : str Key to store the magic as. func : function Callable object to store. """ if magic_kind == 'line_cell': dct['line'][magic_name] = dct['cell'][magic_name] = func else: dct[magic_kind][magic_name] = func
def _method_magic_marker(magic_kind): """Decorator factory for methods in Magics subclasses. """ validate_type(magic_kind) # This is a closure to capture the magic_kind. We could also use a class, # but it's overkill for just that one bit of state. def magic_deco(arg): call = lambda f, *a, **k: f(*a, **k) if callable(arg): # "Naked" decorator call (just @foo, no args) func = arg name = func.func_name retval = decorator(call, func) record_magic(magics, magic_kind, name, name) elif isinstance(arg, basestring): # Decorator called with arguments (@foo('bar')) name = arg def mark(func, *a, **kw): record_magic(magics, magic_kind, name, func.func_name) return decorator(call, func) retval = mark else: raise TypeError("Decorator can only be called with " "string or function") return retval # Ensure the resulting decorator has a usable docstring magic_deco.__doc__ = _docstring_template.format('method', magic_kind) return magic_deco
def _function_magic_marker(magic_kind): """Decorator factory for standalone functions. """ validate_type(magic_kind) # This is a closure to capture the magic_kind. We could also use a class, # but it's overkill for just that one bit of state. def magic_deco(arg): call = lambda f, *a, **k: f(*a, **k) # Find get_ipython() in the caller's namespace caller = sys._getframe(1) for ns in ['f_locals', 'f_globals', 'f_builtins']: get_ipython = getattr(caller, ns).get('get_ipython') if get_ipython is not None: break else: raise NameError('Decorator can only run in context where ' '`get_ipython` exists') ip = get_ipython() if callable(arg): # "Naked" decorator call (just @foo, no args) func = arg name = func.func_name ip.register_magic_function(func, magic_kind, name) retval = decorator(call, func) elif isinstance(arg, basestring): # Decorator called with arguments (@foo('bar')) name = arg def mark(func, *a, **kw): ip.register_magic_function(func, magic_kind, name) return decorator(call, func) retval = mark else: raise TypeError("Decorator can only be called with " "string or function") return retval # Ensure the resulting decorator has a usable docstring ds = _docstring_template.format('function', magic_kind) ds += dedent(""" Note: this decorator can only be used in a context where IPython is already active, so that the `get_ipython()` call succeeds. You can therefore use it in your startup files loaded after IPython initializes, but *not* in the IPython configuration file itself, which is executed before IPython is fully up and running. Any file located in the `startup` subdirectory of your configuration profile will be OK in this sense. """) magic_deco.__doc__ = ds return magic_deco
def lsmagic_docs(self, brief=False, missing=''): """Return dict of documentation of magic functions. The return dict has the keys 'line' and 'cell', corresponding to the two types of magics we support. Each value is a dict keyed by magic name whose value is the function docstring. If a docstring is unavailable, the value of `missing` is used instead. If brief is True, only the first line of each docstring will be returned. """ docs = {} for m_type in self.magics: m_docs = {} for m_name, m_func in self.magics[m_type].iteritems(): if m_func.__doc__: if brief: m_docs[m_name] = m_func.__doc__.split('\n', 1)[0] else: m_docs[m_name] = m_func.__doc__.rstrip() else: m_docs[m_name] = missing docs[m_type] = m_docs return docs
def register(self, *magic_objects): """Register one or more instances of Magics. Take one or more classes or instances of classes that subclass the main `core.Magic` class, and register them with IPython to use the magic functions they provide. The registration process will then ensure that any methods that have decorated to provide line and/or cell magics will be recognized with the `%x`/`%%x` syntax as a line/cell magic respectively. If classes are given, they will be instantiated with the default constructor. If your classes need a custom constructor, you should instanitate them first and pass the instance. The provided arguments can be an arbitrary mix of classes and instances. Parameters ---------- magic_objects : one or more classes or instances """ # Start by validating them to ensure they have all had their magic # methods registered at the instance level for m in magic_objects: if not m.registered: raise ValueError("Class of magics %r was constructed without " "the @register_magics class decorator") if type(m) in (type, MetaHasTraits): # If we're given an uninstantiated class m = m(shell=self.shell) # Now that we have an instance, we can register it and update the # table of callables self.registry[m.__class__.__name__] = m for mtype in magic_kinds: self.magics[mtype].update(m.magics[mtype])
def register_function(self, func, magic_kind='line', magic_name=None): """Expose a standalone function as magic function for IPython. This will create an IPython magic (line, cell or both) from a standalone function. The functions should have the following signatures: * For line magics: `def f(line)` * For cell magics: `def f(line, cell)` * For a function that does both: `def f(line, cell=None)` In the latter case, the function will be called with `cell==None` when invoked as `%f`, and with cell as a string when invoked as `%%f`. Parameters ---------- func : callable Function to be registered as a magic. magic_kind : str Kind of magic, one of 'line', 'cell' or 'line_cell' magic_name : optional str If given, the name the magic will have in the IPython namespace. By default, the name of the function itself is used. """ # Create the new method in the user_magics and register it in the # global table validate_type(magic_kind) magic_name = func.func_name if magic_name is None else magic_name setattr(self.user_magics, magic_name, func) record_magic(self.magics, magic_kind, magic_name, func)
def define_magic(self, name, func): """[Deprecated] Expose own function as magic function for IPython. Example:: def foo_impl(self, parameter_s=''): 'My very own magic!. (Use docstrings, IPython reads them).' print 'Magic function. Passed parameter is between < >:' print '<%s>' % parameter_s print 'The self object is:', self ip.define_magic('foo',foo_impl) """ meth = types.MethodType(func, self.user_magics) setattr(self.user_magics, name, meth) record_magic(self.magics, 'line', name, meth)
def format_latex(self, strng): """Format a string for latex inclusion.""" # Characters that need to be escaped for latex: escape_re = re.compile(r'(%|_|\$|#|&)',re.MULTILINE) # Magic command names as headers: cmd_name_re = re.compile(r'^(%s.*?):' % ESC_MAGIC, re.MULTILINE) # Magic commands cmd_re = re.compile(r'(?P<cmd>%s.+?\b)(?!\}\}:)' % ESC_MAGIC, re.MULTILINE) # Paragraph continue par_re = re.compile(r'\\$',re.MULTILINE) # The "\n" symbol newline_re = re.compile(r'\\n') # Now build the string for output: #strng = cmd_name_re.sub(r'\n\\texttt{\\textsl{\\large \1}}:',strng) strng = cmd_name_re.sub(r'\n\\bigskip\n\\texttt{\\textbf{ \1}}:', strng) strng = cmd_re.sub(r'\\texttt{\g<cmd>}',strng) strng = par_re.sub(r'\\\\',strng) strng = escape_re.sub(r'\\\1',strng) strng = newline_re.sub(r'\\textbackslash{}n',strng) return strng
def parse_options(self, arg_str, opt_str, *long_opts, **kw): """Parse options passed to an argument string. The interface is similar to that of getopt(), but it returns back a Struct with the options as keys and the stripped argument string still as a string. arg_str is quoted as a true sys.argv vector by using shlex.split. This allows us to easily expand variables, glob files, quote arguments, etc. Options: -mode: default 'string'. If given as 'list', the argument string is returned as a list (split on whitespace) instead of a string. -list_all: put all option values in lists. Normally only options appearing more than once are put in a list. -posix (True): whether to split the input line in POSIX mode or not, as per the conventions outlined in the shlex module from the standard library.""" # inject default options at the beginning of the input line caller = sys._getframe(1).f_code.co_name arg_str = '%s %s' % (self.options_table.get(caller,''),arg_str) mode = kw.get('mode','string') if mode not in ['string','list']: raise ValueError,'incorrect mode given: %s' % mode # Get options list_all = kw.get('list_all',0) posix = kw.get('posix', os.name == 'posix') strict = kw.get('strict', True) # Check if we have more than one argument to warrant extra processing: odict = {} # Dictionary with options args = arg_str.split() if len(args) >= 1: # If the list of inputs only has 0 or 1 thing in it, there's no # need to look for options argv = arg_split(arg_str, posix, strict) # Do regular option processing try: opts,args = getopt(argv, opt_str, long_opts) except GetoptError,e: raise UsageError('%s ( allowed: "%s" %s)' % (e.msg,opt_str, " ".join(long_opts))) for o,a in opts: if o.startswith('--'): o = o[2:] else: o = o[1:] try: odict[o].append(a) except AttributeError: odict[o] = [odict[o],a] except KeyError: if list_all: odict[o] = [a] else: odict[o] = a # Prepare opts,args for return opts = Struct(odict) if mode == 'string': args = ' '.join(args) return opts,args
def default_option(self, fn, optstr): """Make an entry in the options_table for fn, with value optstr""" if fn not in self.lsmagic(): error("%s is not a magic function" % fn) self.options_table[fn] = optstr
def page_guiref(arg_s=None): """Show a basic reference about the GUI Console.""" from IPython.core import page page.page(gui_reference, auto_html=True)
def get_member(thing_obj, member_string): """Get a member from an object by (string) name""" mems = {x[0]: x[1] for x in inspect.getmembers(thing_obj)} if member_string in mems: return mems[member_string]
def func_from_string(callable_str): """Return a live function from a full dotted path. Must be either a plain function directly in a module, a class function, or a static function. (No modules, classes, or instance methods, since those can't be called as tasks.)""" components = callable_str.split('.') func = None if len(components) < 2: raise ValueError("Need full dotted path to task function") elif len(components) == 2: mod_name = components[0] func_name = components[1] try: mod = import_module(mod_name) except ModuleNotFoundError: raise ValueError(f"Module {mod_name} not found") func = get_member(mod, func_name) if func is None: raise ValueError(f"{func_name} is not a member of {mod_name}") else: mod_name = '.'.join(components[:-1]) func_name = components[-1] try: mod = import_module(mod_name) except ModuleNotFoundError: mod_name = '.'.join(components[:-2]) class_name = components[-2] try: mod = import_module(mod_name) except ModuleNotFoundError: raise ValueError(f"Module {mod_name} not found") klass = get_member(mod, class_name) if klass is None: raise ValueError(f"Class {class_name} is not a member of {mod_name}") func = get_member(klass, func_name) if func is None: raise ValueError(f"Function {func_name} is not a member of {mod_name}.{class_name}") if func is None: func = get_member(mod, func_name) if func is None: raise ValueError(f"Function {func_name} is not a member of {mod_name}") if inspect.ismodule(func): raise ValueError("Cannot call module directly") if inspect.isclass(func): raise ValueError("Cannot call class directly") try: sig = [x for x in inspect.signature(func).parameters] except TypeError: raise ValueError(f"{callable_str} ({str(type(func))[1:-1]}) is not a callable object") if len(sig) == 1: if sig[0] == 'message': return func else: raise ValueError("Task function must have one parameter, named 'message'") elif len(sig)==2 and sig[0]=='self' and sig[1]=='message': # We only check for the conventional 'self', but if you're using something else, # you deserve the pain you'll have trying to debug this. raise ValueError("Can't call instance method without an instance! (Try sisy.models.task_with_callable)") else: raise ValueError("Improper signature for task function (needs only 'message')")
def task_with_callable(the_callable, label=None, schedule=DEFAULT_SCHEDULE, userdata=None, pk_override=None): """Factory function to create a properly initialized task.""" task = Task() if isinstance(the_callable, str): if pk_override is not None: components = the_callable.split('.') info = dict( func_type='instancemethod', module_name='.'.join(components[:-2]), class_name=components[-2], class_path='.'.join(components[:-1]), model_pk=pk_override, func_name=components[-1], func_path=the_callable, ) task.funcinfo = info else: task.funcinfo = get_func_info(func_from_string(the_callable)) else: task.funcinfo = get_func_info(the_callable) if label is None: task.label = task.funcinfo['func_path'] else: task.label = label task.schedule = schedule if not croniter.is_valid(task.schedule): raise ValueError(f"Cron schedule {task.schedule} is not valid") if userdata is None: task.userdata = dict() else: if isinstance(userdata, dict): task.userdata = userdata else: raise ValueError("Userdata must be a dictionary of JSON-serializable data") return task
def taskinfo_with_label(label): """Return task info dictionary from task label. Internal function, pretty much only used in migrations since the model methods aren't there.""" task = Task.objects.get(label=label) info = json.loads(task._func_info) return info
def func_from_info(self): """Find and return a callable object from a task info dictionary""" info = self.funcinfo functype = info['func_type'] if functype in ['instancemethod', 'classmethod', 'staticmethod']: the_modelclass = get_module_member_by_dottedpath(info['class_path']) if functype == 'instancemethod': the_modelobject = the_modelclass.objects.get(pk=info['model_pk']) the_callable = get_member(the_modelobject, info['func_name']) else: the_callable = get_member(the_modelclass, info['func_name']) return the_callable elif functype == 'function': mod = import_module(info['module_name']) the_callable = get_member(mod, info['func_name']) return the_callable else: raise ValueError(f"Unknown functype '{functype} in task {self.pk} ({self.label})")
def run_tasks(cls): """Internal task-runner class method, called by :py:func:`sisy.consumers.run_heartbeat`""" now = timezone.now() tasks = cls.objects.filter(enabled=True) for task in tasks: if task.next_run == HAS_NOT_RUN: task.calc_next_run() if task.next_run < now: if (task.start_running < now): if (task.end_running > now): task.run_asap() else: task.enabled = False task.save() Channel(KILL_TASK_CHANNEL).send({'id': task.pk})
def calc_next_run(self): """Calculate next run time of this task""" base_time = self.last_run if self.last_run == HAS_NOT_RUN: if self.wait_for_schedule is False: self.next_run = timezone.now() self.wait_for_schedule = False # reset so we don't run on every clock tick self.save() return else: base_time = timezone.now() self.next_run = croniter(self.schedule, base_time).get_next(datetime) self.save()
def submit(self, timestamp): """Internal instance method to submit this task for running immediately. Does not handle any iteration, end-date, etc., processing.""" Channel(RUN_TASK_CHANNEL).send({'id':self.pk, 'ts': timestamp.timestamp()})
def run(self, message): """Internal instance method run by worker process to actually run the task callable.""" the_callable = self.func_from_info() try: task_message = dict( task=self, channel_message=message, ) the_callable(task_message) finally: if self.end_running < self.next_run: self.enabled=False Channel(KILL_TASK_CHANNEL).send({'id': self.pk}) return if self.iterations == 0: return else: self.iterations -= 1 if self.iterations == 0: self.enabled = False Channel(KILL_TASK_CHANNEL).send({'id':self.pk}) self.save()
def run_asap(self): """Instance method to run this task immediately.""" now = timezone.now() self.last_run = now self.calc_next_run() self.save() self.submit(now)
def run_iterations(cls, the_callable, iterations=1, label=None, schedule='* * * * * *', userdata = None, run_immediately=False, delay_until=None): """Class method to run a callable with a specified number of iterations""" task = task_with_callable(the_callable, label=label, schedule=schedule, userdata=userdata) task.iterations = iterations if delay_until is not None: if isinstance(delay_until, datetime): if delay_until > timezone.now(): task.start_running = delay_until else: raise ValueError("Task cannot start running in the past") else: raise ValueError("delay_until must be a datetime.datetime instance") if run_immediately: task.next_run = timezone.now() else: task.calc_next_run() task.save()
def run_once(cls, the_callable, userdata=None, delay_until=None): """Class method to run a one-shot task, immediately.""" cls.run_iterations(the_callable, userdata=userdata, run_immediately=True, delay_until=delay_until)
def find_url_file(self): """Set the url file. Here we don't try to actually see if it exists for is valid as that is hadled by the connection logic. """ config = self.config # Find the actual controller key file if not self.url_file: self.url_file = os.path.join( self.profile_dir.security_dir, self.url_file_name )
def load_connector_file(self): """load config from a JSON connector file, at a *lower* priority than command-line/config files. """ self.log.info("Loading url_file %r", self.url_file) config = self.config with open(self.url_file) as f: d = json.loads(f.read()) if 'exec_key' in d: config.Session.key = cast_bytes(d['exec_key']) try: config.EngineFactory.location except AttributeError: config.EngineFactory.location = d['location'] d['url'] = disambiguate_url(d['url'], config.EngineFactory.location) try: config.EngineFactory.url except AttributeError: config.EngineFactory.url = d['url'] try: config.EngineFactory.sshserver except AttributeError: config.EngineFactory.sshserver = d['ssh']
def bind_kernel(self, **kwargs): """Promote engine to listening kernel, accessible to frontends.""" if self.kernel_app is not None: return self.log.info("Opening ports for direct connections as an IPython kernel") kernel = self.kernel kwargs.setdefault('config', self.config) kwargs.setdefault('log', self.log) kwargs.setdefault('profile_dir', self.profile_dir) kwargs.setdefault('session', self.engine.session) app = self.kernel_app = IPKernelApp(**kwargs) # allow IPKernelApp.instance(): IPKernelApp._instance = app app.init_connection_file() # relevant contents of init_sockets: app.shell_port = app._bind_socket(kernel.shell_streams[0], app.shell_port) app.log.debug("shell ROUTER Channel on port: %i", app.shell_port) app.iopub_port = app._bind_socket(kernel.iopub_socket, app.iopub_port) app.log.debug("iopub PUB Channel on port: %i", app.iopub_port) kernel.stdin_socket = self.engine.context.socket(zmq.ROUTER) app.stdin_port = app._bind_socket(kernel.stdin_socket, app.stdin_port) app.log.debug("stdin ROUTER Channel on port: %i", app.stdin_port) # start the heartbeat, and log connection info: app.init_heartbeat() app.log_connection_info() app.write_connection_file()
def pid_exists(pid): """Check whether pid exists in the current process table.""" if not isinstance(pid, int): raise TypeError('an integer is required') if pid < 0: return False try: os.kill(pid, 0) except OSError: e = sys.exc_info()[1] return e.errno == errno.EPERM else: return True
def get_disk_usage(path): """Return disk usage associated with path.""" st = os.statvfs(path) free = (st.f_bavail * st.f_frsize) total = (st.f_blocks * st.f_frsize) used = (st.f_blocks - st.f_bfree) * st.f_frsize percent = usage_percent(used, total, _round=1) # NB: the percentage is -5% than what shown by df due to # reserved blocks that we are currently not considering: # http://goo.gl/sWGbH return nt_diskinfo(total, used, free, percent)
def timid(ctxt, test, key=None, check=False, exts=None): """ Execute a test described by a YAML file. :param ctxt: A ``timid.context.Context`` object. :param test: The name of a YAML file containing the test description. Note that the current working directory set up in ``ctxt.environment`` does not affect the resolution of this file. :param key: An optional key into the test description file. If not ``None``, the file named by ``test`` must be a YAML dictionary of lists of steps; otherwise, it must be a simple list of steps. :param check: If ``True``, only performs a syntax check of the test steps indicated by ``test`` and ``key``; the test itself is not run. :param exts: An instance of ``timid.extensions.ExtensionSet`` describing the extensions to be called while processing the test steps. """ # Normalize the extension set if exts is None: exts = extensions.ExtensionSet() # Begin by reading the steps and adding them to the list in the # context (which may already have elements thanks to the # extensions) ctxt.emit('Reading test steps from %s%s...' % (test, '[%s]' % key if key else ''), debug=True) ctxt.steps += exts.read_steps(ctxt, steps.Step.parse_file(ctxt, test, key)) # If all we were supposed to do was check, well, we've # accomplished that... if check: return None # Now we execute each step in turn for idx, step in enumerate(ctxt.steps): # Emit information about what we're doing ctxt.emit('[Step %d]: %s . . .' % (idx, step.name)) # Run through extension hooks if exts.pre_step(ctxt, step, idx): ctxt.emit('[Step %d]: `- Step %s' % (idx, steps.states[steps.SKIPPED])) continue # Now execute the step result = step(ctxt) # Let the extensions process the result of the step exts.post_step(ctxt, step, idx, result) # Emit the result ctxt.emit('[Step %d]: `- Step %s%s' % (idx, steps.states[result.state], ' (ignored)' if result.ignore else '')) # Was the step a success? if not result: msg = 'Test step failure' if result.msg: msg += ': %s' % result.msg return msg # All done! And a success, to boot... return None
def _processor(args): """ A ``cli_tools`` processor function that interfaces between the command line and the ``timid()`` function. This function is responsible for allocating a ``timid.context.Context`` object and initializing the activated extensions, and for calling those extensions' ``finalize()`` method. :param args: The ``argparse.Namespace`` object containing the results of argument processing. """ # Begin by initializing a context args.ctxt = context.Context(args.verbose, args.debug, args.directory) # Now set up the extension set args.exts = extensions.ExtensionSet.activate(args.ctxt, args) # Update the environment and the variables args.ctxt.environment.update(args.environment) args.ctxt.variables.update(args.variables) # Call the actual timid() function try: result = yield # If an exception occurred, give the extensions an opportunity to # handle it except Exception as exc: if args.debug: # Make sure we emit a proper traceback traceback.print_exc(file=sys.stderr) # The exception is the result, from the point of view of the # extensions result = exc # Allow the extensions to handle the result result = args.exts.finalize(args.ctxt, result) # If the final result is an exception, convert it to a string for # yielding back to cli_tools if isinstance(result, Exception): result = str(result) # This line is covered, but coverage appears to be missing it for # some reason yield result
def create_lang_instance(var_map = None): """ >>> lang_instance = create_lang_instance() >>> lang_instance.aml_evaluate(lang_instance.aml_compile('1 = 1')) True >>> li = create_lang_instance() >>> c = li.aml_compile >>> e = li.aml_evaluate >>> p = li.aml_translate_python >>> s = li.aml_translate_sql >>> u = li.aml_suggest >>> e(c('1 = 0')) False >>> e(c('"1" = "1"')) True >>> e(c('(1=1)')) True >>> e(c('1 > 1')) False >>> e(c('not 1 > 1')) True >>> e(c('1 != 1')) False >>> e(c('-2 = -2')) True >>> eval(p(c('-2 = -2'))) True >>> eval(p(c('-2 >= -1'))) False >>> eval(p(c('-2 <= -1'))) True >>> eval(p(c('2 >= 1'))) True >>> eval(p(c('2 <= 1'))) False >>> eval(p(c('null = null'))) True >>> eval(p(c('1 = null'))) False >>> e(c('"foo" = "foo"')) True >>> e(c('"foo"' '=' "'foo'")) True >>> e(c('"foo" = \\'foo\\'')) True >>> e(c('"fo\\'o" = "fo\\'o"')) True >>> e(c("'foo'" + '=' + '"foo"')) True >>> li = create_lang_instance({'foo' : 1}); >>> c = li.aml_compile >>> e = li.aml_evaluate >>> e(c('foo = 1')) True >>> li = create_lang_instance({'foo' : 1.00}) >>> c = li.aml_compile >>> e = li.aml_evaluate >>> e(c('foo = 1')) True >>> li = create_lang_instance({'foo' : 2.24}) >>> c = li.aml_compile >>> e = li.aml_evaluate >>> e(c('foo = 2.24')) True >>> e(c('foo = 2.2399 or foo = 2.24')) True >>> e(c('foo = 2.2399 or foo = 2.2401')) False >>> e(c('foo in (2.2399, 2.24, null,)')) True >>> e(c('foo in (2.2399, 2.2401, null,)')) False >>> e(c('null in (2.2399, 2.2401, null)')) True >>> e(c('"null" in (2.2399, 2.2401, null)')) False >>> e(c('"null"' 'in' "(2.2399, 'null', null)")) True >>> li = create_lang_instance({'foo' : 'foo'}) >>> c = li.aml_compile >>> e = li.aml_evaluate >>> e(c('foo = "foo"')) True >>> li = create_lang_instance() >>> c = li.aml_compile >>> p = li.aml_translate_python >>> s = li.aml_translate_sql >>> s(c('null = null')) u'null is null' >>> p(c('null = null')) u'None == None' >>> s(c('null != null')) u'null is not null' >>> p(c('null != null')) u'None != None' >>> s(c('5 != 3')) u'5 <> 3' >>> p(c('5 != 3')) u'5 != 3' >>> p(c('5 in (3, 4, 5)')) u'5 in (3, 4, 5,)' >>> p(s('5 in (3, 4, 5)')) u'5 in (3, 4, 5)' >>> li = create_lang_instance({'foo' : 'bar', 'fo2' : 'ba2'}) >>> c = li.aml_compile >>> p = li.aml_translate_python >>> e = li.aml_evaluate >>> u = li.aml_suggest >>> u('1 = fo') [u'fo2', u'foo'] >>> u('1 = FO') [u'fo2', u'foo'] >>> p(c('null = null')) u'None == None' >>> e(c('foo = "bar"')) True >>> e(c('fo2 = "ba2"')) True """ def py_bool_to_lit(py_bool): return parse( 'true' if py_bool else 'false', BooleanLiteral) if not var_map: class Identifier(str): grammar = re.compile(r'$a') # This will match nothing. else: class Identifier(Keyword): grammar = Enum(*[K(v) for v in var_map.iterkeys()]) class StringLiteral(str): def __new__(cls, s): return super(StringLiteral, cls).__new__(cls, ast.literal_eval(s)) grammar = [re.compile(r'"[^\\\n\r]+?"'), re.compile(r"'[^\\\n\r]+?'")] class IntegerLiteral(int): grammar = re.compile(r'-?\d+') class FloatLiteral(float): grammar = re.compile(r'-?\d+.\d+') class BooleanLiteral(Keyword): grammar = Enum(K('true'), K('false')) class NullLiteral(Keyword): grammar = Enum(K('null')) Comparable = [NullLiteral, FloatLiteral, IntegerLiteral, StringLiteral, Identifier] class ListOfComparables(List): pass ListOfComparables.grammar = ( '(', Comparable, maybe_some( ',', blank, Comparable, ), optional(','), ')' ) class ComparisonOperator(str): grammar = re.compile(r'=|>=|<=|>|<|!=|in') class BooleanFunctionName(Keyword): grammar = Enum(K('and'), K('or')) class ComparisonOperation(List): pass ComparisonOperation.grammar = ( Comparable, blank, attr('comp_op', ComparisonOperator), blank, [Comparable, ListOfComparables], ) class BooleanOperationSimple(List): # The flag() pypeg2 function works great when parsing but does not work when # composing (the flag gets output whether it was in the source text or not). So # a workaround is this: grammar = ( attr('negated', optional(RE_NOT)), ComparisonOperation, ) class BooleanOperation(List): pass BooleanOperation.grammar = ( BooleanOperationSimple, maybe_some( blank, BooleanFunctionName, blank, BooleanOperationSimple, ), ) class Expression(List): pass Expression.grammar = ( [BooleanOperationSimple, ('(', Expression, ')')], maybe_some( blank, BooleanFunctionName, blank, [BooleanOperationSimple, ('(', Expression, ')')], ), ) def eval_node(node): en = lambda n: eval_node(n) if isinstance(node, Identifier): return var_map[node] elif isinstance(node, StringLiteral): return node elif isinstance(node, IntegerLiteral): return node elif isinstance(node, FloatLiteral): return node elif isinstance(node, BooleanLiteral): if node == 'true': return True elif node == 'false': return False elif isinstance(node, NullLiteral): return None elif isinstance(node, ListOfComparables): return node elif isinstance(node, ComparisonOperation): opa, opb = node[0:2] if node.comp_op == '=': return en(opa) == en(opb) elif node.comp_op == '>': return en(opa) > en(opb) elif node.comp_op == '<': return en(opa) < en(opb) elif node.comp_op == '!=': return en(opa) != en(opb) elif node.comp_op == '>=': return en(opa) >= en(opb) elif node.comp_op == '<=': return en(opa) <= en(opb) elif node.comp_op == 'in': enopa = en(opa) enopb = en(opb) for other_node in list(enopb): virtual_node = ComparisonOperation([opa, other_node]) virtual_node.comp_op = '=' if en(virtual_node): return True return False elif isinstance(node, BooleanOperationSimple): a = en(node[0]) if node.negated: a = not a return a elif isinstance(node, BooleanOperation): if len(node) == 1: return en(node[0]) fn_map = { 'and': lambda a,b: a and b, 'or': lambda a,b: a or b, } def simple_eval(tr): return py_bool_to_lit(fn_map[tr[1]]( en(tr[0]), en(tr[2]))) for fname in ['and', 'or']: for i in xrange(1, len(node), 2): if node[i] == fname: new_self = ( node[:i-1] + [simple_eval(node[i-1:i+2])] + node[i+2:] ) return en(BooleanOperation(new_self)) elif isinstance(node, Expression): def iter_over_relevant(): for eli, el in enumerate(node): if eli % 2 == 0: yield eli, el if all( isinstance(el, BooleanOperationSimple) or isinstance(el, BooleanLiteral) for eli, el in iter_over_relevant() ): res = en(BooleanOperation(node)) return res else: for eli, el in iter_over_relevant(): if isinstance(el, Expression): new_self = ( node[:eli] + [py_bool_to_lit(en(el))] + node[eli+1:] ) return en(Expression(new_self)) def compose_node_to_python(node): return compose(node) def compose_node_to_sql(node): return compose(node) def aml_compile(source): return parse(source, Expression) def aml_evaluate(aml_c): result = eval_node(aml_c) return result def aml_translate_python(aml_c): def comp_op_compose(self, *args, **kwargs): if self == '=': return '==' else: return self def null_compose(self, *args, **kwargs): return 'None' def string_compose(self, *args, **kwargs): return '"' + self.replace('"', r'\"') + '"' ComparisonOperator.compose = comp_op_compose NullLiteral.compose = null_compose StringLiteral.compose = string_compose result = compose_node_to_python(aml_c) delattr(ComparisonOperator, 'compose') delattr(NullLiteral, 'compose') delattr(StringLiteral, 'compose') return result def aml_translate_sql(aml_c): def comp_op_compose(self, *args, **kwargs): if self == '!=': return '<>' else: return self def null_compose(self, *args, **kwargs): return 'null' def string_compose(self, *args, **kwargs): return "'" + self.replace("'", "''") + "'" def comp_operation_compose(self, *args, **kwargs): if ( ( isinstance(self[0], NullLiteral) or isinstance(self[1], NullLiteral) ) and ( self.comp_op in ('=', '!=') ) ): if self.comp_op == '=': middle = 'is' else: middle = 'is not' else: middle = compose(self.comp_op) return ' '.join([ compose(self[0]), middle, compose(self[1]), ]) ComparisonOperator.compose = comp_op_compose NullLiteral.compose = null_compose ComparisonOperation.compose = comp_operation_compose StringLiteral.compose = string_compose result = compose_node_to_sql(aml_c) delattr(ComparisonOperator, 'compose') delattr(NullLiteral, 'compose') delattr(ComparisonOperation, 'compose') delattr(StringLiteral, 'compose') return result def aml_suggest(source): suggestions = [ ] if var_map: if not source: suggestions = list(var_map.iterkeys()) else: split = [el for el in re.split(r'(?m)\s+', source) if el] if split: for candidate in var_map.iterkeys(): if candidate.lower().startswith(split[-1].lower()): suggestions.append(candidate) suggestions.sort() return suggestions lang_instance = LangInstance() lang_instance.aml_compile = aml_compile lang_instance.aml_evaluate = aml_evaluate lang_instance.aml_translate_python = aml_translate_python lang_instance.aml_translate_sql = aml_translate_sql lang_instance.aml_suggest = aml_suggest return lang_instance
def create_interrupt_event(): """ Create an interrupt event handle. The parent process should use this static method for creating the interrupt event that is passed to the child process. It should store this handle and use it with ``send_interrupt`` to interrupt the child process. """ # Create a security attributes struct that permits inheritance of the # handle by new processes. # FIXME: We can clean up this mess by requiring pywin32 for IPython. class SECURITY_ATTRIBUTES(ctypes.Structure): _fields_ = [ ("nLength", ctypes.c_int), ("lpSecurityDescriptor", ctypes.c_void_p), ("bInheritHandle", ctypes.c_int) ] sa = SECURITY_ATTRIBUTES() sa_p = ctypes.pointer(sa) sa.nLength = ctypes.sizeof(SECURITY_ATTRIBUTES) sa.lpSecurityDescriptor = 0 sa.bInheritHandle = 1 return ctypes.windll.kernel32.CreateEventA( sa_p, # lpEventAttributes False, # bManualReset False, # bInitialState '')
def run(self): """ Run the poll loop. This method never returns. """ try: from _winapi import WAIT_OBJECT_0, INFINITE except ImportError: from _subprocess import WAIT_OBJECT_0, INFINITE # Build the list of handle to listen on. handles = [] if self.interrupt_handle: handles.append(self.interrupt_handle) if self.parent_handle: handles.append(self.parent_handle) arch = platform.architecture()[0] c_int = ctypes.c_int64 if arch.startswith('64') else ctypes.c_int # Listen forever. while True: result = ctypes.windll.kernel32.WaitForMultipleObjects( len(handles), # nCount (c_int * len(handles))(*handles), # lpHandles False, # bWaitAll INFINITE) # dwMilliseconds if WAIT_OBJECT_0 <= result < len(handles): handle = handles[result - WAIT_OBJECT_0] if handle == self.interrupt_handle: interrupt_main() elif handle == self.parent_handle: os._exit(1) elif result < 0: # wait failed, just give up and stop polling. warn("""Parent poll failed. If the frontend dies, the kernel may be left running. Please let us know about your system (bitness, Python, etc.) at ipython-dev@scipy.org""") return
def initialize(): """ Function to initialize settings from command line and/or custom settings file :return: Returns str with operation type """ if len(sys.argv) == 1: usage() sys.exit() command = _get_command(sys.argv[1]) try: opts, args = getopt.getopt(sys.argv[2:], 'h:e:p:u:l:P:s:m:', ['help', 'email=', 'password=', 'url=', 'locale=', 'po-path=', 'settings=', 'message=']) except getopt.GetoptError: usage() sys.exit() params = _get_params_from_options(opts) _set_settings_file(settings, params) if command == 'push': if 'GIT_MESSAGE' in params: return 'push', params['GIT_MESSAGE'] return 'push', None return command, None
def create_typestr2type_dicts(dont_include_in_type2typestr=["lambda"]): """Return dictionaries mapping lower case typename (e.g. 'tuple') to type objects from the types package, and vice versa.""" typenamelist = [tname for tname in dir(types) if tname.endswith("Type")] typestr2type, type2typestr = {}, {} for tname in typenamelist: name = tname[:-4].lower() # Cut 'Type' off the end of the name obj = getattr(types, tname) typestr2type[name] = obj if name not in dont_include_in_type2typestr: type2typestr[obj] = name return typestr2type, type2typestr
def is_type(obj, typestr_or_type): """is_type(obj, typestr_or_type) verifies if obj is of a certain type. It can take strings or actual python types for the second argument, i.e. 'tuple'<->TupleType. 'all' matches all types. TODO: Should be extended for choosing more than one type.""" if typestr_or_type == "all": return True if type(typestr_or_type) == types.TypeType: test_type = typestr_or_type else: test_type = typestr2type.get(typestr_or_type, False) if test_type: return isinstance(obj, test_type) return False
def dict_dir(obj): """Produce a dictionary of an object's attributes. Builds on dir2 by checking that a getattr() call actually succeeds.""" ns = {} for key in dir2(obj): # This seemingly unnecessary try/except is actually needed # because there is code out there with metaclasses that # create 'write only' attributes, where a getattr() call # will fail even if the attribute appears listed in the # object's dictionary. Properties can actually do the same # thing. In particular, Traits use this pattern try: ns[key] = getattr(obj, key) except AttributeError: pass return ns
def filter_ns(ns, name_pattern="*", type_pattern="all", ignore_case=True, show_all=True): """Filter a namespace dictionary by name pattern and item type.""" pattern = name_pattern.replace("*",".*").replace("?",".") if ignore_case: reg = re.compile(pattern+"$", re.I) else: reg = re.compile(pattern+"$") # Check each one matches regex; shouldn't be hidden; of correct type. return dict((key,obj) for key, obj in ns.iteritems() if reg.match(key) \ and show_hidden(key, show_all) \ and is_type(obj, type_pattern) )
def list_namespace(namespace, type_pattern, filter, ignore_case=False, show_all=False): """Return dictionary of all objects in a namespace dictionary that match type_pattern and filter.""" pattern_list=filter.split(".") if len(pattern_list) == 1: return filter_ns(namespace, name_pattern=pattern_list[0], type_pattern=type_pattern, ignore_case=ignore_case, show_all=show_all) else: # This is where we can change if all objects should be searched or # only modules. Just change the type_pattern to module to search only # modules filtered = filter_ns(namespace, name_pattern=pattern_list[0], type_pattern="all", ignore_case=ignore_case, show_all=show_all) results = {} for name, obj in filtered.iteritems(): ns = list_namespace(dict_dir(obj), type_pattern, ".".join(pattern_list[1:]), ignore_case=ignore_case, show_all=show_all) for inner_name, inner_obj in ns.iteritems(): results["%s.%s"%(name,inner_name)] = inner_obj return results
def mutex_opts(dict,ex_op): """Check for presence of mutually exclusive keys in a dict. Call: mutex_opts(dict,[[op1a,op1b],[op2a,op2b]...]""" for op1,op2 in ex_op: if op1 in dict and op2 in dict: raise ValueError,'\n*** ERROR in Arguments *** '\ 'Options '+op1+' and '+op2+' are mutually exclusive.'
def map_method(method,object_list,*argseq,**kw): """map_method(method,object_list,*args,**kw) -> list Return a list of the results of applying the methods to the items of the argument sequence(s). If more than one sequence is given, the method is called with an argument list consisting of the corresponding item of each sequence. All sequences must be of the same length. Keyword arguments are passed verbatim to all objects called. This is Python code, so it's not nearly as fast as the builtin map().""" out_list = [] idx = 0 for object in object_list: try: handler = getattr(object, method) except AttributeError: out_list.append(None) else: if argseq: args = map(lambda lst:lst[idx],argseq) #print 'ob',object,'hand',handler,'ar',args # dbg out_list.append(handler(args,**kw)) else: out_list.append(handler(**kw)) idx += 1 return out_list
def popkey(dct,key,default=NotGiven): """Return dct[key] and delete dct[key]. If default is given, return it if dct[key] doesn't exist, otherwise raise KeyError. """ try: val = dct[key] except KeyError: if default is NotGiven: raise else: return default else: del dct[key] return val
def show(close=None): """Show all figures as SVG/PNG payloads sent to the IPython clients. Parameters ---------- close : bool, optional If true, a ``plt.close('all')`` call is automatically issued after sending all the figures. If this is set, the figures will entirely removed from the internal list of figures. """ if close is None: close = InlineBackend.instance().close_figures try: for figure_manager in Gcf.get_all_fig_managers(): send_figure(figure_manager.canvas.figure) finally: show._to_draw = [] if close: matplotlib.pyplot.close('all')
def draw_if_interactive(): """ Is called after every pylab drawing command """ # signal that the current active figure should be sent at the end of # execution. Also sets the _draw_called flag, signaling that there will be # something to send. At the end of the code execution, a separate call to # flush_figures() will act upon these values fig = Gcf.get_active().canvas.figure # Hack: matplotlib FigureManager objects in interacive backends (at least # in some of them) monkeypatch the figure object and add a .show() method # to it. This applies the same monkeypatch in order to support user code # that might expect `.show()` to be part of the official API of figure # objects. # For further reference: # https://github.com/ipython/ipython/issues/1612 # https://github.com/matplotlib/matplotlib/issues/835 if not hasattr(fig, 'show'): # Queue up `fig` for display fig.show = lambda *a: send_figure(fig) # If matplotlib was manually set to non-interactive mode, this function # should be a no-op (otherwise we'll generate duplicate plots, since a user # who set ioff() manually expects to make separate draw/show calls). if not matplotlib.is_interactive(): return # ensure current figure will be drawn, and each subsequent call # of draw_if_interactive() moves the active figure to ensure it is # drawn last try: show._to_draw.remove(fig) except ValueError: # ensure it only appears in the draw list once pass # Queue up the figure for drawing in next show() call show._to_draw.append(fig) show._draw_called = True
def flush_figures(): """Send all figures that changed This is meant to be called automatically and will call show() if, during prior code execution, there had been any calls to draw_if_interactive. This function is meant to be used as a post_execute callback in IPython, so user-caused errors are handled with showtraceback() instead of being allowed to raise. If this function is not called from within IPython, then these exceptions will raise. """ if not show._draw_called: return if InlineBackend.instance().close_figures: # ignore the tracking, just draw and close all figures try: return show(True) except Exception as e: # safely show traceback if in IPython, else raise try: get_ipython except NameError: raise e else: get_ipython().showtraceback() return try: # exclude any figures that were closed: active = set([fm.canvas.figure for fm in Gcf.get_all_fig_managers()]) for fig in [ fig for fig in show._to_draw if fig in active ]: try: send_figure(fig) except Exception as e: # safely show traceback if in IPython, else raise try: get_ipython except NameError: raise e else: get_ipython().showtraceback() break finally: # clear flags for next round show._to_draw = [] show._draw_called = False
def send_figure(fig): """Draw the given figure and send it as a PNG payload. """ fmt = InlineBackend.instance().figure_format data = print_figure(fig, fmt) # print_figure will return None if there's nothing to draw: if data is None: return mimetypes = { 'png' : 'image/png', 'svg' : 'image/svg+xml' } mime = mimetypes[fmt] # flush text streams before sending figures, helps a little with output # synchronization in the console (though it's a bandaid, not a real sln) sys.stdout.flush(); sys.stderr.flush() publish_display_data( 'IPython.zmq.pylab.backend_inline.send_figure', {mime : data} )
def load_extension(self, module_str): """Load an IPython extension by its module name. If :func:`load_ipython_extension` returns anything, this function will return that object. """ from IPython.utils.syspathcontext import prepended_to_syspath if module_str not in sys.modules: with prepended_to_syspath(self.ipython_extension_dir): __import__(module_str) mod = sys.modules[module_str] return self._call_load_ipython_extension(mod)
def unload_extension(self, module_str): """Unload an IPython extension by its module name. This function looks up the extension's name in ``sys.modules`` and simply calls ``mod.unload_ipython_extension(self)``. """ if module_str in sys.modules: mod = sys.modules[module_str] self._call_unload_ipython_extension(mod)
def install_extension(self, url, filename=None): """Download and install an IPython extension. If filename is given, the file will be so named (inside the extension directory). Otherwise, the name from the URL will be used. The file must have a .py or .zip extension; otherwise, a ValueError will be raised. Returns the full path to the installed file. """ # Ensure the extension directory exists if not os.path.isdir(self.ipython_extension_dir): os.makedirs(self.ipython_extension_dir, mode = 0777) if os.path.isfile(url): src_filename = os.path.basename(url) copy = copyfile else: src_filename = urlparse(url).path.split('/')[-1] copy = urlretrieve if filename is None: filename = src_filename if os.path.splitext(filename)[1] not in ('.py', '.zip'): raise ValueError("The file must have a .py or .zip extension", filename) filename = os.path.join(self.ipython_extension_dir, filename) copy(url, filename) return filename
def externals_finder(dirname, filename): """Find any 'svn:externals' directories""" found = False f = open(filename,'rt') for line in iter(f.readline, ''): # can't use direct iter! parts = line.split() if len(parts)==2: kind,length = parts data = f.read(int(length)) if kind=='K' and data=='svn:externals': found = True elif kind=='V' and found: f.close() break else: f.close() return for line in data.splitlines(): parts = line.split() if parts: yield joinpath(dirname, parts[0])
def random_ports(port, n): """Generate a list of n random ports near the given port. The first 5 ports will be sequential, and the remaining n-5 will be randomly selected in the range [port-2*n, port+2*n]. """ for i in range(min(5, n)): yield port + i for i in range(n-5): yield port + random.randint(-2*n, 2*n)
def init_webapp(self): """initialize tornado webapp and httpserver""" self.web_app = NotebookWebApplication( self, self.kernel_manager, self.notebook_manager, self.cluster_manager, self.log, self.base_project_url, self.webapp_settings ) if self.certfile: ssl_options = dict(certfile=self.certfile) if self.keyfile: ssl_options['keyfile'] = self.keyfile else: ssl_options = None self.web_app.password = self.password self.http_server = httpserver.HTTPServer(self.web_app, ssl_options=ssl_options) if ssl_options is None and not self.ip and not (self.read_only and not self.password): self.log.critical('WARNING: the notebook server is listening on all IP addresses ' 'but not using any encryption or authentication. This is highly ' 'insecure and not recommended.') success = None for port in random_ports(self.port, self.port_retries+1): try: self.http_server.listen(port, self.ip) except socket.error, e: if e.errno != errno.EADDRINUSE: raise self.log.info('The port %i is already in use, trying another random port.' % port) else: self.port = port success = True break if not success: self.log.critical('ERROR: the notebook server could not be started because ' 'no available port could be found.') self.exit(1)
def _handle_sigint(self, sig, frame): """SIGINT handler spawns confirmation dialog""" # register more forceful signal handler for ^C^C case signal.signal(signal.SIGINT, self._signal_stop) # request confirmation dialog in bg thread, to avoid # blocking the App thread = threading.Thread(target=self._confirm_exit) thread.daemon = True thread.start()
def _confirm_exit(self): """confirm shutdown on ^C A second ^C, or answering 'y' within 5s will cause shutdown, otherwise original SIGINT handler will be restored. This doesn't work on Windows. """ # FIXME: remove this delay when pyzmq dependency is >= 2.1.11 time.sleep(0.1) sys.stdout.write("Shutdown Notebook Server (y/[n])? ") sys.stdout.flush() r,w,x = select.select([sys.stdin], [], [], 5) if r: line = sys.stdin.readline() if line.lower().startswith('y'): self.log.critical("Shutdown confirmed") ioloop.IOLoop.instance().stop() return else: print "No answer for 5s:", print "resuming operation..." # no answer, or answer is no: # set it back to original SIGINT handler # use IOLoop.add_callback because signal.signal must be called # from main thread ioloop.IOLoop.instance().add_callback(self._restore_sigint_handler)
def cleanup_kernels(self): """shutdown all kernels The kernels will shutdown themselves when this process no longer exists, but explicit shutdown allows the KernelManagers to cleanup the connection files. """ self.log.info('Shutting down kernels') km = self.kernel_manager # copy list, since shutdown_kernel deletes keys for kid in list(km.kernel_ids): km.shutdown_kernel(kid)
def price_options(S=100.0, K=100.0, sigma=0.25, r=0.05, days=260, paths=10000): """ Price European and Asian options using a Monte Carlo method. Parameters ---------- S : float The initial price of the stock. K : float The strike price of the option. sigma : float The volatility of the stock. r : float The risk free interest rate. days : int The number of days until the option expires. paths : int The number of Monte Carlo paths used to price the option. Returns ------- A tuple of (E. call, E. put, A. call, A. put) option prices. """ import numpy as np from math import exp,sqrt h = 1.0/days const1 = exp((r-0.5*sigma**2)*h) const2 = sigma*sqrt(h) stock_price = S*np.ones(paths, dtype='float64') stock_price_sum = np.zeros(paths, dtype='float64') for j in range(days): growth_factor = const1*np.exp(const2*np.random.standard_normal(paths)) stock_price = stock_price*growth_factor stock_price_sum = stock_price_sum + stock_price stock_price_avg = stock_price_sum/days zeros = np.zeros(paths, dtype='float64') r_factor = exp(-r*h*days) euro_put = r_factor*np.mean(np.maximum(zeros, K-stock_price)) asian_put = r_factor*np.mean(np.maximum(zeros, K-stock_price_avg)) euro_call = r_factor*np.mean(np.maximum(zeros, stock_price-K)) asian_call = r_factor*np.mean(np.maximum(zeros, stock_price_avg-K)) return (euro_call, euro_put, asian_call, asian_put)
def multiple_replace(dict, text): """ Replace in 'text' all occurences of any key in the given dictionary by its corresponding value. Returns the new string.""" # Function by Xavier Defrang, originally found at: # http://aspn.activestate.com/ASPN/Cookbook/Python/Recipe/81330 # Create a regular expression from the dictionary keys regex = re.compile("(%s)" % "|".join(map(re.escape, dict.keys()))) # For each match, look-up corresponding value in dictionary return regex.sub(lambda mo: dict[mo.string[mo.start():mo.end()]], text)
def cwd_filt(depth): """Return the last depth elements of the current working directory. $HOME is always replaced with '~'. If depth==0, the full path is returned.""" cwd = os.getcwdu().replace(HOME,"~") out = os.sep.join(cwd.split(os.sep)[-depth:]) return out or os.sep
def cwd_filt2(depth): """Return the last depth elements of the current working directory. $HOME is always replaced with '~'. If depth==0, the full path is returned.""" full_cwd = os.getcwdu() cwd = full_cwd.replace(HOME,"~").split(os.sep) if '~' in cwd and len(cwd) == depth+1: depth += 1 drivepart = '' if sys.platform == 'win32' and len(cwd) > depth: drivepart = os.path.splitdrive(full_cwd)[0] out = drivepart + '/'.join(cwd[-depth:]) return out or os.sep
def update_prompt(self, name, new_template=None): """This is called when a prompt template is updated. It processes abbreviations used in the prompt template (like \#) and calculates how many invisible characters (ANSI colour escapes) the resulting prompt contains. It is also called for each prompt on changing the colour scheme. In both cases, traitlets should take care of calling this automatically. """ if new_template is not None: self.templates[name] = multiple_replace(prompt_abbreviations, new_template) # We count invisible characters (colour escapes) on the last line of the # prompt, to calculate the width for lining up subsequent prompts. invis_chars = _lenlastline(self._render(name, color=True)) - \ _lenlastline(self._render(name, color=False)) self.invisible_chars[name] = invis_chars
def _render(self, name, color=True, **kwargs): """Render but don't justify, or update the width or txtwidth attributes. """ if name == 'rewrite': return self._render_rewrite(color=color) if color: scheme = self.color_scheme_table.active_colors if name=='out': colors = color_lists['normal'] colors.number, colors.prompt, colors.normal = \ scheme.out_number, scheme.out_prompt, scheme.normal else: colors = color_lists['inp'] colors.number, colors.prompt, colors.normal = \ scheme.in_number, scheme.in_prompt, scheme.in_normal if name=='in2': colors.prompt = scheme.in_prompt2 else: # No color colors = color_lists['nocolor'] colors.number, colors.prompt, colors.normal = '', '', '' count = self.shell.execution_count # Shorthand # Build the dictionary to be passed to string formatting fmtargs = dict(color=colors, count=count, dots="."*len(str(count)), width=self.width, txtwidth=self.txtwidth ) fmtargs.update(self.lazy_evaluate_fields) fmtargs.update(kwargs) # Prepare the prompt prompt = colors.prompt + self.templates[name] + colors.normal # Fill in required fields return self._formatter.format(prompt, **fmtargs)
def _render_rewrite(self, color=True): """Render the ---> rewrite prompt.""" if color: scheme = self.color_scheme_table.active_colors # We need a non-input version of these escapes color_prompt = scheme.in_prompt.replace("\001","").replace("\002","") color_normal = scheme.normal else: color_prompt, color_normal = '', '' return color_prompt + "-> ".rjust(self.txtwidth, "-") + color_normal
def render(self, name, color=True, just=None, **kwargs): """ Render the selected prompt. Parameters ---------- name : str Which prompt to render. One of 'in', 'in2', 'out', 'rewrite' color : bool If True (default), include ANSI escape sequences for a coloured prompt. just : bool If True, justify the prompt to the width of the last prompt. The default is stored in self.justify. **kwargs : Additional arguments will be passed to the string formatting operation, so they can override the values that would otherwise fill in the template. Returns ------- A string containing the rendered prompt. """ res = self._render(name, color=color, **kwargs) # Handle justification of prompt invis_chars = self.invisible_chars[name] if color else 0 self.txtwidth = _lenlastline(res) - invis_chars just = self.justify if (just is None) else just # If the prompt spans more than one line, don't try to justify it: if just and name != 'in' and ('\n' not in res) and ('\r' not in res): res = res.rjust(self.width + invis_chars) self.width = _lenlastline(res) - invis_chars return res