repo_name
stringlengths
5
100
ref
stringlengths
12
67
path
stringlengths
4
244
copies
stringlengths
1
8
content
stringlengths
0
1.05M
ramonsaraiva/sgce
refs/heads/master
sgce/sgce/unislugify.py
2
import re from django.template.defaultfilters import slugify def unique_slugify(instance, value, slug_field_name='slug', queryset=None, slug_separator='-'): """ Calculates and stores a unique slug of ``value`` for an instance. ``slug_field_name`` should be a string matching the name of the field to store the slug in (and the field to check against for uniqueness). ``queryset`` usually doesn't need to be explicitly provided - it'll default to using the ``.all()`` queryset from the model's default manager. """ slug_field = instance._meta.get_field(slug_field_name) slug = getattr(instance, slug_field.attname) slug_len = slug_field.max_length # Sort out the initial slug, limiting its length if necessary. slug = slugify(value) if slug_len: slug = slug[:slug_len] slug = _slug_strip(slug, slug_separator) original_slug = slug # Create the queryset if one wasn't explicitly provided and exclude the # current instance from the queryset. if queryset is None: queryset = instance.__class__._default_manager.all() if instance.pk: queryset = queryset.exclude(pk=instance.pk) # Find a unique slug. If one matches, at '-2' to the end and try again # (then '-3', etc). next = 2 while not slug or queryset.filter(**{slug_field_name: slug}): slug = original_slug end = '%s%s' % (slug_separator, next) if slug_len and len(slug) + len(end) > slug_len: slug = slug[:slug_len-len(end)] slug = _slug_strip(slug, slug_separator) slug = '%s%s' % (slug, end) next += 1 setattr(instance, slug_field.attname, slug) def _slug_strip(value, separator='-'): """ Cleans up a slug by removing slug separator characters that occur at the beginning or end of a slug. If an alternate separator is used, it will also replace any instances of the default '-' separator with the new separator. """ separator = separator or '' if separator == '-' or not separator: re_sep = '-' else: re_sep = '(?:-|%s)' % re.escape(separator) # Remove multiple instances and if an alternate separator is provided, # replace the default '-' separator. if separator != re_sep: value = re.sub('%s+' % re_sep, separator, value) # Remove separator from the beginning and end of the slug. if separator: if separator != '-': re_sep = re.escape(separator) value = re.sub(r'^%s+|%s+$' % (re_sep, re_sep), '', value) return value
ahotam/micropython
refs/heads/master
tests/basics/frozenset_binop.py
76
try: frozenset except NameError: print("SKIP") import sys sys.exit() sets = [ frozenset(), frozenset({1}), frozenset({1, 2}), frozenset({1, 2, 3}), frozenset({2, 3}), frozenset({2, 3, 5}), frozenset({5}), frozenset({7}) ] for s in sets: for t in sets: print(sorted(s), '|', sorted(t), '=', sorted(s | t)) print(sorted(s), '^', sorted(t), '=', sorted(s ^ t)) print(sorted(s), '&', sorted(t), '=', sorted(s & t)) print(sorted(s), '-', sorted(t), '=', sorted(s - t)) u = s.copy() u |= t print(sorted(s), "|=", sorted(t), '-->', sorted(u)) u = s.copy() u ^= t print(sorted(s), "^=", sorted(t), '-->', sorted(u)) u = s.copy() u &= t print(sorted(s), "&=", sorted(t), "-->", sorted(u)) u = s.copy() u -= t print(sorted(s), "-=", sorted(t), "-->", sorted(u)) print(sorted(s), '==', sorted(t), '=', s == t) print(sorted(s), '!=', sorted(t), '=', s != t) print(sorted(s), '>', sorted(t), '=', s > t) print(sorted(s), '>=', sorted(t), '=', s >= t) print(sorted(s), '<', sorted(t), '=', s < t) print(sorted(s), '<=', sorted(t), '=', s <= t)
switchkiller/nodejs-play
refs/heads/master
node_modules/socket.io/node_modules/socket.io-client/node_modules/engine.io-client/node_modules/engine.io-parser/node_modules/utf8/tests/generate-test-data.py
1788
#!/usr/bin/env python import re import json # https://mathiasbynens.be/notes/javascript-encoding#surrogate-formulae # http://stackoverflow.com/a/13436167/96656 def unisymbol(codePoint): if codePoint >= 0x0000 and codePoint <= 0xFFFF: return unichr(codePoint) elif codePoint >= 0x010000 and codePoint <= 0x10FFFF: highSurrogate = int((codePoint - 0x10000) / 0x400) + 0xD800 lowSurrogate = int((codePoint - 0x10000) % 0x400) + 0xDC00 return unichr(highSurrogate) + unichr(lowSurrogate) else: return 'Error' def hexify(codePoint): return 'U+' + hex(codePoint)[2:].upper().zfill(6) def writeFile(filename, contents): print filename with open(filename, 'w') as f: f.write(contents.strip() + '\n') data = [] for codePoint in range(0x000000, 0x10FFFF + 1): # Skip non-scalar values. if codePoint >= 0xD800 and codePoint <= 0xDFFF: continue symbol = unisymbol(codePoint) # http://stackoverflow.com/a/17199950/96656 bytes = symbol.encode('utf8').decode('latin1') data.append({ 'codePoint': codePoint, 'decoded': symbol, 'encoded': bytes }); jsonData = json.dumps(data, sort_keys=False, indent=2, separators=(',', ': ')) # Use tabs instead of double spaces for indentation jsonData = jsonData.replace(' ', '\t') # Escape hexadecimal digits in escape sequences jsonData = re.sub( r'\\u([a-fA-F0-9]{4})', lambda match: r'\u{}'.format(match.group(1).upper()), jsonData ) writeFile('data.json', jsonData)
DoubleNegativeVisualEffects/cortex
refs/heads/master
test/IECore/BGEOParticleReader.py
12
########################################################################## # # Copyright (c) 2009, Image Engine Design Inc. All rights reserved. # # Redistribution and use in source and binary forms, with or without # modification, are permitted provided that the following conditions are # met: # # * Redistributions of source code must retain the above copyright # notice, this list of conditions and the following disclaimer. # # * Redistributions in binary form must reproduce the above copyright # notice, this list of conditions and the following disclaimer in the # documentation and/or other materials provided with the distribution. # # * Neither the name of Image Engine Design nor the names of any # other contributors to this software may be used to endorse or # promote products derived from this software without specific prior # written permission. # # THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS # IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, # THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR # PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT OWNER OR # CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, # EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, # PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR # PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF # LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING # NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS # SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. # ########################################################################## import unittest import sys import IECore class TestBGEOReader( unittest.TestCase ) : def testConstruction( self ) : r = IECore.Reader.create( "test/IECore/data/bgeoFiles/particleTest.0006.bgeo" ) self.assert_( r.isInstanceOf( "ParticleReader" ) ) self.assertEqual( type( r ), IECore.BGEOParticleReader ) self.assertEqual( r["fileName"].getValue().value, "test/IECore/data/bgeoFiles/particleTest.0006.bgeo" ) def testRead( self ) : r = IECore.Reader.create( "test/IECore/data/bgeoFiles/particleTest.0006.bgeo" ) self.assertEqual( type( r ), IECore.BGEOParticleReader ) self.assertEqual( r.numParticles(), 26 ) attrNames = r.attributeNames() expectedAttrNamesAndTypes = { "P" : IECore.V3fVectorData, "v" : IECore.V3fVectorData, "accel" : IECore.V3fVectorData, "life" : IECore.V2fVectorData, "pstate" : IECore.IntVectorData, "id" : IECore.IntVectorData, "parent" : IECore.IntVectorData, "spriteshop" : IECore.StringVectorData, "spriterot" : IECore.FloatVectorData, "spritescale" : IECore.V3fVectorData, } self.assertEqual( len( attrNames ), len( expectedAttrNamesAndTypes ) ) for i in expectedAttrNamesAndTypes.keys() : self.assert_( i in attrNames ) c = r.read() self.assertEqual( type( c ), IECore.PointsPrimitive ) self.assertEqual( len( c ), len( expectedAttrNamesAndTypes ) ) for i in expectedAttrNamesAndTypes.keys() : self.assert_( i in c ) self.assertEqual( type(c[i].data), expectedAttrNamesAndTypes[i] ) self.assertEqual( len(c[i].data), r.numParticles() ) for p in c["P"].data : self.assert_( abs( p.x ) < 0.2 ) self.assert_( abs( p.x ) < 0.2 ) self.assert_( abs( p.z ) < 0.2 ) for s in c["spriteshop"].data : self.assertEqual( s, '/shop/sprite/sprite' ) self.assertEqual( c["id"].data, IECore.IntVectorData( range( 1, 27 ) ) ) def testFiltering( self ) : r = IECore.Reader.create( "test/IECore/data/bgeoFiles/particleTest.0006.bgeo" ) attributesToLoad = [ "P", "life" ] r.parameters()["percentage"].setValue( IECore.FloatData( 50 ) ) r.parameters()["attributes"].setValue( IECore.StringVectorData( attributesToLoad ) ) p = r.read() # what the acceptable thresholds should be are somewhat debatable, # especially for such a small number of particles self.assert_( p.numPoints < 15 ) self.assert_( p.numPoints > 7 ) for attr in attributesToLoad : self.assertEqual( p.numPoints, p[attr].data.size() ) a = r.readAttribute( "P" ) self.assert_( len( a ) < 15 ) self.assert_( len( a ) > 7 ) def testConversion( self ) : r = IECore.Reader.create( "test/IECore/data/bgeoFiles/particleTest.0006.bgeo" ) self.assertEqual( type( r ), IECore.BGEOParticleReader ) r.parameters()["realType"].setValue( "double" ) attrNames = r.attributeNames() expectedAttrNamesAndTypes = { "P" : IECore.V3dVectorData, "v" : IECore.V3dVectorData, "accel" : IECore.V3dVectorData, "life" : IECore.V2dVectorData, "pstate" : IECore.IntVectorData, "id" : IECore.IntVectorData, "parent" : IECore.IntVectorData, "spriteshop" : IECore.StringVectorData, "spriterot" : IECore.DoubleVectorData, "spritescale" : IECore.V3dVectorData, } c = r.read() self.assertEqual( type( c ), IECore.PointsPrimitive ) self.assertEqual( len( c ), len( expectedAttrNamesAndTypes ) ) for i in expectedAttrNamesAndTypes.keys() : self.assert_( i in c ) self.assertEqual( type(c[i].data), expectedAttrNamesAndTypes[i] ) self.assertEqual( len(c[i].data), r.numParticles() ) for p in c["P"].data : self.assert_( abs( p.x ) < 0.2 ) self.assert_( abs( p.y ) < 0.2 ) self.assert_( abs( p.z ) < 0.2 ) def testFileNameChange( self ) : """Now Readers are Ops, the filename can be changed and read() can be called again. So we need to check that that works.""" r = IECore.Reader.create( "test/IECore/data/bgeoFiles/particleTest.0006.bgeo" ) self.assertEqual( type( r ), IECore.BGEOParticleReader ) r.parameters()["realType"].setValue( "float" ) attrNames = r.attributeNames() expectedAttrNamesAndTypes = { "P" : IECore.V3fVectorData, "v" : IECore.V3fVectorData, "accel" : IECore.V3fVectorData, "life" : IECore.V2fVectorData, "pstate" : IECore.IntVectorData, "id" : IECore.IntVectorData, "parent" : IECore.IntVectorData, "spriteshop" : IECore.StringVectorData, "spriterot" : IECore.FloatVectorData, "spritescale" : IECore.V3fVectorData, } c = r.read() self.assertEqual( type( c ), IECore.PointsPrimitive ) self.assertEqual( len( c ), len( expectedAttrNamesAndTypes ) ) for i in expectedAttrNamesAndTypes.keys() : self.assert_( i in c ) self.assertEqual( type(c[i].data), expectedAttrNamesAndTypes[i] ) self.assertEqual( len(c[i].data), r.numParticles() ) for p in c["P"].data : self.assert_( abs( p.x ) < 0.2 ) self.assert_( abs( p.y ) < 0.2 ) self.assert_( abs( p.z ) < 0.2 ) r["fileName"].setValue( IECore.StringData( "test/IECore/data/bgeoFiles/particleTest.0004.bgeo") ) self.assertEqual( r.numParticles(), 17 ) c = r.read() self.assertEqual( type( c ), IECore.PointsPrimitive ) self.assertEqual( len( c ), len( expectedAttrNamesAndTypes ) ) for i in expectedAttrNamesAndTypes.keys() : self.assert_( i in c ) self.assertEqual( type(c[i].data), expectedAttrNamesAndTypes[i] ) self.assertEqual( len(c[i].data), 17 ) for p in c["P"].data : self.assert_( abs( p.x ) < 0.12 ) self.assert_( abs( p.y ) < 0.12 ) self.assert_( abs( p.z ) < 0.12 ) def testParameterTypes( self ) : p = IECore.BGEOParticleReader() self.assert_( p.resultParameter().isInstanceOf( "ObjectParameter" ) ) self.assertEqual( p.resultParameter().validTypes(), [IECore.TypeId.PointsPrimitive] ) if __name__ == "__main__": unittest.main()
40223222/2015cd_midterm
refs/heads/master
static/Brython3.1.1-20150328-091302/Lib/xml/sax/_exceptions.py
625
"""Different kinds of SAX Exceptions""" #in brython the 4 lines below causes an $globals['Exception'] error #import sys #if sys.platform[:4] == "java": # from java.lang import Exception #del sys # ===== SAXEXCEPTION ===== class SAXException(Exception): """Encapsulate an XML error or warning. This class can contain basic error or warning information from either the XML parser or the application: you can subclass it to provide additional functionality, or to add localization. Note that although you will receive a SAXException as the argument to the handlers in the ErrorHandler interface, you are not actually required to raise the exception; instead, you can simply read the information in it.""" def __init__(self, msg, exception=None): """Creates an exception. The message is required, but the exception is optional.""" self._msg = msg self._exception = exception Exception.__init__(self, msg) def getMessage(self): "Return a message for this exception." return self._msg def getException(self): "Return the embedded exception, or None if there was none." return self._exception def __str__(self): "Create a string representation of the exception." return self._msg def __getitem__(self, ix): """Avoids weird error messages if someone does exception[ix] by mistake, since Exception has __getitem__ defined.""" raise AttributeError("__getitem__") # ===== SAXPARSEEXCEPTION ===== class SAXParseException(SAXException): """Encapsulate an XML parse error or warning. This exception will include information for locating the error in the original XML document. Note that although the application will receive a SAXParseException as the argument to the handlers in the ErrorHandler interface, the application is not actually required to raise the exception; instead, it can simply read the information in it and take a different action. Since this exception is a subclass of SAXException, it inherits the ability to wrap another exception.""" def __init__(self, msg, exception, locator): "Creates the exception. The exception parameter is allowed to be None." SAXException.__init__(self, msg, exception) self._locator = locator # We need to cache this stuff at construction time. # If this exception is raised, the objects through which we must # traverse to get this information may be deleted by the time # it gets caught. self._systemId = self._locator.getSystemId() self._colnum = self._locator.getColumnNumber() self._linenum = self._locator.getLineNumber() def getColumnNumber(self): """The column number of the end of the text where the exception occurred.""" return self._colnum def getLineNumber(self): "The line number of the end of the text where the exception occurred." return self._linenum def getPublicId(self): "Get the public identifier of the entity where the exception occurred." return self._locator.getPublicId() def getSystemId(self): "Get the system identifier of the entity where the exception occurred." return self._systemId def __str__(self): "Create a string representation of the exception." sysid = self.getSystemId() if sysid is None: sysid = "<unknown>" linenum = self.getLineNumber() if linenum is None: linenum = "?" colnum = self.getColumnNumber() if colnum is None: colnum = "?" return "%s:%s:%s: %s" % (sysid, linenum, colnum, self._msg) # ===== SAXNOTRECOGNIZEDEXCEPTION ===== class SAXNotRecognizedException(SAXException): """Exception class for an unrecognized identifier. An XMLReader will raise this exception when it is confronted with an unrecognized feature or property. SAX applications and extensions may use this class for similar purposes.""" pass # ===== SAXNOTSUPPORTEDEXCEPTION ===== class SAXNotSupportedException(SAXException): """Exception class for an unsupported operation. An XMLReader will raise this exception when a service it cannot perform is requested (specifically setting a state or value). SAX applications and extensions may use this class for similar purposes.""" pass # ===== SAXNOTSUPPORTEDEXCEPTION ===== class SAXReaderNotAvailable(SAXNotSupportedException): """Exception class for a missing driver. An XMLReader module (driver) should raise this exception when it is first imported, e.g. when a support module cannot be imported. It also may be raised during parsing, e.g. if executing an external program is not permitted.""" pass
resba/gnuradio
refs/heads/master
grc/python/Param.py
6
""" Copyright 2008-2011 Free Software Foundation, Inc. This file is part of GNU Radio GNU Radio Companion is free software; you can redistribute it and/or modify it under the terms of the GNU General Public License as published by the Free Software Foundation; either version 2 of the License, or (at your option) any later version. GNU Radio Companion is distributed in the hope that it will be useful, but WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for more details. You should have received a copy of the GNU General Public License along with this program; if not, write to the Free Software Foundation, Inc., 51 Franklin Street, Fifth Floor, Boston, MA 02110-1301, USA """ from .. base.Param import Param as _Param from .. gui.Param import Param as _GUIParam from .. gui.Param import EntryParam import Constants import numpy import os import pygtk pygtk.require('2.0') import gtk from gnuradio import eng_notation import re from gnuradio import gr _check_id_matcher = re.compile('^[a-z|A-Z]\w*$') _show_id_matcher = re.compile('^(variable\w*|parameter|options|notebook)$') class FileParam(EntryParam): """Provide an entry box for filename and a button to browse for a file.""" def __init__(self, *args, **kwargs): EntryParam.__init__(self, *args, **kwargs) input = gtk.Button('...') input.connect('clicked', self._handle_clicked) self.pack_start(input, False) def _handle_clicked(self, widget=None): """ If the button was clicked, open a file dialog in open/save format. Replace the text in the entry with the new filename from the file dialog. """ #get the paths file_path = self.param.is_valid() and self.param.get_evaluated() or '' (dirname, basename) = os.path.isfile(file_path) and os.path.split(file_path) or (file_path, '') if not os.path.exists(dirname): dirname = os.getcwd() #fix bad paths #build the dialog if self.param.get_type() == 'file_open': file_dialog = gtk.FileChooserDialog('Open a Data File...', None, gtk.FILE_CHOOSER_ACTION_OPEN, ('gtk-cancel',gtk.RESPONSE_CANCEL,'gtk-open',gtk.RESPONSE_OK)) elif self.param.get_type() == 'file_save': file_dialog = gtk.FileChooserDialog('Save a Data File...', None, gtk.FILE_CHOOSER_ACTION_SAVE, ('gtk-cancel',gtk.RESPONSE_CANCEL, 'gtk-save',gtk.RESPONSE_OK)) file_dialog.set_do_overwrite_confirmation(True) file_dialog.set_current_name(basename) #show the current filename file_dialog.set_current_folder(dirname) #current directory file_dialog.set_select_multiple(False) file_dialog.set_local_only(True) if gtk.RESPONSE_OK == file_dialog.run(): #run the dialog file_path = file_dialog.get_filename() #get the file path self._input.set_text(file_path) self._handle_changed() file_dialog.destroy() #destroy the dialog #blacklist certain ids, its not complete, but should help import __builtin__ ID_BLACKLIST = ['self', 'options', 'gr', 'blks2', 'wxgui', 'wx', 'math', 'forms', 'firdes'] + \ filter(lambda x: not x.startswith('_'), dir(gr.top_block())) + dir(__builtin__) #define types, native python + numpy VECTOR_TYPES = (tuple, list, set, numpy.ndarray) COMPLEX_TYPES = [complex, numpy.complex, numpy.complex64, numpy.complex128] REAL_TYPES = [float, numpy.float, numpy.float32, numpy.float64] INT_TYPES = [int, long, numpy.int, numpy.int8, numpy.int16, numpy.int32, numpy.uint64, numpy.uint, numpy.uint8, numpy.uint16, numpy.uint32, numpy.uint64] #cast to tuple for isinstance, concat subtypes COMPLEX_TYPES = tuple(COMPLEX_TYPES + REAL_TYPES + INT_TYPES) REAL_TYPES = tuple(REAL_TYPES + INT_TYPES) INT_TYPES = tuple(INT_TYPES) class Param(_Param, _GUIParam): def __init__(self, **kwargs): _Param.__init__(self, **kwargs) _GUIParam.__init__(self) self._init = False self._hostage_cells = list() def get_types(self): return ( 'raw', 'enum', 'complex', 'real', 'int', 'complex_vector', 'real_vector', 'int_vector', 'hex', 'string', 'bool', 'file_open', 'file_save', 'id', 'stream_id', 'grid_pos', 'notebook', 'gui_hint', 'import', ) def __repr__(self): """ Get the repr (nice string format) for this param. @return the string representation """ ################################################## # truncate helper method ################################################## def _truncate(string, style=0): max_len = max(27 - len(self.get_name()), 3) if len(string) > max_len: if style < 0: #front truncate string = '...' + string[3-max_len:] elif style == 0: #center truncate string = string[:max_len/2 -3] + '...' + string[-max_len/2:] elif style > 0: #rear truncate string = string[:max_len-3] + '...' return string ################################################## # simple conditions ################################################## if not self.is_valid(): return _truncate(self.get_value()) if self.get_value() in self.get_option_keys(): return self.get_option(self.get_value()).get_name() ################################################## # display logic for numbers ################################################## def num_to_str(num): if isinstance(num, COMPLEX_TYPES): num = complex(num) #cast to python complex if num == 0: return '0' #value is zero elif num.imag == 0: return '%s'%eng_notation.num_to_str(num.real) #value is real elif num.real == 0: return '%sj'%eng_notation.num_to_str(num.imag) #value is imaginary elif num.imag < 0: return '%s-%sj'%(eng_notation.num_to_str(num.real), eng_notation.num_to_str(abs(num.imag))) else: return '%s+%sj'%(eng_notation.num_to_str(num.real), eng_notation.num_to_str(num.imag)) else: return str(num) ################################################## # split up formatting by type ################################################## truncate = 0 #default center truncate e = self.get_evaluated() t = self.get_type() if isinstance(e, bool): return str(e) elif isinstance(e, COMPLEX_TYPES): dt_str = num_to_str(e) elif isinstance(e, VECTOR_TYPES): #vector types if len(e) > 8: dt_str = self.get_value() #large vectors use code truncate = 1 else: dt_str = ', '.join(map(num_to_str, e)) #small vectors use eval elif t in ('file_open', 'file_save'): dt_str = self.get_value() truncate = -1 else: dt_str = str(e) #other types ################################################## # done ################################################## return _truncate(dt_str, truncate) def get_input(self, *args, **kwargs): if self.get_type() in ('file_open', 'file_save'): return FileParam(self, *args, **kwargs) return _GUIParam.get_input(self, *args, **kwargs) def get_color(self): """ Get the color that represents this param's type. @return a hex color code. """ try: return { #number types 'complex': Constants.COMPLEX_COLOR_SPEC, 'real': Constants.FLOAT_COLOR_SPEC, 'int': Constants.INT_COLOR_SPEC, #vector types 'complex_vector': Constants.COMPLEX_VECTOR_COLOR_SPEC, 'real_vector': Constants.FLOAT_VECTOR_COLOR_SPEC, 'int_vector': Constants.INT_VECTOR_COLOR_SPEC, #special 'bool': Constants.INT_COLOR_SPEC, 'hex': Constants.INT_COLOR_SPEC, 'string': Constants.BYTE_VECTOR_COLOR_SPEC, 'id': Constants.ID_COLOR_SPEC, 'stream_id': Constants.ID_COLOR_SPEC, 'grid_pos': Constants.INT_VECTOR_COLOR_SPEC, 'notebook': Constants.INT_VECTOR_COLOR_SPEC, 'raw': Constants.WILDCARD_COLOR_SPEC, }[self.get_type()] except: return _Param.get_color(self) def get_hide(self): """ Get the hide value from the base class. Hide the ID parameter for most blocks. Exceptions below. If the parameter controls a port type, vlen, or nports, return part. If the parameter is an empty grid position, return part. These parameters are redundant to display in the flow graph view. @return hide the hide property string """ hide = _Param.get_hide(self) if hide: return hide #hide ID in non variable blocks if self.get_key() == 'id' and not _show_id_matcher.match(self.get_parent().get_key()): return 'part' #hide port controllers for type and nports if self.get_key() in ' '.join(map( lambda p: ' '.join([p._type, p._nports]), self.get_parent().get_ports()) ): return 'part' #hide port controllers for vlen, when == 1 if self.get_key() in ' '.join(map( lambda p: p._vlen, self.get_parent().get_ports()) ): try: if int(self.get_evaluated()) == 1: return 'part' except: pass #hide empty grid positions if self.get_key() in ('grid_pos', 'notebook') and not self.get_value(): return 'part' return hide def validate(self): """ Validate the param. A test evaluation is performed """ _Param.validate(self) #checks type self._evaluated = None try: self._evaluated = self.evaluate() except Exception, e: self.add_error_message(str(e)) def get_evaluated(self): return self._evaluated def evaluate(self): """ Evaluate the value. @return evaluated type """ self._init = True self._lisitify_flag = False self._stringify_flag = False self._hostage_cells = list() def eval_string(v): try: e = self.get_parent().get_parent().evaluate(v) if isinstance(e, str): return e raise Exception #want to stringify except: self._stringify_flag = True return v t = self.get_type() v = self.get_value() ######################### # Enum Type ######################### if self.is_enum(): return v ######################### # Numeric Types ######################### elif t in ('raw', 'complex', 'real', 'int', 'hex', 'bool'): #raise exception if python cannot evaluate this value try: e = self.get_parent().get_parent().evaluate(v) except Exception, e: raise Exception, 'Value "%s" cannot be evaluated:\n%s'%(v, e) #raise an exception if the data is invalid if t == 'raw': return e elif t == 'complex': if not isinstance(e, COMPLEX_TYPES): raise Exception, 'Expression "%s" is invalid for type complex.'%str(e) return e elif t == 'real': if not isinstance(e, REAL_TYPES): raise Exception, 'Expression "%s" is invalid for type real.'%str(e) return e elif t == 'int': if not isinstance(e, INT_TYPES): raise Exception, 'Expression "%s" is invalid for type integer.'%str(e) return e elif t == 'hex': return hex(e) elif t == 'bool': if not isinstance(e, bool): raise Exception, 'Expression "%s" is invalid for type bool.'%str(e) return e else: raise TypeError, 'Type "%s" not handled'%t ######################### # Numeric Vector Types ######################### elif t in ('complex_vector', 'real_vector', 'int_vector'): if not v: v = '()' #turn a blank string into an empty list, so it will eval #raise exception if python cannot evaluate this value try: e = self.get_parent().get_parent().evaluate(v) except Exception, e: raise Exception, 'Value "%s" cannot be evaluated:\n%s'%(v, e) #raise an exception if the data is invalid if t == 'complex_vector': if not isinstance(e, VECTOR_TYPES): self._lisitify_flag = True e = [e] if not all([isinstance(ei, COMPLEX_TYPES) for ei in e]): raise Exception, 'Expression "%s" is invalid for type complex vector.'%str(e) return e elif t == 'real_vector': if not isinstance(e, VECTOR_TYPES): self._lisitify_flag = True e = [e] if not all([isinstance(ei, REAL_TYPES) for ei in e]): raise Exception, 'Expression "%s" is invalid for type real vector.'%str(e) return e elif t == 'int_vector': if not isinstance(e, VECTOR_TYPES): self._lisitify_flag = True e = [e] if not all([isinstance(ei, INT_TYPES) for ei in e]): raise Exception, 'Expression "%s" is invalid for type integer vector.'%str(e) return e ######################### # String Types ######################### elif t in ('string', 'file_open', 'file_save'): #do not check if file/directory exists, that is a runtime issue e = eval_string(v) return str(e) ######################### # Unique ID Type ######################### elif t == 'id': #can python use this as a variable? if not _check_id_matcher.match(v): raise Exception, 'ID "%s" must begin with a letter and may contain letters, numbers, and underscores.'%v ids = [param.get_value() for param in self.get_all_params(t)] if ids.count(v) > 1: #id should only appear once, or zero times if block is disabled raise Exception, 'ID "%s" is not unique.'%v if v in ID_BLACKLIST: raise Exception, 'ID "%s" is blacklisted.'%v return v ######################### # Stream ID Type ######################### elif t == 'stream_id': #get a list of all stream ids used in the virtual sinks ids = [param.get_value() for param in filter( lambda p: p.get_parent().is_virtual_sink(), self.get_all_params(t), )] #check that the virtual sink's stream id is unique if self.get_parent().is_virtual_sink(): if ids.count(v) > 1: #id should only appear once, or zero times if block is disabled raise Exception, 'Stream ID "%s" is not unique.'%v #check that the virtual source's steam id is found if self.get_parent().is_virtual_source(): if v not in ids: raise Exception, 'Stream ID "%s" is not found.'%v return v ######################### # GUI Position/Hint ######################### elif t == 'gui_hint': if ':' in v: tab, pos = v.split(':') elif '@' in v: tab, pos = v, '' else: tab, pos = '', v if '@' in tab: tab, index = tab.split('@') else: index = '?' widget_str = ({ (True, True): 'self.%(tab)s_grid_layout_%(index)s.addWidget(%(widget)s, %(pos)s)', (True, False): 'self.%(tab)s_layout_%(index)s.addWidget(%(widget)s)', (False, True): 'self.top_grid_layout.addWidget(%(widget)s, %(pos)s)', (False, False): 'self.top_layout.addWidget(%(widget)s)', }[bool(tab), bool(pos)])%{'tab': tab, 'index': index, 'widget': '%s', 'pos': pos} def gui_hint(ws, w): if 'layout' in w: ws = ws.replace('addWidget', 'addLayout') return ws%w return lambda w: gui_hint(widget_str, w) ######################### # Grid Position Type ######################### elif t == 'grid_pos': if not v: return '' #allow for empty grid pos e = self.get_parent().get_parent().evaluate(v) if not isinstance(e, (list, tuple)) or len(e) != 4 or not all([isinstance(ei, int) for ei in e]): raise Exception, 'A grid position must be a list of 4 integers.' row, col, row_span, col_span = e #check row, col if row < 0 or col < 0: raise Exception, 'Row and column must be non-negative.' #check row span, col span if row_span <= 0 or col_span <= 0: raise Exception, 'Row and column span must be greater than zero.' #get hostage cell parent try: my_parent = self.get_parent().get_param('notebook').evaluate() except: my_parent = '' #calculate hostage cells for r in range(row_span): for c in range(col_span): self._hostage_cells.append((my_parent, (row+r, col+c))) #avoid collisions params = filter(lambda p: p is not self, self.get_all_params('grid_pos')) for param in params: for parent, cell in param._hostage_cells: if (parent, cell) in self._hostage_cells: raise Exception, 'Another graphical element is using parent "%s", cell "%s".'%(str(parent), str(cell)) return e ######################### # Notebook Page Type ######################### elif t == 'notebook': if not v: return '' #allow for empty notebook #get a list of all notebooks notebook_blocks = filter(lambda b: b.get_key() == 'notebook', self.get_parent().get_parent().get_enabled_blocks()) #check for notebook param syntax try: notebook_id, page_index = map(str.strip, v.split(',')) except: raise Exception, 'Bad notebook page format.' #check that the notebook id is valid try: notebook_block = filter(lambda b: b.get_id() == notebook_id, notebook_blocks)[0] except: raise Exception, 'Notebook id "%s" is not an existing notebook id.'%notebook_id #check that page index exists if int(page_index) not in range(len(notebook_block.get_param('labels').evaluate())): raise Exception, 'Page index "%s" is not a valid index number.'%page_index return notebook_id, page_index ######################### # Import Type ######################### elif t == 'import': n = dict() #new namespace try: exec v in n except ImportError: raise Exception, 'Import "%s" failed.'%v except Exception: raise Exception, 'Bad import syntax: "%s".'%v return filter(lambda k: str(k) != '__builtins__', n.keys()) ######################### else: raise TypeError, 'Type "%s" not handled'%t def to_code(self): """ Convert the value to code. For string and list types, check the init flag, call evaluate(). This ensures that evaluate() was called to set the xxxify_flags. @return a string representing the code """ v = self.get_value() t = self.get_type() if t in ('string', 'file_open', 'file_save'): #string types if not self._init: self.evaluate() if self._stringify_flag: return '"%s"'%v.replace('"', '\"') else: return v elif t in ('complex_vector', 'real_vector', 'int_vector'): #vector types if not self._init: self.evaluate() if self._lisitify_flag: return '(%s, )'%v else: return '(%s)'%v else: return v def get_all_params(self, type): """ Get all the params from the flowgraph that have the given type. @param type the specified type @return a list of params """ return sum([filter(lambda p: p.get_type() == type, block.get_params()) for block in self.get_parent().get_parent().get_enabled_blocks()], [])
Agana/MyBlogAgain
refs/heads/master
django/contrib/gis/geos/point.py
403
from ctypes import c_uint from django.contrib.gis.geos.error import GEOSException from django.contrib.gis.geos.geometry import GEOSGeometry from django.contrib.gis.geos import prototypes as capi class Point(GEOSGeometry): _minlength = 2 _maxlength = 3 def __init__(self, x, y=None, z=None, srid=None): """ The Point object may be initialized with either a tuple, or individual parameters. For Example: >>> p = Point((5, 23)) # 2D point, passed in as a tuple >>> p = Point(5, 23, 8) # 3D point, passed in with individual parameters """ if isinstance(x, (tuple, list)): # Here a tuple or list was passed in under the `x` parameter. ndim = len(x) coords = x elif isinstance(x, (int, float, long)) and isinstance(y, (int, float, long)): # Here X, Y, and (optionally) Z were passed in individually, as parameters. if isinstance(z, (int, float, long)): ndim = 3 coords = [x, y, z] else: ndim = 2 coords = [x, y] else: raise TypeError('Invalid parameters given for Point initialization.') point = self._create_point(ndim, coords) # Initializing using the address returned from the GEOS # createPoint factory. super(Point, self).__init__(point, srid=srid) def _create_point(self, ndim, coords): """ Create a coordinate sequence, set X, Y, [Z], and create point """ if ndim < 2 or ndim > 3: raise TypeError('Invalid point dimension: %s' % str(ndim)) cs = capi.create_cs(c_uint(1), c_uint(ndim)) i = iter(coords) capi.cs_setx(cs, 0, i.next()) capi.cs_sety(cs, 0, i.next()) if ndim == 3: capi.cs_setz(cs, 0, i.next()) return capi.create_point(cs) def _set_list(self, length, items): ptr = self._create_point(length, items) if ptr: capi.destroy_geom(self.ptr) self._ptr = ptr self._set_cs() else: # can this happen? raise GEOSException('Geometry resulting from slice deletion was invalid.') def _set_single(self, index, value): self._cs.setOrdinate(index, 0, value) def __iter__(self): "Allows iteration over coordinates of this Point." for i in xrange(len(self)): yield self[i] def __len__(self): "Returns the number of dimensions for this Point (either 0, 2 or 3)." if self.empty: return 0 if self.hasz: return 3 else: return 2 def _get_single_external(self, index): if index == 0: return self.x elif index == 1: return self.y elif index == 2: return self.z _get_single_internal = _get_single_external def get_x(self): "Returns the X component of the Point." return self._cs.getOrdinate(0, 0) def set_x(self, value): "Sets the X component of the Point." self._cs.setOrdinate(0, 0, value) def get_y(self): "Returns the Y component of the Point." return self._cs.getOrdinate(1, 0) def set_y(self, value): "Sets the Y component of the Point." self._cs.setOrdinate(1, 0, value) def get_z(self): "Returns the Z component of the Point." if self.hasz: return self._cs.getOrdinate(2, 0) else: return None def set_z(self, value): "Sets the Z component of the Point." if self.hasz: self._cs.setOrdinate(2, 0, value) else: raise GEOSException('Cannot set Z on 2D Point.') # X, Y, Z properties x = property(get_x, set_x) y = property(get_y, set_y) z = property(get_z, set_z) ### Tuple setting and retrieval routines. ### def get_coords(self): "Returns a tuple of the point." return self._cs.tuple def set_coords(self, tup): "Sets the coordinates of the point with the given tuple." self._cs[0] = tup # The tuple and coords properties tuple = property(get_coords, set_coords) coords = tuple
Rosuav/shed
refs/heads/master
channel_split.py
1
# Split the audio channel from a file into its channels # Keeps video and subtitles tracks untouched import json import subprocess import sys if len(sys.argv) < 3: sys.exit(1, "USAGE: python3 %s inputfile outputfile") _, infile, outfile, *_ = sys.argv # Determine the channel layout of the input file p = subprocess.run(["ffprobe", "-print_format", "json", "-show_streams", "-v", "quiet", infile], check=True, capture_output=True) info = json.loads(p.stdout) if not isinstance(info, dict): raise Exception("shouldn't happen - non-dict output from ffprobe") map_video = map_subs = [] audio = None # Scan streams. If we have any video streams, make sure we grab them, ditto subs. for strm in info["streams"]: if strm["codec_type"] == "video": map_video = ["-map", "0:v"] # If there are multiple, these still only need one map argument elif strm["codec_type"] == "subtitle": map_subs = ["-map", "0:s"] elif audio is None: audio = strm if audio is None: sys.exit(1, "No audio stream found") # Good. There's an audio stream. (We'll use the first if there are multiple.) # Ask ffmpeg what channel layouts there are and how the channels should be described. p = subprocess.run(["ffmpeg", "-layouts"], check=True, capture_output=True) channel_desc, layouts = {}, {} cur = None for line in p.stdout.decode().split("\n"): if line == "": pass elif line == "Individual channels:": cur = channel_desc elif line == "Standard channel layouts:": cur = layouts else: k, v = line.split(maxsplit=1) cur[k] = v # Give each channel a descriptive title eg "[FR] front right". audio_titles = [] for i, channel in enumerate(layouts[audio["channel_layout"]].split("+")): audio_titles.extend(["-metadata:s:a:%d" % i, f"title=[{channel}] {channel_desc[channel]}"]) # Let's do this. args = (["ffmpeg", "-y", "-i", infile, "-filter_complex", "channelsplit=channel_layout=" + audio["channel_layout"]] + map_video + map_subs + audio_titles + ["-c:v", "copy", "-c:s", "copy", "-loglevel", "error", "-stats", outfile]) # print(args) sys.exit(subprocess.run(args).returncode)
j2sol/ansible-modules-core
refs/heads/devel
database/postgresql/postgresql_user.py
25
#!/usr/bin/python # -*- coding: utf-8 -*- # This file is part of Ansible # # Ansible is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # Ansible is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with Ansible. If not, see <http://www.gnu.org/licenses/>. DOCUMENTATION = ''' --- module: postgresql_user short_description: Adds or removes a users (roles) from a PostgreSQL database. description: - Add or remove PostgreSQL users (roles) from a remote host and, optionally, grant the users access to an existing database or tables. - The fundamental function of the module is to create, or delete, roles from a PostgreSQL cluster. Privilege assignment, or removal, is an optional step, which works on one database at a time. This allows for the module to be called several times in the same module to modify the permissions on different databases, or to grant permissions to already existing users. - A user cannot be removed until all the privileges have been stripped from the user. In such situation, if the module tries to remove the user it will fail. To avoid this from happening the fail_on_user option signals the module to try to remove the user, but if not possible keep going; the module will report if changes happened and separately if the user was removed or not. version_added: "0.6" options: name: description: - name of the user (role) to add or remove required: true default: null password: description: - set the user's password, before 1.4 this was required. - "When passing an encrypted password, the encrypted parameter must also be true, and it must be generated with the format C('str[\\"md5\\"] + md5[ password + username ]'), resulting in a total of 35 characters. An easy way to do this is: C(echo \\"md5`echo -n \\"verysecretpasswordJOE\\" | md5`\\")." required: false default: null db: description: - name of database where permissions will be granted required: false default: null fail_on_user: description: - if C(yes), fail when user can't be removed. Otherwise just log and continue required: false default: 'yes' choices: [ "yes", "no" ] port: description: - Database port to connect to. required: false default: 5432 login_user: description: - User (role) used to authenticate with PostgreSQL required: false default: postgres login_password: description: - Password used to authenticate with PostgreSQL required: false default: null login_host: description: - Host running PostgreSQL. required: false default: localhost login_unix_socket: description: - Path to a Unix domain socket for local connections required: false default: null priv: description: - "PostgreSQL privileges string in the format: C(table:priv1,priv2)" required: false default: null role_attr_flags: description: - "PostgreSQL role attributes string in the format: CREATEDB,CREATEROLE,SUPERUSER" required: false default: null choices: [ "[NO]SUPERUSER","[NO]CREATEROLE", "[NO]CREATEUSER", "[NO]CREATEDB", "[NO]INHERIT", "[NO]LOGIN", "[NO]REPLICATION" ] state: description: - The user (role) state required: false default: present choices: [ "present", "absent" ] encrypted: description: - denotes if the password is already encrypted. boolean. required: false default: false version_added: '1.4' expires: description: - sets the user's password expiration. required: false default: null version_added: '1.4' notes: - The default authentication assumes that you are either logging in as or sudo'ing to the postgres account on the host. - This module uses psycopg2, a Python PostgreSQL database adapter. You must ensure that psycopg2 is installed on the host before using this module. If the remote host is the PostgreSQL server (which is the default case), then PostgreSQL must also be installed on the remote host. For Ubuntu-based systems, install the postgresql, libpq-dev, and python-psycopg2 packages on the remote host before using this module. - If you specify PUBLIC as the user, then the privilege changes will apply to all users. You may not specify password or role_attr_flags when the PUBLIC user is specified. requirements: [ psycopg2 ] author: Lorin Hochstein ''' EXAMPLES = ''' # Create django user and grant access to database and products table - postgresql_user: db=acme name=django password=ceec4eif7ya priv=CONNECT/products:ALL # Create rails user, grant privilege to create other databases and demote rails from super user status - postgresql_user: name=rails password=secret role_attr_flags=CREATEDB,NOSUPERUSER # Remove test user privileges from acme - postgresql_user: db=acme name=test priv=ALL/products:ALL state=absent fail_on_user=no # Remove test user from test database and the cluster - postgresql_user: db=test name=test priv=ALL state=absent # Example privileges string format INSERT,UPDATE/table:SELECT/anothertable:ALL # Remove an existing user's password - postgresql_user: db=test user=test password=NULL ''' import re import itertools try: import psycopg2 except ImportError: postgresqldb_found = False else: postgresqldb_found = True _flags = ('SUPERUSER', 'CREATEROLE', 'CREATEUSER', 'CREATEDB', 'INHERIT', 'LOGIN', 'REPLICATION') VALID_FLAGS = frozenset(itertools.chain(_flags, ('NO%s' % f for f in _flags))) VALID_PRIVS = dict(table=frozenset(('SELECT', 'INSERT', 'UPDATE', 'DELETE', 'TRUNCATE', 'REFERENCES', 'TRIGGER', 'ALL', 'USAGE')), database=frozenset(('CREATE', 'CONNECT', 'TEMPORARY', 'TEMP', 'ALL', 'USAGE')), ) class InvalidFlagsError(Exception): pass class InvalidPrivsError(Exception): pass # =========================================== # PostgreSQL module specific support methods. # def user_exists(cursor, user): # The PUBLIC user is a special case that is always there if user == 'PUBLIC': return True query = "SELECT rolname FROM pg_roles WHERE rolname=%(user)s" cursor.execute(query, {'user': user}) return cursor.rowcount > 0 def user_add(cursor, user, password, role_attr_flags, encrypted, expires): """Create a new database user (role).""" # Note: role_attr_flags escaped by parse_role_attrs and encrypted is a literal query_password_data = dict(password=password, expires=expires) query = ['CREATE USER %(user)s' % { "user": pg_quote_identifier(user, 'role')}] if password is not None: query.append("WITH %(crypt)s" % { "crypt": encrypted }) query.append("PASSWORD %(password)s") if expires is not None: query.append("VALID UNTIL %(expires)s") query.append(role_attr_flags) query = ' '.join(query) cursor.execute(query, query_password_data) return True def user_alter(cursor, module, user, password, role_attr_flags, encrypted, expires): """Change user password and/or attributes. Return True if changed, False otherwise.""" changed = False # Note: role_attr_flags escaped by parse_role_attrs and encrypted is a literal if user == 'PUBLIC': if password is not None: module.fail_json(msg="cannot change the password for PUBLIC user") elif role_attr_flags != '': module.fail_json(msg="cannot change the role_attr_flags for PUBLIC user") else: return False # Handle passwords. if password is not None or role_attr_flags is not None: # Select password and all flag-like columns in order to verify changes. query_password_data = dict(password=password, expires=expires) select = "SELECT * FROM pg_authid where rolname=%(user)s" cursor.execute(select, {"user": user}) # Grab current role attributes. current_role_attrs = cursor.fetchone() alter = ['ALTER USER %(user)s' % {"user": pg_quote_identifier(user, 'role')}] if password is not None: alter.append("WITH %(crypt)s" % {"crypt": encrypted}) alter.append("PASSWORD %(password)s") alter.append(role_attr_flags) elif role_attr_flags: alter.append('WITH %s' % role_attr_flags) if expires is not None: alter.append("VALID UNTIL %(expires)s") try: cursor.execute(' '.join(alter), query_password_data) except psycopg2.InternalError, e: if e.pgcode == '25006': # Handle errors due to read-only transactions indicated by pgcode 25006 # ERROR: cannot execute ALTER ROLE in a read-only transaction changed = False module.fail_json(msg=e.pgerror) return changed else: raise psycopg2.InternalError, e # Grab new role attributes. cursor.execute(select, {"user": user}) new_role_attrs = cursor.fetchone() # Detect any differences between current_ and new_role_attrs. for i in range(len(current_role_attrs)): if current_role_attrs[i] != new_role_attrs[i]: changed = True return changed def user_delete(cursor, user): """Try to remove a user. Returns True if successful otherwise False""" cursor.execute("SAVEPOINT ansible_pgsql_user_delete") try: cursor.execute("DROP USER %s" % pg_quote_identifier(user, 'role')) except: cursor.execute("ROLLBACK TO SAVEPOINT ansible_pgsql_user_delete") cursor.execute("RELEASE SAVEPOINT ansible_pgsql_user_delete") return False cursor.execute("RELEASE SAVEPOINT ansible_pgsql_user_delete") return True def has_table_privilege(cursor, user, table, priv): query = 'SELECT has_table_privilege(%s, %s, %s)' cursor.execute(query, (user, table, priv)) return cursor.fetchone()[0] def get_table_privileges(cursor, user, table): if '.' in table: schema, table = table.split('.', 1) else: schema = 'public' query = '''SELECT privilege_type FROM information_schema.role_table_grants WHERE grantee=%s AND table_name=%s AND table_schema=%s''' cursor.execute(query, (user, table, schema)) return set([x[0] for x in cursor.fetchall()]) def grant_table_privilege(cursor, user, table, priv): # Note: priv escaped by parse_privs prev_priv = get_table_privileges(cursor, user, table) query = 'GRANT %s ON TABLE %s TO %s' % ( priv, pg_quote_identifier(table, 'table'), pg_quote_identifier(user, 'role') ) cursor.execute(query) curr_priv = get_table_privileges(cursor, user, table) return len(curr_priv) > len(prev_priv) def revoke_table_privilege(cursor, user, table, priv): # Note: priv escaped by parse_privs prev_priv = get_table_privileges(cursor, user, table) query = 'REVOKE %s ON TABLE %s FROM %s' % ( priv, pg_quote_identifier(table, 'table'), pg_quote_identifier(user, 'role') ) cursor.execute(query) curr_priv = get_table_privileges(cursor, user, table) return len(curr_priv) < len(prev_priv) def get_database_privileges(cursor, user, db): priv_map = { 'C':'CREATE', 'T':'TEMPORARY', 'c':'CONNECT', } query = 'SELECT datacl FROM pg_database WHERE datname = %s' cursor.execute(query, (db,)) datacl = cursor.fetchone()[0] if datacl is None: return [] r = re.search('%s=(C?T?c?)/[a-z]+\,?' % user, datacl) if r is None: return [] o = [] for v in r.group(1): o.append(priv_map[v]) return o def has_database_privilege(cursor, user, db, priv): query = 'SELECT has_database_privilege(%s, %s, %s)' cursor.execute(query, (user, db, priv)) return cursor.fetchone()[0] def grant_database_privilege(cursor, user, db, priv): # Note: priv escaped by parse_privs prev_priv = get_database_privileges(cursor, user, db) if user == "PUBLIC": query = 'GRANT %s ON DATABASE %s TO PUBLIC' % ( priv, pg_quote_identifier(db, 'database')) else: query = 'GRANT %s ON DATABASE %s TO %s' % ( priv, pg_quote_identifier(db, 'database'), pg_quote_identifier(user, 'role')) cursor.execute(query) curr_priv = get_database_privileges(cursor, user, db) return len(curr_priv) > len(prev_priv) def revoke_database_privilege(cursor, user, db, priv): # Note: priv escaped by parse_privs prev_priv = get_database_privileges(cursor, user, db) if user == "PUBLIC": query = 'REVOKE %s ON DATABASE %s FROM PUBLIC' % ( priv, pg_quote_identifier(db, 'database')) else: query = 'REVOKE %s ON DATABASE %s FROM %s' % ( priv, pg_quote_identifier(db, 'database'), pg_quote_identifier(user, 'role')) cursor.execute(query) curr_priv = get_database_privileges(cursor, user, db) return len(curr_priv) < len(prev_priv) def revoke_privileges(cursor, user, privs): if privs is None: return False changed = False for type_ in privs: revoke_func = { 'table':revoke_table_privilege, 'database':revoke_database_privilege }[type_] for name, privileges in privs[type_].iteritems(): for privilege in privileges: changed = revoke_func(cursor, user, name, privilege)\ or changed return changed def grant_privileges(cursor, user, privs): if privs is None: return False changed = False for type_ in privs: grant_func = { 'table':grant_table_privilege, 'database':grant_database_privilege }[type_] for name, privileges in privs[type_].iteritems(): for privilege in privileges: changed = grant_func(cursor, user, name, privilege)\ or changed return changed def parse_role_attrs(role_attr_flags): """ Parse role attributes string for user creation. Format: attributes[,attributes,...] Where: attributes := CREATEDB,CREATEROLE,NOSUPERUSER,... [ "[NO]SUPERUSER","[NO]CREATEROLE", "[NO]CREATEUSER", "[NO]CREATEDB", "[NO]INHERIT", "[NO]LOGIN", "[NO]REPLICATION" ] """ if ',' in role_attr_flags: flag_set = frozenset(r.upper() for r in role_attr_flags.split(",")) elif role_attr_flags: flag_set = frozenset((role_attr_flags.upper(),)) else: flag_set = frozenset() if not flag_set.issubset(VALID_FLAGS): raise InvalidFlagsError('Invalid role_attr_flags specified: %s' % ' '.join(flag_set.difference(VALID_FLAGS))) o_flags = ' '.join(flag_set) return o_flags def parse_privs(privs, db): """ Parse privilege string to determine permissions for database db. Format: privileges[/privileges/...] Where: privileges := DATABASE_PRIVILEGES[,DATABASE_PRIVILEGES,...] | TABLE_NAME:TABLE_PRIVILEGES[,TABLE_PRIVILEGES,...] """ if privs is None: return privs o_privs = { 'database':{}, 'table':{} } for token in privs.split('/'): if ':' not in token: type_ = 'database' name = db priv_set = frozenset(x.strip().upper() for x in token.split(',') if x.strip()) else: type_ = 'table' name, privileges = token.split(':', 1) priv_set = frozenset(x.strip().upper() for x in privileges.split(',') if x.strip()) if not priv_set.issubset(VALID_PRIVS[type_]): raise InvalidPrivsError('Invalid privs specified for %s: %s' % (type_, ' '.join(priv_set.difference(VALID_PRIVS[type_])))) o_privs[type_][name] = priv_set return o_privs # =========================================== # Module execution. # def main(): module = AnsibleModule( argument_spec=dict( login_user=dict(default="postgres"), login_password=dict(default=""), login_host=dict(default=""), login_unix_socket=dict(default=""), user=dict(required=True, aliases=['name']), password=dict(default=None), state=dict(default="present", choices=["absent", "present"]), priv=dict(default=None), db=dict(default=''), port=dict(default='5432'), fail_on_user=dict(type='bool', default='yes'), role_attr_flags=dict(default=''), encrypted=dict(type='bool', default='no'), expires=dict(default=None) ), supports_check_mode = True ) user = module.params["user"] password = module.params["password"] state = module.params["state"] fail_on_user = module.params["fail_on_user"] db = module.params["db"] if db == '' and module.params["priv"] is not None: module.fail_json(msg="privileges require a database to be specified") privs = parse_privs(module.params["priv"], db) port = module.params["port"] try: role_attr_flags = parse_role_attrs(module.params["role_attr_flags"]) except InvalidFlagsError, e: module.fail_json(msg=str(e)) if module.params["encrypted"]: encrypted = "ENCRYPTED" else: encrypted = "UNENCRYPTED" expires = module.params["expires"] if not postgresqldb_found: module.fail_json(msg="the python psycopg2 module is required") # To use defaults values, keyword arguments must be absent, so # check which values are empty and don't include in the **kw # dictionary params_map = { "login_host":"host", "login_user":"user", "login_password":"password", "port":"port", "db":"database" } kw = dict( (params_map[k], v) for (k, v) in module.params.iteritems() if k in params_map and v != "" ) # If a login_unix_socket is specified, incorporate it here. is_localhost = "host" not in kw or kw["host"] == "" or kw["host"] == "localhost" if is_localhost and module.params["login_unix_socket"] != "": kw["host"] = module.params["login_unix_socket"] try: db_connection = psycopg2.connect(**kw) cursor = db_connection.cursor() except Exception, e: module.fail_json(msg="unable to connect to database: %s" % e) kw = dict(user=user) changed = False user_removed = False if state == "present": if user_exists(cursor, user): try: changed = user_alter(cursor, module, user, password, role_attr_flags, encrypted, expires) except SQLParseError, e: module.fail_json(msg=str(e)) else: try: changed = user_add(cursor, user, password, role_attr_flags, encrypted, expires) except SQLParseError, e: module.fail_json(msg=str(e)) try: changed = grant_privileges(cursor, user, privs) or changed except SQLParseError, e: module.fail_json(msg=str(e)) else: if user_exists(cursor, user): if module.check_mode: changed = True kw['user_removed'] = True else: try: changed = revoke_privileges(cursor, user, privs) user_removed = user_delete(cursor, user) except SQLParseError, e: module.fail_json(msg=str(e)) changed = changed or user_removed if fail_on_user and not user_removed: msg = "unable to remove user" module.fail_json(msg=msg) kw['user_removed'] = user_removed if changed: if module.check_mode: db_connection.rollback() else: db_connection.commit() kw['changed'] = changed module.exit_json(**kw) # import module snippets from ansible.module_utils.basic import * from ansible.module_utils.database import * main()
wangli1426/heron
refs/heads/master
heron/pyheron/src/python/component/base_component.py
8
# Copyright 2016 - Parsely, Inc. (d/b/a Parse.ly) # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. '''base_component.py''' class NotCompatibleError(RuntimeError): """Method is not compatible in PyHeron""" pass class BaseComponent(object): """Base component for PyHeron spout and bolt""" def __init__(self, delegate): """Initializes BaseComponent :param delegate: SpoutInstance or BoltInstance """ self.delegate = delegate self.logger = self.delegate.logger def log(self, message, level=None): """Log message, optionally providing a logging level It is compatible with StreamParse API. :type message: str :param message: the log message to send :type level: str :param level: the logging level, one of: trace (=debug), debug, info, warn or error (default: info) """ self.delegate.log(message, level) # pylint: disable=unused-argument # pylint: disable=no-self-use @staticmethod def is_heartbeat(tup): """Streamparse API that is not applicable for PyHeron""" raise NotCompatibleError("is_heartbeat() method is not applicable in Heron.") def raise_exception(self, exception, tup=None): """Streamparse API that is not applicable for PyHeron""" raise NotCompatibleError("raise_exception() method is not applicable in Heron.") def read_handshake(self): """Streamparse API that is not applicable for PyHeron""" raise NotCompatibleError("read_handshake() method is not applicable in Heron.") def read_message(self): """Streamparse API that is not applicable for PyHeron""" raise NotCompatibleError("read_message() method is not applicable in Heron.") def report_metric(self, name, value): """Streamparse API that is not applicable for PyHeron""" raise NotCompatibleError("report_metric() method is not applicable in Heron.\n" "Please use a metrics collector instead.") def run(self): """Streamparse API that is not applicable for PyHeron""" raise NotCompatibleError("run() method is not applicable in Heron.") def send_message(self, message): """Streamparse API that is not applicable for PyHeron""" raise NotCompatibleError("send_message() method is not applicable in Heron.")
ryfeus/lambda-packs
refs/heads/master
Rasterio_osgeo_shapely_PIL_pyproj_numpy/source/dateutil/parser.py
49
# -*- coding:iso-8859-1 -*- """ This module offers a generic date/time string parser which is able to parse most known formats to represent a date and/or time. This module attempts to be forgiving with regards to unlikely input formats, returning a datetime object even for dates which are ambiguous. If an element of a date/time stamp is omitted, the following rules are applied: - If AM or PM is left unspecified, a 24-hour clock is assumed, however, an hour on a 12-hour clock (``0 <= hour <= 12``) *must* be specified if AM or PM is specified. - If a time zone is omitted, a timezone-naive datetime is returned. If any other elements are missing, they are taken from the :class:`datetime.datetime` object passed to the parameter ``default``. If this results in a day number exceeding the valid number of days per month, the value falls back to the end of the month. Additional resources about date/time string formats can be found below: - `A summary of the international standard date and time notation <http://www.cl.cam.ac.uk/~mgk25/iso-time.html>`_ - `W3C Date and Time Formats <http://www.w3.org/TR/NOTE-datetime>`_ - `Time Formats (Planetary Rings Node) <http://pds-rings.seti.org/tools/time_formats.html>`_ - `CPAN ParseDate module <http://search.cpan.org/~muir/Time-modules-2013.0912/lib/Time/ParseDate.pm>`_ - `Java SimpleDateFormat Class <https://docs.oracle.com/javase/6/docs/api/java/text/SimpleDateFormat.html>`_ """ from __future__ import unicode_literals import datetime import string import time import collections import re from io import StringIO from calendar import monthrange, isleap from six import text_type, binary_type, integer_types from . import relativedelta from . import tz __all__ = ["parse", "parserinfo"] class _timelex(object): # Fractional seconds are sometimes split by a comma _split_decimal = re.compile("([\.,])") def __init__(self, instream): if isinstance(instream, binary_type): instream = instream.decode() if isinstance(instream, text_type): instream = StringIO(instream) if getattr(instream, 'read', None) is None: raise TypeError('Parser must be a string or character stream, not ' '{itype}'.format(itype=instream.__class__.__name__)) self.instream = instream self.charstack = [] self.tokenstack = [] self.eof = False def get_token(self): """ This function breaks the time string into lexical units (tokens), which can be parsed by the parser. Lexical units are demarcated by changes in the character set, so any continuous string of letters is considered one unit, any continuous string of numbers is considered one unit. The main complication arises from the fact that dots ('.') can be used both as separators (e.g. "Sep.20.2009") or decimal points (e.g. "4:30:21.447"). As such, it is necessary to read the full context of any dot-separated strings before breaking it into tokens; as such, this function maintains a "token stack", for when the ambiguous context demands that multiple tokens be parsed at once. """ if self.tokenstack: return self.tokenstack.pop(0) seenletters = False token = None state = None while not self.eof: # We only realize that we've reached the end of a token when we # find a character that's not part of the current token - since # that character may be part of the next token, it's stored in the # charstack. if self.charstack: nextchar = self.charstack.pop(0) else: nextchar = self.instream.read(1) while nextchar == '\x00': nextchar = self.instream.read(1) if not nextchar: self.eof = True break elif not state: # First character of the token - determines if we're starting # to parse a word, a number or something else. token = nextchar if self.isword(nextchar): state = 'a' elif self.isnum(nextchar): state = '0' elif self.isspace(nextchar): token = ' ' break # emit token else: break # emit token elif state == 'a': # If we've already started reading a word, we keep reading # letters until we find something that's not part of a word. seenletters = True if self.isword(nextchar): token += nextchar elif nextchar == '.': token += nextchar state = 'a.' else: self.charstack.append(nextchar) break # emit token elif state == '0': # If we've already started reading a number, we keep reading # numbers until we find something that doesn't fit. if self.isnum(nextchar): token += nextchar elif nextchar == '.' or (nextchar == ',' and len(token) >= 2): token += nextchar state = '0.' else: self.charstack.append(nextchar) break # emit token elif state == 'a.': # If we've seen some letters and a dot separator, continue # parsing, and the tokens will be broken up later. seenletters = True if nextchar == '.' or self.isword(nextchar): token += nextchar elif self.isnum(nextchar) and token[-1] == '.': token += nextchar state = '0.' else: self.charstack.append(nextchar) break # emit token elif state == '0.': # If we've seen at least one dot separator, keep going, we'll # break up the tokens later. if nextchar == '.' or self.isnum(nextchar): token += nextchar elif self.isword(nextchar) and token[-1] == '.': token += nextchar state = 'a.' else: self.charstack.append(nextchar) break # emit token if (state in ('a.', '0.') and (seenletters or token.count('.') > 1 or token[-1] in '.,')): l = self._split_decimal.split(token) token = l[0] for tok in l[1:]: if tok: self.tokenstack.append(tok) if state == '0.' and token.count('.') == 0: token = token.replace(',', '.') return token def __iter__(self): return self def __next__(self): token = self.get_token() if token is None: raise StopIteration return token def next(self): return self.__next__() # Python 2.x support @classmethod def split(cls, s): return list(cls(s)) @classmethod def isword(cls, nextchar): """ Whether or not the next character is part of a word """ return nextchar.isalpha() @classmethod def isnum(cls, nextchar): """ Whether the next character is part of a number """ return nextchar.isdigit() @classmethod def isspace(cls, nextchar): """ Whether the next character is whitespace """ return nextchar.isspace() class _resultbase(object): def __init__(self): for attr in self.__slots__: setattr(self, attr, None) def _repr(self, classname): l = [] for attr in self.__slots__: value = getattr(self, attr) if value is not None: l.append("%s=%s" % (attr, repr(value))) return "%s(%s)" % (classname, ", ".join(l)) def __len__(self): return (sum(getattr(self, attr) is not None for attr in self.__slots__)) def __repr__(self): return self._repr(self.__class__.__name__) class parserinfo(object): """ Class which handles what inputs are accepted. Subclass this to customize the language and acceptable values for each parameter. :param dayfirst: Whether to interpret the first value in an ambiguous 3-integer date (e.g. 01/05/09) as the day (``True``) or month (``False``). If ``yearfirst`` is set to ``True``, this distinguishes between YDM and YMD. Default is ``False``. :param yearfirst: Whether to interpret the first value in an ambiguous 3-integer date (e.g. 01/05/09) as the year. If ``True``, the first number is taken to be the year, otherwise the last number is taken to be the year. Default is ``False``. """ # m from a.m/p.m, t from ISO T separator JUMP = [" ", ".", ",", ";", "-", "/", "'", "at", "on", "and", "ad", "m", "t", "of", "st", "nd", "rd", "th"] WEEKDAYS = [("Mon", "Monday"), ("Tue", "Tuesday"), ("Wed", "Wednesday"), ("Thu", "Thursday"), ("Fri", "Friday"), ("Sat", "Saturday"), ("Sun", "Sunday")] MONTHS = [("Jan", "January"), ("Feb", "February"), ("Mar", "March"), ("Apr", "April"), ("May", "May"), ("Jun", "June"), ("Jul", "July"), ("Aug", "August"), ("Sep", "Sept", "September"), ("Oct", "October"), ("Nov", "November"), ("Dec", "December")] HMS = [("h", "hour", "hours"), ("m", "minute", "minutes"), ("s", "second", "seconds")] AMPM = [("am", "a"), ("pm", "p")] UTCZONE = ["UTC", "GMT", "Z"] PERTAIN = ["of"] TZOFFSET = {} def __init__(self, dayfirst=False, yearfirst=False): self._jump = self._convert(self.JUMP) self._weekdays = self._convert(self.WEEKDAYS) self._months = self._convert(self.MONTHS) self._hms = self._convert(self.HMS) self._ampm = self._convert(self.AMPM) self._utczone = self._convert(self.UTCZONE) self._pertain = self._convert(self.PERTAIN) self.dayfirst = dayfirst self.yearfirst = yearfirst self._year = time.localtime().tm_year self._century = self._year // 100 * 100 def _convert(self, lst): dct = {} for i, v in enumerate(lst): if isinstance(v, tuple): for v in v: dct[v.lower()] = i else: dct[v.lower()] = i return dct def jump(self, name): return name.lower() in self._jump def weekday(self, name): if len(name) >= 3: try: return self._weekdays[name.lower()] except KeyError: pass return None def month(self, name): if len(name) >= 3: try: return self._months[name.lower()] + 1 except KeyError: pass return None def hms(self, name): try: return self._hms[name.lower()] except KeyError: return None def ampm(self, name): try: return self._ampm[name.lower()] except KeyError: return None def pertain(self, name): return name.lower() in self._pertain def utczone(self, name): return name.lower() in self._utczone def tzoffset(self, name): if name in self._utczone: return 0 return self.TZOFFSET.get(name) def convertyear(self, year, century_specified=False): if year < 100 and not century_specified: year += self._century if abs(year - self._year) >= 50: if year < self._year: year += 100 else: year -= 100 return year def validate(self, res): # move to info if res.year is not None: res.year = self.convertyear(res.year, res.century_specified) if res.tzoffset == 0 and not res.tzname or res.tzname == 'Z': res.tzname = "UTC" res.tzoffset = 0 elif res.tzoffset != 0 and res.tzname and self.utczone(res.tzname): res.tzoffset = 0 return True class _ymd(list): def __init__(self, tzstr, *args, **kwargs): super(self.__class__, self).__init__(*args, **kwargs) self.century_specified = False self.tzstr = tzstr @staticmethod def token_could_be_year(token, year): try: return int(token) == year except ValueError: return False @staticmethod def find_potential_year_tokens(year, tokens): return [token for token in tokens if _ymd.token_could_be_year(token, year)] def find_probable_year_index(self, tokens): """ attempt to deduce if a pre 100 year was lost due to padded zeros being taken off """ for index, token in enumerate(self): potential_year_tokens = _ymd.find_potential_year_tokens(token, tokens) if len(potential_year_tokens) == 1 and len(potential_year_tokens[0]) > 2: return index def append(self, val): if hasattr(val, '__len__'): if val.isdigit() and len(val) > 2: self.century_specified = True elif val > 100: self.century_specified = True super(self.__class__, self).append(int(val)) def resolve_ymd(self, mstridx, yearfirst, dayfirst): len_ymd = len(self) year, month, day = (None, None, None) if len_ymd > 3: raise ValueError("More than three YMD values") elif len_ymd == 1 or (mstridx != -1 and len_ymd == 2): # One member, or two members with a month string if mstridx != -1: month = self[mstridx] del self[mstridx] if len_ymd > 1 or mstridx == -1: if self[0] > 31: year = self[0] else: day = self[0] elif len_ymd == 2: # Two members with numbers if self[0] > 31: # 99-01 year, month = self elif self[1] > 31: # 01-99 month, year = self elif dayfirst and self[1] <= 12: # 13-01 day, month = self else: # 01-13 month, day = self elif len_ymd == 3: # Three members if mstridx == 0: month, day, year = self elif mstridx == 1: if self[0] > 31 or (yearfirst and self[2] <= 31): # 99-Jan-01 year, month, day = self else: # 01-Jan-01 # Give precendence to day-first, since # two-digit years is usually hand-written. day, month, year = self elif mstridx == 2: # WTF!? if self[1] > 31: # 01-99-Jan day, year, month = self else: # 99-01-Jan year, day, month = self else: if self[0] > 31 or \ self.find_probable_year_index(_timelex.split(self.tzstr)) == 0 or \ (yearfirst and self[1] <= 12 and self[2] <= 31): # 99-01-01 if dayfirst and self[2] <= 12: year, day, month = self else: year, month, day = self elif self[0] > 12 or (dayfirst and self[1] <= 12): # 13-01-01 day, month, year = self else: # 01-13-01 month, day, year = self return year, month, day class parser(object): def __init__(self, info=None): self.info = info or parserinfo() def parse(self, timestr, default=None, ignoretz=False, tzinfos=None, **kwargs): """ Parse the date/time string into a :class:`datetime.datetime` object. :param timestr: Any date/time string using the supported formats. :param default: The default datetime object, if this is a datetime object and not ``None``, elements specified in ``timestr`` replace elements in the default object. :param ignoretz: If set ``True``, time zones in parsed strings are ignored and a naive :class:`datetime.datetime` object is returned. :param tzinfos: Additional time zone names / aliases which may be present in the string. This argument maps time zone names (and optionally offsets from those time zones) to time zones. This parameter can be a dictionary with timezone aliases mapping time zone names to time zones or a function taking two parameters (``tzname`` and ``tzoffset``) and returning a time zone. The timezones to which the names are mapped can be an integer offset from UTC in minutes or a :class:`tzinfo` object. .. doctest:: :options: +NORMALIZE_WHITESPACE >>> from dateutil.parser import parse >>> from dateutil.tz import gettz >>> tzinfos = {"BRST": -10800, "CST": gettz("America/Chicago")} >>> parse("2012-01-19 17:21:00 BRST", tzinfos=tzinfos) datetime.datetime(2012, 1, 19, 17, 21, tzinfo=tzoffset(u'BRST', -10800)) >>> parse("2012-01-19 17:21:00 CST", tzinfos=tzinfos) datetime.datetime(2012, 1, 19, 17, 21, tzinfo=tzfile('/usr/share/zoneinfo/America/Chicago')) This parameter is ignored if ``ignoretz`` is set. :param **kwargs: Keyword arguments as passed to ``_parse()``. :return: Returns a :class:`datetime.datetime` object or, if the ``fuzzy_with_tokens`` option is ``True``, returns a tuple, the first element being a :class:`datetime.datetime` object, the second a tuple containing the fuzzy tokens. :raises ValueError: Raised for invalid or unknown string format, if the provided :class:`tzinfo` is not in a valid format, or if an invalid date would be created. :raises OverflowError: Raised if the parsed date exceeds the largest valid C integer on your system. """ if default is None: effective_dt = datetime.datetime.now() default = datetime.datetime.now().replace(hour=0, minute=0, second=0, microsecond=0) else: effective_dt = default res, skipped_tokens = self._parse(timestr, **kwargs) if res is None: raise ValueError("Unknown string format") if len(res) == 0: raise ValueError("String does not contain a date.") repl = {} for attr in ("year", "month", "day", "hour", "minute", "second", "microsecond"): value = getattr(res, attr) if value is not None: repl[attr] = value if 'day' not in repl: # If the default day exceeds the last day of the month, fall back to # the end of the month. cyear = default.year if res.year is None else res.year cmonth = default.month if res.month is None else res.month cday = default.day if res.day is None else res.day if cday > monthrange(cyear, cmonth)[1]: repl['day'] = monthrange(cyear, cmonth)[1] ret = default.replace(**repl) if res.weekday is not None and not res.day: ret = ret+relativedelta.relativedelta(weekday=res.weekday) if not ignoretz: if (isinstance(tzinfos, collections.Callable) or tzinfos and res.tzname in tzinfos): if isinstance(tzinfos, collections.Callable): tzdata = tzinfos(res.tzname, res.tzoffset) else: tzdata = tzinfos.get(res.tzname) if isinstance(tzdata, datetime.tzinfo): tzinfo = tzdata elif isinstance(tzdata, text_type): tzinfo = tz.tzstr(tzdata) elif isinstance(tzdata, integer_types): tzinfo = tz.tzoffset(res.tzname, tzdata) else: raise ValueError("Offset must be tzinfo subclass, " "tz string, or int offset.") ret = ret.replace(tzinfo=tzinfo) elif res.tzname and res.tzname in time.tzname: ret = ret.replace(tzinfo=tz.tzlocal()) elif res.tzoffset == 0: ret = ret.replace(tzinfo=tz.tzutc()) elif res.tzoffset: ret = ret.replace(tzinfo=tz.tzoffset(res.tzname, res.tzoffset)) if kwargs.get('fuzzy_with_tokens', False): return ret, skipped_tokens else: return ret class _result(_resultbase): __slots__ = ["year", "month", "day", "weekday", "hour", "minute", "second", "microsecond", "tzname", "tzoffset", "ampm"] def _parse(self, timestr, dayfirst=None, yearfirst=None, fuzzy=False, fuzzy_with_tokens=False): """ Private method which performs the heavy lifting of parsing, called from ``parse()``, which passes on its ``kwargs`` to this function. :param timestr: The string to parse. :param dayfirst: Whether to interpret the first value in an ambiguous 3-integer date (e.g. 01/05/09) as the day (``True``) or month (``False``). If ``yearfirst`` is set to ``True``, this distinguishes between YDM and YMD. If set to ``None``, this value is retrieved from the current :class:`parserinfo` object (which itself defaults to ``False``). :param yearfirst: Whether to interpret the first value in an ambiguous 3-integer date (e.g. 01/05/09) as the year. If ``True``, the first number is taken to be the year, otherwise the last number is taken to be the year. If this is set to ``None``, the value is retrieved from the current :class:`parserinfo` object (which itself defaults to ``False``). :param fuzzy: Whether to allow fuzzy parsing, allowing for string like "Today is January 1, 2047 at 8:21:00AM". :param fuzzy_with_tokens: If ``True``, ``fuzzy`` is automatically set to True, and the parser will return a tuple where the first element is the parsed :class:`datetime.datetime` datetimestamp and the second element is a tuple containing the portions of the string which were ignored: .. doctest:: >>> from dateutil.parser import parse >>> parse("Today is January 1, 2047 at 8:21:00AM", fuzzy_with_tokens=True) (datetime.datetime(2047, 1, 1, 8, 21), (u'Today is ', u' ', u'at ')) """ if fuzzy_with_tokens: fuzzy = True info = self.info if dayfirst is None: dayfirst = info.dayfirst if yearfirst is None: yearfirst = info.yearfirst res = self._result() l = _timelex.split(timestr) # Splits the timestr into tokens # keep up with the last token skipped so we can recombine # consecutively skipped tokens (-2 for when i begins at 0). last_skipped_token_i = -2 skipped_tokens = list() try: # year/month/day list ymd = _ymd(timestr) # Index of the month string in ymd mstridx = -1 len_l = len(l) i = 0 while i < len_l: # Check if it's a number try: value_repr = l[i] value = float(value_repr) except ValueError: value = None if value is not None: # Token is a number len_li = len(l[i]) i += 1 if (len(ymd) == 3 and len_li in (2, 4) and res.hour is None and (i >= len_l or (l[i] != ':' and info.hms(l[i]) is None))): # 19990101T23[59] s = l[i-1] res.hour = int(s[:2]) if len_li == 4: res.minute = int(s[2:]) elif len_li == 6 or (len_li > 6 and l[i-1].find('.') == 6): # YYMMDD or HHMMSS[.ss] s = l[i-1] if not ymd and l[i-1].find('.') == -1: #ymd.append(info.convertyear(int(s[:2]))) ymd.append(s[:2]) ymd.append(s[2:4]) ymd.append(s[4:]) else: # 19990101T235959[.59] res.hour = int(s[:2]) res.minute = int(s[2:4]) res.second, res.microsecond = _parsems(s[4:]) elif len_li in (8, 12, 14): # YYYYMMDD s = l[i-1] ymd.append(s[:4]) ymd.append(s[4:6]) ymd.append(s[6:8]) if len_li > 8: res.hour = int(s[8:10]) res.minute = int(s[10:12]) if len_li > 12: res.second = int(s[12:]) elif ((i < len_l and info.hms(l[i]) is not None) or (i+1 < len_l and l[i] == ' ' and info.hms(l[i+1]) is not None)): # HH[ ]h or MM[ ]m or SS[.ss][ ]s if l[i] == ' ': i += 1 idx = info.hms(l[i]) while True: if idx == 0: res.hour = int(value) if value % 1: res.minute = int(60*(value % 1)) elif idx == 1: res.minute = int(value) if value % 1: res.second = int(60*(value % 1)) elif idx == 2: res.second, res.microsecond = \ _parsems(value_repr) i += 1 if i >= len_l or idx == 2: break # 12h00 try: value_repr = l[i] value = float(value_repr) except ValueError: break else: i += 1 idx += 1 if i < len_l: newidx = info.hms(l[i]) if newidx is not None: idx = newidx elif (i == len_l and l[i-2] == ' ' and info.hms(l[i-3]) is not None): # X h MM or X m SS idx = info.hms(l[i-3]) + 1 if idx == 1: res.minute = int(value) if value % 1: res.second = int(60*(value % 1)) elif idx == 2: res.second, res.microsecond = \ _parsems(value_repr) i += 1 elif i+1 < len_l and l[i] == ':': # HH:MM[:SS[.ss]] res.hour = int(value) i += 1 value = float(l[i]) res.minute = int(value) if value % 1: res.second = int(60*(value % 1)) i += 1 if i < len_l and l[i] == ':': res.second, res.microsecond = _parsems(l[i+1]) i += 2 elif i < len_l and l[i] in ('-', '/', '.'): sep = l[i] ymd.append(value_repr) i += 1 if i < len_l and not info.jump(l[i]): try: # 01-01[-01] ymd.append(l[i]) except ValueError: # 01-Jan[-01] value = info.month(l[i]) if value is not None: ymd.append(value) assert mstridx == -1 mstridx = len(ymd)-1 else: return None, None i += 1 if i < len_l and l[i] == sep: # We have three members i += 1 value = info.month(l[i]) if value is not None: ymd.append(value) mstridx = len(ymd)-1 assert mstridx == -1 else: ymd.append(l[i]) i += 1 elif i >= len_l or info.jump(l[i]): if i+1 < len_l and info.ampm(l[i+1]) is not None: # 12 am res.hour = int(value) if res.hour < 12 and info.ampm(l[i+1]) == 1: res.hour += 12 elif res.hour == 12 and info.ampm(l[i+1]) == 0: res.hour = 0 i += 1 else: # Year, month or day ymd.append(value) i += 1 elif info.ampm(l[i]) is not None: # 12am res.hour = int(value) if res.hour < 12 and info.ampm(l[i]) == 1: res.hour += 12 elif res.hour == 12 and info.ampm(l[i]) == 0: res.hour = 0 i += 1 elif not fuzzy: return None, None else: i += 1 continue # Check weekday value = info.weekday(l[i]) if value is not None: res.weekday = value i += 1 continue # Check month name value = info.month(l[i]) if value is not None: ymd.append(value) assert mstridx == -1 mstridx = len(ymd)-1 i += 1 if i < len_l: if l[i] in ('-', '/'): # Jan-01[-99] sep = l[i] i += 1 ymd.append(l[i]) i += 1 if i < len_l and l[i] == sep: # Jan-01-99 i += 1 ymd.append(l[i]) i += 1 elif (i+3 < len_l and l[i] == l[i+2] == ' ' and info.pertain(l[i+1])): # Jan of 01 # In this case, 01 is clearly year try: value = int(l[i+3]) except ValueError: # Wrong guess pass else: # Convert it here to become unambiguous ymd.append(str(info.convertyear(value))) i += 4 continue # Check am/pm value = info.ampm(l[i]) if value is not None: # For fuzzy parsing, 'a' or 'am' (both valid English words) # may erroneously trigger the AM/PM flag. Deal with that # here. val_is_ampm = True # If there's already an AM/PM flag, this one isn't one. if fuzzy and res.ampm is not None: val_is_ampm = False # If AM/PM is found and hour is not, raise a ValueError if res.hour is None: if fuzzy: val_is_ampm = False else: raise ValueError('No hour specified with ' + 'AM or PM flag.') elif not 0 <= res.hour <= 12: # If AM/PM is found, it's a 12 hour clock, so raise # an error for invalid range if fuzzy: val_is_ampm = False else: raise ValueError('Invalid hour specified for ' + '12-hour clock.') if val_is_ampm: if value == 1 and res.hour < 12: res.hour += 12 elif value == 0 and res.hour == 12: res.hour = 0 res.ampm = value i += 1 continue # Check for a timezone name if (res.hour is not None and len(l[i]) <= 5 and res.tzname is None and res.tzoffset is None and not [x for x in l[i] if x not in string.ascii_uppercase]): res.tzname = l[i] res.tzoffset = info.tzoffset(res.tzname) i += 1 # Check for something like GMT+3, or BRST+3. Notice # that it doesn't mean "I am 3 hours after GMT", but # "my time +3 is GMT". If found, we reverse the # logic so that timezone parsing code will get it # right. if i < len_l and l[i] in ('+', '-'): l[i] = ('+', '-')[l[i] == '+'] res.tzoffset = None if info.utczone(res.tzname): # With something like GMT+3, the timezone # is *not* GMT. res.tzname = None continue # Check for a numbered timezone if res.hour is not None and l[i] in ('+', '-'): signal = (-1, 1)[l[i] == '+'] i += 1 len_li = len(l[i]) if len_li == 4: # -0300 res.tzoffset = int(l[i][:2])*3600+int(l[i][2:])*60 elif i+1 < len_l and l[i+1] == ':': # -03:00 res.tzoffset = int(l[i])*3600+int(l[i+2])*60 i += 2 elif len_li <= 2: # -[0]3 res.tzoffset = int(l[i][:2])*3600 else: return None, None i += 1 res.tzoffset *= signal # Look for a timezone name between parenthesis if (i+3 < len_l and info.jump(l[i]) and l[i+1] == '(' and l[i+3] == ')' and 3 <= len(l[i+2]) <= 5 and not [x for x in l[i+2] if x not in string.ascii_uppercase]): # -0300 (BRST) res.tzname = l[i+2] i += 4 continue # Check jumps if not (info.jump(l[i]) or fuzzy): return None, None if last_skipped_token_i == i - 1: # recombine the tokens skipped_tokens[-1] += l[i] else: # just append skipped_tokens.append(l[i]) last_skipped_token_i = i i += 1 # Process year/month/day year, month, day = ymd.resolve_ymd(mstridx, yearfirst, dayfirst) if year is not None: res.year = year res.century_specified = ymd.century_specified if month is not None: res.month = month if day is not None: res.day = day except (IndexError, ValueError, AssertionError): return None, None if not info.validate(res): return None, None if fuzzy_with_tokens: return res, tuple(skipped_tokens) else: return res, None DEFAULTPARSER = parser() def parse(timestr, parserinfo=None, **kwargs): """ Parse a string in one of the supported formats, using the ``parserinfo`` parameters. :param timestr: A string containing a date/time stamp. :param parserinfo: A :class:`parserinfo` object containing parameters for the parser. If ``None``, the default arguments to the :class:`parserinfo` constructor are used. The ``**kwargs`` parameter takes the following keyword arguments: :param default: The default datetime object, if this is a datetime object and not ``None``, elements specified in ``timestr`` replace elements in the default object. :param ignoretz: If set ``True``, time zones in parsed strings are ignored and a naive :class:`datetime` object is returned. :param tzinfos: Additional time zone names / aliases which may be present in the string. This argument maps time zone names (and optionally offsets from those time zones) to time zones. This parameter can be a dictionary with timezone aliases mapping time zone names to time zones or a function taking two parameters (``tzname`` and ``tzoffset``) and returning a time zone. The timezones to which the names are mapped can be an integer offset from UTC in minutes or a :class:`tzinfo` object. .. doctest:: :options: +NORMALIZE_WHITESPACE >>> from dateutil.parser import parse >>> from dateutil.tz import gettz >>> tzinfos = {"BRST": -10800, "CST": gettz("America/Chicago")} >>> parse("2012-01-19 17:21:00 BRST", tzinfos=tzinfos) datetime.datetime(2012, 1, 19, 17, 21, tzinfo=tzoffset(u'BRST', -10800)) >>> parse("2012-01-19 17:21:00 CST", tzinfos=tzinfos) datetime.datetime(2012, 1, 19, 17, 21, tzinfo=tzfile('/usr/share/zoneinfo/America/Chicago')) This parameter is ignored if ``ignoretz`` is set. :param dayfirst: Whether to interpret the first value in an ambiguous 3-integer date (e.g. 01/05/09) as the day (``True``) or month (``False``). If ``yearfirst`` is set to ``True``, this distinguishes between YDM and YMD. If set to ``None``, this value is retrieved from the current :class:`parserinfo` object (which itself defaults to ``False``). :param yearfirst: Whether to interpret the first value in an ambiguous 3-integer date (e.g. 01/05/09) as the year. If ``True``, the first number is taken to be the year, otherwise the last number is taken to be the year. If this is set to ``None``, the value is retrieved from the current :class:`parserinfo` object (which itself defaults to ``False``). :param fuzzy: Whether to allow fuzzy parsing, allowing for string like "Today is January 1, 2047 at 8:21:00AM". :param fuzzy_with_tokens: If ``True``, ``fuzzy`` is automatically set to True, and the parser will return a tuple where the first element is the parsed :class:`datetime.datetime` datetimestamp and the second element is a tuple containing the portions of the string which were ignored: .. doctest:: >>> from dateutil.parser import parse >>> parse("Today is January 1, 2047 at 8:21:00AM", fuzzy_with_tokens=True) (datetime.datetime(2011, 1, 1, 8, 21), (u'Today is ', u' ', u'at ')) :return: Returns a :class:`datetime.datetime` object or, if the ``fuzzy_with_tokens`` option is ``True``, returns a tuple, the first element being a :class:`datetime.datetime` object, the second a tuple containing the fuzzy tokens. :raises ValueError: Raised for invalid or unknown string format, if the provided :class:`tzinfo` is not in a valid format, or if an invalid date would be created. :raises OverflowError: Raised if the parsed date exceeds the largest valid C integer on your system. """ if parserinfo: return parser(parserinfo).parse(timestr, **kwargs) else: return DEFAULTPARSER.parse(timestr, **kwargs) class _tzparser(object): class _result(_resultbase): __slots__ = ["stdabbr", "stdoffset", "dstabbr", "dstoffset", "start", "end"] class _attr(_resultbase): __slots__ = ["month", "week", "weekday", "yday", "jyday", "day", "time"] def __repr__(self): return self._repr("") def __init__(self): _resultbase.__init__(self) self.start = self._attr() self.end = self._attr() def parse(self, tzstr): res = self._result() l = _timelex.split(tzstr) try: len_l = len(l) i = 0 while i < len_l: # BRST+3[BRDT[+2]] j = i while j < len_l and not [x for x in l[j] if x in "0123456789:,-+"]: j += 1 if j != i: if not res.stdabbr: offattr = "stdoffset" res.stdabbr = "".join(l[i:j]) else: offattr = "dstoffset" res.dstabbr = "".join(l[i:j]) i = j if (i < len_l and (l[i] in ('+', '-') or l[i][0] in "0123456789")): if l[i] in ('+', '-'): # Yes, that's right. See the TZ variable # documentation. signal = (1, -1)[l[i] == '+'] i += 1 else: signal = -1 len_li = len(l[i]) if len_li == 4: # -0300 setattr(res, offattr, (int(l[i][:2])*3600 + int(l[i][2:])*60)*signal) elif i+1 < len_l and l[i+1] == ':': # -03:00 setattr(res, offattr, (int(l[i])*3600+int(l[i+2])*60)*signal) i += 2 elif len_li <= 2: # -[0]3 setattr(res, offattr, int(l[i][:2])*3600*signal) else: return None i += 1 if res.dstabbr: break else: break if i < len_l: for j in range(i, len_l): if l[j] == ';': l[j] = ',' assert l[i] == ',' i += 1 if i >= len_l: pass elif (8 <= l.count(',') <= 9 and not [y for x in l[i:] if x != ',' for y in x if y not in "0123456789"]): # GMT0BST,3,0,30,3600,10,0,26,7200[,3600] for x in (res.start, res.end): x.month = int(l[i]) i += 2 if l[i] == '-': value = int(l[i+1])*-1 i += 1 else: value = int(l[i]) i += 2 if value: x.week = value x.weekday = (int(l[i])-1) % 7 else: x.day = int(l[i]) i += 2 x.time = int(l[i]) i += 2 if i < len_l: if l[i] in ('-', '+'): signal = (-1, 1)[l[i] == "+"] i += 1 else: signal = 1 res.dstoffset = (res.stdoffset+int(l[i]))*signal elif (l.count(',') == 2 and l[i:].count('/') <= 2 and not [y for x in l[i:] if x not in (',', '/', 'J', 'M', '.', '-', ':') for y in x if y not in "0123456789"]): for x in (res.start, res.end): if l[i] == 'J': # non-leap year day (1 based) i += 1 x.jyday = int(l[i]) elif l[i] == 'M': # month[-.]week[-.]weekday i += 1 x.month = int(l[i]) i += 1 assert l[i] in ('-', '.') i += 1 x.week = int(l[i]) if x.week == 5: x.week = -1 i += 1 assert l[i] in ('-', '.') i += 1 x.weekday = (int(l[i])-1) % 7 else: # year day (zero based) x.yday = int(l[i])+1 i += 1 if i < len_l and l[i] == '/': i += 1 # start time len_li = len(l[i]) if len_li == 4: # -0300 x.time = (int(l[i][:2])*3600+int(l[i][2:])*60) elif i+1 < len_l and l[i+1] == ':': # -03:00 x.time = int(l[i])*3600+int(l[i+2])*60 i += 2 if i+1 < len_l and l[i+1] == ':': i += 2 x.time += int(l[i]) elif len_li <= 2: # -[0]3 x.time = (int(l[i][:2])*3600) else: return None i += 1 assert i == len_l or l[i] == ',' i += 1 assert i >= len_l except (IndexError, ValueError, AssertionError): return None return res DEFAULTTZPARSER = _tzparser() def _parsetz(tzstr): return DEFAULTTZPARSER.parse(tzstr) def _parsems(value): """Parse a I[.F] seconds value into (seconds, microseconds).""" if "." not in value: return int(value), 0 else: i, f = value.split(".") return int(i), int(f.ljust(6, "0")[:6]) # vim:ts=4:sw=4:et
Lab603/PicEncyclopedias
refs/heads/master
jni-build/jni-build/jni/include/tensorflow/python/framework/tensor_shape.py
7
# Copyright 2015 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== """Helper classes for tensor shape inference.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function from tensorflow.core.framework import tensor_shape_pb2 from tensorflow.python.util import compat class Dimension(object): """Represents the value of one dimension in a TensorShape.""" def __init__(self, value): """Creates a new Dimension with the given value.""" if value is None: self._value = None else: self._value = int(value) if (not isinstance(value, compat.bytes_or_text_types) and self._value != value): raise ValueError("Ambiguous dimension: %s" % value) if self._value < 0: raise ValueError("Dimension %d must be >= 0" % self._value) def __repr__(self): return "Dimension(%s)" % repr(self._value) def __str__(self): value = self._value return "?" if value is None else str(value) def __eq__(self, other): """Returns true if `other` has the same known value as this Dimension.""" try: other = as_dimension(other) except ValueError: return NotImplemented if self._value is None or other.value is None: return None return self._value == other.value def __ne__(self, other): """Returns true if `other` has a different known value from `self`.""" try: other = as_dimension(other) except ValueError: return NotImplemented if self._value is None or other.value is None: return None return self._value != other.value def __int__(self): return self._value def __index__(self): # Allow use in Python 3 range return self._value @property def value(self): """The value of this dimension, or None if it is unknown.""" return self._value def is_compatible_with(self, other): """Returns true if `other` is compatible with this Dimension. Two known Dimensions are compatible if they have the same value. An unknown Dimension is compatible with all other Dimensions. Args: other: Another Dimension. Returns: True if this Dimension and `other` are compatible. """ other = as_dimension(other) return (self._value is None or other.value is None or self._value == other.value) def assert_is_compatible_with(self, other): """Raises an exception if `other` is not compatible with this Dimension. Args: other: Another Dimension. Raises: ValueError: If `self` and `other` are not compatible (see is_compatible_with). """ if not self.is_compatible_with(other): raise ValueError("Dimensions %s and %s are not compatible" % (self, other)) def merge_with(self, other): """Returns a Dimension that combines the information in `self` and `other`. Dimensions are combined as follows: Dimension(n) .merge_with(Dimension(n)) == Dimension(n) Dimension(n) .merge_with(Dimension(None)) == Dimension(n) Dimension(None).merge_with(Dimension(n)) == Dimension(n) Dimension(None).merge_with(Dimension(None)) == Dimension(None) Dimension(n) .merge_with(Dimension(m)) raises ValueError for n != m Args: other: Another Dimension. Returns: A Dimension containing the combined information of `self` and `other`. Raises: ValueError: If `self` and `other` are not compatible (see is_compatible_with). """ other = as_dimension(other) self.assert_is_compatible_with(other) if self._value is None: return Dimension(other.value) else: return Dimension(self._value) def __add__(self, other): """Returns the sum of `self` and `other`. Dimensions are summed as follows: Dimension(m) + Dimension(n) == Dimension(m + n) Dimension(m) + Dimension(None) == Dimension(None) Dimension(None) + Dimension(n) == Dimension(None) Dimension(None) + Dimension(None) == Dimension(None) Args: other: Another Dimension. Returns: A Dimension whose value is the sum of `self` and `other`. """ other = as_dimension(other) if self._value is None or other.value is None: return Dimension(None) else: return Dimension(self._value + other.value) def __sub__(self, other): """Returns the subtraction of `other` from `self`. Dimensions are subtracted as follows: Dimension(m) - Dimension(n) == Dimension(m - n) Dimension(m) - Dimension(None) == Dimension(None) Dimension(None) - Dimension(n) == Dimension(None) Dimension(None) - Dimension(None) == Dimension(None) Args: other: Another Dimension. Returns: A Dimension whose value is the subtraction of sum of `other` from `self`. """ other = as_dimension(other) if self._value is None or other.value is None: return Dimension(None) else: return Dimension(self._value - other.value) def __mul__(self, other): """Returns the product of `self` and `other`. Dimensions are summed as follows: Dimension(m) * Dimension(n) == Dimension(m * n) Dimension(m) * Dimension(None) == Dimension(None) Dimension(None) * Dimension(n) == Dimension(None) Dimension(None) * Dimension(None) == Dimension(None) Args: other: Another Dimension. Returns: A Dimension whose value is the sum of `self` and `other`. """ other = as_dimension(other) if self._value is None or other.value is None: return Dimension(None) else: return Dimension(self._value * other.value) def __floordiv__(self, other): """Returns the quotient of `self` and `other` rounded down. Dimensions are divided as follows: Dimension(m) // Dimension(n) == Dimension(m // n) Dimension(m) // Dimension(None) == Dimension(None) Dimension(None) // Dimension(n) == Dimension(None) Dimension(None) // Dimension(None) == Dimension(None) Args: other: Another `Dimension`. Returns: A `Dimension` whose value is the integer quotient of `self` and `other`. """ other = as_dimension(other) if self._value is None or other.value is None: return Dimension(None) else: return Dimension(self._value // other.value) def __div__(self, other): """DEPRECATED: Use `__floordiv__` via `x // y` instead. This function exists only for backwards compatibility purposes; new code should use `__floordiv__` via the syntax `x // y`. Using `x // y` communicates clearly that the result rounds down, and is forward compatible to Python 3. Args: other: Another `Dimension`. Returns: A `Dimension` whose value is the integer quotient of `self` and `other`. """ return self // other def __mod__(self, other): """Returns `self` modulo `other. Dimension moduli are computed as follows: Dimension(m) % Dimension(n) == Dimension(m % n) Dimension(m) % Dimension(None) == Dimension(None) Dimension(None) % Dimension(n) == Dimension(None) Dimension(None) % Dimension(None) == Dimension(None) Args: other: Another Dimension. Returns: A Dimension whose value is `self` modulo `other`. """ other = as_dimension(other) if self._value is None or other.value is None: return Dimension(None) else: return Dimension(self._value % other.value) def __lt__(self, other): """Returns True if `self` is known to be less than `other`. Dimensions are compared as follows: Dimension(m) < Dimension(n) == m < n Dimension(m) < Dimension(None) == None Dimension(None) < Dimension(n) == None Dimension(None) < Dimension(None) == None Args: other: Another Dimension. Returns: The value of `self.value < other.value` if both are known, otherwise None. """ other = as_dimension(other) if self._value is None or other.value is None: return None else: return self._value < other.value def __le__(self, other): """Returns True if `self` is known to be less than or equal to `other`. Dimensions are compared as follows: Dimension(m) <= Dimension(n) == m <= n Dimension(m) <= Dimension(None) == None Dimension(None) <= Dimension(n) == None Dimension(None) <= Dimension(None) == None Args: other: Another Dimension. Returns: The value of `self.value <= other.value` if both are known, otherwise None. """ other = as_dimension(other) if self._value is None or other.value is None: return None else: return self._value <= other.value def __gt__(self, other): """Returns True if `self` is known to be greater than `other`. Dimensions are compared as follows: Dimension(m) > Dimension(n) == m > n Dimension(m) > Dimension(None) == None Dimension(None) > Dimension(n) == None Dimension(None) > Dimension(None) == None Args: other: Another Dimension. Returns: The value of `self.value > other.value` if both are known, otherwise None. """ other = as_dimension(other) if self._value is None or other.value is None: return None else: return self._value > other.value def __ge__(self, other): """Returns True if `self` is known to be greater than or equal to `other`. Dimensions are compared as follows: Dimension(m) >= Dimension(n) == m >= n Dimension(m) >= Dimension(None) == None Dimension(None) >= Dimension(n) == None Dimension(None) >= Dimension(None) == None Args: other: Another Dimension. Returns: The value of `self.value >= other.value` if both are known, otherwise None. """ other = as_dimension(other) if self._value is None or other.value is None: return None else: return self._value >= other.value def as_dimension(value): """Converts the given value to a Dimension. A Dimenson input will be returned unmodified. An input of `None` will be converted to an unknown Dimension. An integer input will be converted to a Dimension with that value. Args: value: The value to be converted. Returns: A Dimension corresponding to the given value. """ if isinstance(value, Dimension): return value else: return Dimension(value) class TensorShape(object): """Represents the shape of a `Tensor`. A `TensorShape` represents a possibly-partial shape specification for a `Tensor`. It may be one of the following: * *Fully-known shape:* has a known number of dimensions and a known size for each dimension. * *Partially-known shape:* has a known number of dimensions, and an unknown size for one or more dimension. * *Unknown shape:* has an unknown number of dimensions, and an unknown size in all dimensions. If a tensor is produced by an operation of type `"Foo"`, its shape may be inferred if there is a registered shape function for `"Foo"`. See [`tf.RegisterShape()`](../../api_docs/python/framework.md#RegisterShape) for details of shape functions and how to register them. Alternatively, the shape may be set explicitly using [`Tensor.set_shape()`](../../api_docs/python/framework.md#Tensor.set_shape). @@merge_with @@concatenate @@ndims @@dims @@as_list @@as_proto @@is_compatible_with @@is_fully_defined @@with_rank @@with_rank_at_least @@with_rank_at_most @@assert_has_rank @@assert_same_rank @@assert_is_compatible_with @@assert_is_fully_defined """ def __init__(self, dims): """Creates a new TensorShape with the given dimensions. Args: dims: A list of Dimensions, or None if the shape is unspecified. DEPRECATED: A single integer is treated as a singleton list. Raises: TypeError: If dims cannot be converted to a list of dimensions. """ # TODO(irving): Eliminate the single integer special case. if dims is None: self._dims = None elif isinstance(dims, compat.bytes_or_text_types): raise TypeError("A string has ambiguous TensorShape, please wrap in a " "list or convert to an int: %s" % dims) elif isinstance(dims, tensor_shape_pb2.TensorShapeProto): if dims.unknown_rank: self._dims = None else: self._dims = [ # Protos store variable-size dimensions as -1 as_dimension(dim.size if dim.size != -1 else None) for dim in dims.dim] else: try: dims_iter = iter(dims) except TypeError: # Treat as a singleton dimension self._dims = [as_dimension(dims)] else: # Got a list of dimensions self._dims = [as_dimension(d) for d in dims_iter] def __repr__(self): return "TensorShape(%r)" % self._dims def __str__(self): if self.ndims is None: return "<unknown>" elif self.ndims == 1: return "(%s,)" % self._dims[0] else: return "(%s)" % ", ".join(str(d) for d in self._dims) @property def dims(self): """Returns a list of Dimensions, or None if the shape is unspecified.""" return self._dims @property def ndims(self): """Returns the rank of this shape, or None if it is unspecified.""" if self._dims is None: return None else: return len(self._dims) def __len__(self): """Returns the rank of this shape, or raises ValueError if unspecified.""" if self._dims is None: raise ValueError("Cannot take the length of Shape with unknown rank.") return len(self._dims) def __bool__(self): """Returns True if this shape contains non-zero information.""" return self._dims is not None # Python 3 wants __bool__, Python 2.7 wants __nonzero__ __nonzero__ = __bool__ def __getitem__(self, key): """Returns the value of a dimension or a shape, depending on the key. Args: key: If `key` is an integer, returns the dimension at that index; otherwise if `key` is a slice, returns a TensorShape whose dimensions are those selected by the slice from `self`. Returns: A dimension if `key` is an integer, or a `TensorShape` if `key` is a slice. Raises: ValueError: If `key` is a slice, and any of its elements are negative, or if `self` is completely unknown and the step is set. """ if self._dims is not None: if isinstance(key, slice): return TensorShape(self._dims[key]) else: return self._dims[key] else: if isinstance(key, slice): start = key.start if key.start is not None else 0 stop = key.stop if key.step is not None: # TODO(mrry): Handle these maybe. raise ValueError("Steps are not yet handled") if stop is None: # NOTE(mrry): This implies that TensorShape(None) is compatible with # TensorShape(None)[1:], which is obviously not true. It would be # possible to track the number of dimensions symbolically, # and perhaps we should do that. return unknown_shape() elif start < 0 or stop < 0: # TODO(mrry): Handle this better, as it will be useful for handling # suffixes of otherwise unknown shapes. return unknown_shape() else: return unknown_shape(ndims=stop-start) else: return Dimension(None) def num_elements(self): """Returns the total number of elements, or none for incomplete shapes.""" if self.is_fully_defined(): size = 1 for dim in self._dims: size *= dim.value return size else: return None def merge_with(self, other): """Returns a `TensorShape` combining the information in `self` and `other`. The dimensions in `self` and `other` are merged elementwise, according to the rules defined for `Dimension.merge_with()`. Args: other: Another `TensorShape`. Returns: A `TensorShape` containing the combined information of `self` and `other`. Raises: ValueError: If `self` and `other` are not compatible. """ other = as_shape(other) if self._dims is None: return other else: try: self.assert_same_rank(other) new_dims = [] for i, dim in enumerate(self._dims): new_dims.append(dim.merge_with(other[i])) return TensorShape(new_dims) except ValueError: raise ValueError("Shapes %s and %s are not compatible" % (self, other)) def concatenate(self, other): """Returns the concatenation of the dimension in `self` and `other`. *N.B.* If either `self` or `other` is completely unknown, concatenation will discard information about the other shape. In future, we might support concatenation that preserves this information for use with slicing. Args: other: Another `TensorShape`. Returns: A `TensorShape` whose dimensions are the concatenation of the dimensions in `self` and `other`. """ # TODO(mrry): Handle the case where we concatenate a known shape with a # completely unknown shape, so that we can use the partial information. other = as_shape(other) if self._dims is None or other.dims is None: return unknown_shape() else: return TensorShape(self._dims + other.dims) def assert_same_rank(self, other): """Raises an exception if `self` and `other` do not have compatible ranks. Args: other: Another `TensorShape`. Raises: ValueError: If `self` and `other` do not represent shapes with the same rank. """ other = as_shape(other) if self.ndims is not None and other.ndims is not None: if self.ndims != other.ndims: raise ValueError( "Shapes %s and %s must have the same rank" % (self, other)) def assert_has_rank(self, rank): """Raises an exception if `self` is not compatible with the given `rank`. Args: rank: An integer. Raises: ValueError: If `self` does not represent a shape with the given `rank`. """ if self.ndims not in (None, rank): raise ValueError("Shape %s must have rank %d" % (self, rank)) def with_rank(self, rank): """Returns a shape based on `self` with the given rank. This method promotes a completely unknown shape to one with a known rank. Args: rank: An integer. Returns: A shape that is at least as specific as `self` with the given rank. Raises: ValueError: If `self` does not represent a shape with the given `rank`. """ try: return self.merge_with(unknown_shape(ndims=rank)) except ValueError: raise ValueError("Shape %s must have rank %d" % (self, rank)) def with_rank_at_least(self, rank): """Returns a shape based on `self` with at least the given rank. Args: rank: An integer. Returns: A shape that is at least as specific as `self` with at least the given rank. Raises: ValueError: If `self` does not represent a shape with at least the given `rank`. """ if self.ndims is not None and self.ndims < rank: raise ValueError("Shape %s must have rank at least %d" % (self, rank)) else: return self def with_rank_at_most(self, rank): """Returns a shape based on `self` with at most the given rank. Args: rank: An integer. Returns: A shape that is at least as specific as `self` with at most the given rank. Raises: ValueError: If `self` does not represent a shape with at most the given `rank`. """ if self.ndims is not None and self.ndims > rank: raise ValueError("Shape %s must have rank at most %d" % (self, rank)) else: return self def is_compatible_with(self, other): """Returns True iff `self` is compatible with `other`. Two possibly-partially-defined shapes are compatible if there exists a fully-defined shape that both shapes can represent. Thus, compatibility allows the shape inference code to reason about partially-defined shapes. For example: * TensorShape(None) is compatible with all shapes. * TensorShape([None, None]) is compatible with all two-dimensional shapes, such as TensorShape([32, 784]), and also TensorShape(None). It is not compatible with, for example, TensorShape([None]) or TensorShape([None, None, None]). * TensorShape([32, None]) is compatible with all two-dimensional shapes with size 32 in the 0th dimension, and also TensorShape([None, None]) and TensorShape(None). It is not compatible with, for example, TensorShape([32]), TensorShape([32, None, 1]) or TensorShape([64, None]). * TensorShape([32, 784]) is compatible with itself, and also TensorShape([32, None]), TensorShape([None, 784]), TensorShape([None, None]) and TensorShape(None). It is not compatible with, for example, TensorShape([32, 1, 784]) or TensorShape([None]). The compatibility relation is reflexive and symmetric, but not transitive. For example, TensorShape([32, 784]) is compatible with TensorShape(None), and TensorShape(None) is compatible with TensorShape([4, 4]), but TensorShape([32, 784]) is not compatible with TensorShape([4, 4]). Args: other: Another TensorShape. Returns: True iff `self` is compatible with `other`. """ other = as_shape(other) if self._dims is not None and other.dims is not None: if self.ndims != other.ndims: return False for x_dim, y_dim in zip(self._dims, other.dims): if not x_dim.is_compatible_with(y_dim): return False return True def assert_is_compatible_with(self, other): """Raises exception if `self` and `other` do not represent the same shape. This method can be used to assert that there exists a shape that both `self` and `other` represent. Args: other: Another TensorShape. Raises: ValueError: If `self` and `other` do not represent the same shape. """ if not self.is_compatible_with(other): raise ValueError("Shapes %s and %s are incompatible" % (self, other)) def is_fully_defined(self): """Returns True iff `self` is fully defined in every dimension.""" return (self._dims is not None and all(dim.value is not None for dim in self._dims)) def assert_is_fully_defined(self): """Raises an exception if `self` is not fully defined in every dimension. Raises: ValueError: If `self` does not have a known value for every dimension. """ if not self.is_fully_defined(): raise ValueError("Shape %s is not fully defined" % self) def as_list(self): """Returns a list of integers or None for each dimension. Returns: A list of integers or None for each dimension. """ return [dim.value for dim in self._dims] def as_proto(self): """Returns this shape as a `TensorShapeProto`.""" if self._dims is None: return tensor_shape_pb2.TensorShapeProto(unknown_rank=True) else: return tensor_shape_pb2.TensorShapeProto(dim=[ tensor_shape_pb2.TensorShapeProto.Dim( size=-1 if d.value is None else d.value) for d in self._dims]) def __eq__(self, other): """Returns True if `self` is equivalent to `other`.""" try: other = as_shape(other) except TypeError: return NotImplemented return self._dims == other.dims def __ne__(self, other): """Returns True if `self` is known to be different from `other`.""" try: other = as_shape(other) except TypeError: return NotImplemented if self.ndims is None or other.ndims is None: raise ValueError("The inequality of unknown TensorShapes is undefined.") if self.ndims != other.ndims: return True return self._dims != other.dims def as_shape(shape): """Converts the given object to a TensorShape.""" if isinstance(shape, TensorShape): return shape else: return TensorShape(shape) def unknown_shape(ndims=None): """Returns an unknown TensorShape, optionally with a known rank. Args: ndims: (Optional) If specified, the number of dimensions in the shape. Returns: An unknown TensorShape. """ if ndims is None: return TensorShape(None) else: return TensorShape([Dimension(None)] * ndims) def scalar(): """Returns a shape representing a scalar.""" return TensorShape([]) def vector(length): """Returns a shape representing a vector. Args: length: The length of the vector, which may be None if unknown. Returns: A TensorShape representing a vector of the given length. """ return TensorShape([length]) def matrix(rows, cols): """Returns a shape representing a matrix. Args: rows: The number of rows in the matrix, which may be None if unknown. cols: The number of columns in the matrix, which may be None if unknown. Returns: A TensorShape representing a matrix of the given size. """ return TensorShape([rows, cols])
tommyanthony/tablib
refs/heads/develop
tablib/formats/_xlsx.py
8
# -*- coding: utf-8 -*- """ Tablib - XLSX Support. """ import sys if sys.version_info[0] > 2: from io import BytesIO else: from cStringIO import StringIO as BytesIO from tablib.compat import openpyxl import tablib Workbook = openpyxl.workbook.Workbook ExcelWriter = openpyxl.writer.excel.ExcelWriter get_column_letter = openpyxl.cell.get_column_letter from tablib.compat import unicode title = 'xlsx' extensions = ('xlsx',) def detect(stream): """Returns True if given stream is a readable excel file.""" try: openpyxl.reader.excel.load_workbook(stream) return True except openpyxl.shared.exc.InvalidFileException: pass def export_set(dataset): """Returns XLSX representation of Dataset.""" wb = Workbook() ws = wb.worksheets[0] ws.title = dataset.title if dataset.title else 'Tablib Dataset' dset_sheet(dataset, ws) stream = BytesIO() wb.save(stream) return stream.getvalue() def export_book(databook): """Returns XLSX representation of DataBook.""" wb = Workbook() wb.worksheets = [] for i, dset in enumerate(databook._datasets): ws = wb.create_sheet() ws.title = dset.title if dset.title else 'Sheet%s' % (i) dset_sheet(dset, ws) stream = BytesIO() wb.save(stream) return stream.getvalue() def import_set(dset, in_stream, headers=True): """Returns databook from XLS stream.""" dset.wipe() xls_book = openpyxl.reader.excel.load_workbook(in_stream) sheet = xls_book.get_active_sheet() dset.title = sheet.title for i, row in enumerate(sheet.rows): row_vals = [c.value for c in row] if (i == 0) and (headers): dset.headers = row_vals else: dset.append(row_vals) def import_book(dbook, in_stream, headers=True): """Returns databook from XLS stream.""" dbook.wipe() xls_book = openpyxl.reader.excel.load_workbook(in_stream) for sheet in xls_book.worksheets: data = tablib.Dataset() data.title = sheet.title for i, row in enumerate(sheet.rows): row_vals = [c.value for c in row] if (i == 0) and (headers): data.headers = row_vals else: data.append(row_vals) dbook.add_sheet(data) def dset_sheet(dataset, ws): """Completes given worksheet from given Dataset.""" _package = dataset._package(dicts=False) for i, sep in enumerate(dataset._separators): _offset = i _package.insert((sep[0] + _offset), (sep[1],)) for i, row in enumerate(_package): row_number = i + 1 for j, col in enumerate(row): col_idx = get_column_letter(j + 1) # bold headers if (row_number == 1) and dataset.headers: # ws.cell('%s%s'%(col_idx, row_number)).value = unicode( # '%s' % col, errors='ignore') ws.cell('%s%s'%(col_idx, row_number)).value = unicode(col) style = ws.get_style('%s%s' % (col_idx, row_number)) style.font.bold = True ws.freeze_panes = 'A2' # bold separators elif len(row) < dataset.width: ws.cell('%s%s'%(col_idx, row_number)).value = unicode( '%s' % col, errors='ignore') style = ws.get_style('%s%s' % (col_idx, row_number)) style.font.bold = True # wrap the rest else: try: if '\n' in col: ws.cell('%s%s'%(col_idx, row_number)).value = unicode( '%s' % col, errors='ignore') style = ws.get_style('%s%s' % (col_idx, row_number)) style.alignment.wrap_text else: ws.cell('%s%s'%(col_idx, row_number)).value = unicode( '%s' % col, errors='ignore') except TypeError: ws.cell('%s%s'%(col_idx, row_number)).value = unicode(col)
tmpgit/intellij-community
refs/heads/master
python/testData/buildout/site.py
83
"""Append module search paths for third-party packages to sys.path. **************************************************************** * This module is automatically imported during initialization. * **************************************************************** In earlier versions of Python (up to 1.5a3), scripts or modules that needed to use site-specific modules would place ``import site'' somewhere near the top of their code. Because of the automatic import, this is no longer necessary (but code that does it still works). This will append site-specific paths to the module search path. On Unix (including Mac OSX), it starts with sys.prefix and sys.exec_prefix (if different) and appends lib/python<version>/site-packages as well as lib/site-python. On other platforms (such as Windows), it tries each of the prefixes directly, as well as with lib/site-packages appended. The resulting directories, if they exist, are appended to sys.path, and also inspected for path configuration files. A path configuration file is a file whose name has the form <package>.pth; its contents are additional directories (one per line) to be added to sys.path. Non-existing directories (or non-directories) are never added to sys.path; no directory is added to sys.path more than once. Blank lines and lines beginning with '#' are skipped. Lines starting with 'import' are executed. For example, suppose sys.prefix and sys.exec_prefix are set to /usr/local and there is a directory /usr/local/lib/python2.5/site-packages with three subdirectories, foo, bar and spam, and two path configuration files, foo.pth and bar.pth. Assume foo.pth contains the following: # foo package configuration foo bar bletch and bar.pth contains: # bar package configuration bar Then the following directories are added to sys.path, in this order: /usr/local/lib/python2.5/site-packages/bar /usr/local/lib/python2.5/site-packages/foo Note that bletch is omitted because it doesn't exist; bar precedes foo because bar.pth comes alphabetically before foo.pth; and spam is omitted because it is not mentioned in either path configuration file. After these path manipulations, an attempt is made to import a module named sitecustomize, which can perform arbitrary additional site-specific customizations. If this import fails with an ImportError exception, it is silently ignored. """ import sys import os import __builtin__ # Prefixes for site-packages; add additional prefixes like /usr/local here PREFIXES = [sys.prefix, sys.exec_prefix] # Enable per user site-packages directory # set it to False to disable the feature or True to force the feature ENABLE_USER_SITE = False # buildout does not support user sites. # for distutils.commands.install USER_SITE = None USER_BASE = None def makepath(*paths): dir = os.path.abspath(os.path.join(*paths)) return dir, os.path.normcase(dir) def abs__file__(): """Set all module' __file__ attribute to an absolute path""" for m in sys.modules.values(): if hasattr(m, '__loader__'): continue # don't mess with a PEP 302-supplied __file__ try: m.__file__ = os.path.abspath(m.__file__) except AttributeError: continue def removeduppaths(): """ Remove duplicate entries from sys.path along with making them absolute""" # This ensures that the initial path provided by the interpreter contains # only absolute pathnames, even if we're running from the build directory. L = [] known_paths = set() for dir in sys.path: # Filter out duplicate paths (on case-insensitive file systems also # if they only differ in case); turn relative paths into absolute # paths. dir, dircase = makepath(dir) if not dircase in known_paths: L.append(dir) known_paths.add(dircase) sys.path[:] = L return known_paths # XXX This should not be part of site.py, since it is needed even when # using the -S option for Python. See http://www.python.org/sf/586680 def addbuilddir(): """Append ./build/lib.<platform> in case we're running in the build dir (especially for Guido :-)""" from distutils.util import get_platform s = "build/lib.%s-%.3s" % (get_platform(), sys.version) if hasattr(sys, 'gettotalrefcount'): s += '-pydebug' s = os.path.join(os.path.dirname(sys.path[-1]), s) sys.path.append(s) def _init_pathinfo(): """Return a set containing all existing directory entries from sys.path""" d = set() for dir in sys.path: try: if os.path.isdir(dir): dir, dircase = makepath(dir) d.add(dircase) except TypeError: continue return d def addpackage(sitedir, name, known_paths): """Process a .pth file within the site-packages directory: For each line in the file, either combine it with sitedir to a path and add that to known_paths, or execute it if it starts with 'import '. """ if known_paths is None: _init_pathinfo() reset = 1 else: reset = 0 fullname = os.path.join(sitedir, name) try: f = open(fullname, "rU") except IOError: return with f: for line in f: if line.startswith("#"): continue if line.startswith(("import ", "import\t")): exec line continue line = line.rstrip() dir, dircase = makepath(sitedir, line) if not dircase in known_paths and os.path.exists(dir): sys.path.append(dir) known_paths.add(dircase) if reset: known_paths = None return known_paths def addsitedir(sitedir, known_paths=None): """Add 'sitedir' argument to sys.path if missing and handle .pth files in 'sitedir'""" if known_paths is None: known_paths = _init_pathinfo() reset = 1 else: reset = 0 sitedir, sitedircase = makepath(sitedir) if not sitedircase in known_paths: sys.path.append(sitedir) # Add path component try: names = os.listdir(sitedir) except os.error: return dotpth = os.extsep + "pth" names = [name for name in names if name.endswith(dotpth)] for name in sorted(names): addpackage(sitedir, name, known_paths) if reset: known_paths = None return known_paths def check_enableusersite(): """Check if user site directory is safe for inclusion The function tests for the command line flag (including environment var), process uid/gid equal to effective uid/gid. None: Disabled for security reasons False: Disabled by user (command line option) True: Safe and enabled """ if sys.flags.no_user_site: return False if hasattr(os, "getuid") and hasattr(os, "geteuid"): # check process uid == effective uid if os.geteuid() != os.getuid(): return None if hasattr(os, "getgid") and hasattr(os, "getegid"): # check process gid == effective gid if os.getegid() != os.getgid(): return None return True def addusersitepackages(known_paths): """Add a per user site-package to sys.path Each user has its own python directory with site-packages in the home directory. USER_BASE is the root directory for all Python versions USER_SITE is the user specific site-packages directory USER_SITE/.. can be used for data. """ global USER_BASE, USER_SITE, ENABLE_USER_SITE env_base = os.environ.get("PYTHONUSERBASE", None) def joinuser(*args): return os.path.expanduser(os.path.join(*args)) #if sys.platform in ('os2emx', 'riscos'): # # Don't know what to put here # USER_BASE = '' # USER_SITE = '' if os.name == "nt": base = os.environ.get("APPDATA") or "~" USER_BASE = env_base if env_base else joinuser(base, "Python") USER_SITE = os.path.join(USER_BASE, "Python" + sys.version[0] + sys.version[2], "site-packages") else: USER_BASE = env_base if env_base else joinuser("~", ".local") USER_SITE = os.path.join(USER_BASE, "lib", "python" + sys.version[:3], "site-packages") if ENABLE_USER_SITE and os.path.isdir(USER_SITE): addsitedir(USER_SITE, known_paths) return known_paths def addsitepackages(known_paths): """Add site packages, as determined by zc.buildout. See original_addsitepackages, below, for the original version.""" setuptools_path = 'c:\\src\\django\\buildout15\\eggs\\setuptools-0.6c12dev_r88124-py2.6.egg' sys.path.append(setuptools_path) known_paths.add(os.path.normcase(setuptools_path)) import pkg_resources buildout_paths = [ 'c:\\src\\django\\buildout15\\src', 'c:\\src\\django\\buildout15\\eggs\\setuptools-0.6c12dev_r88124-py2.6.egg' ] for path in buildout_paths: sitedir, sitedircase = makepath(path) if not sitedircase in known_paths and os.path.exists(sitedir): sys.path.append(sitedir) known_paths.add(sitedircase) pkg_resources.working_set.add_entry(sitedir) sys.__egginsert = len(buildout_paths) # Support distribute. original_paths = [ 'C:\\Python26\\lib\\site-packages' ] for path in original_paths: if path == setuptools_path or path not in known_paths: addsitedir(path, known_paths) return known_paths def original_addsitepackages(known_paths): """Add site-packages (and possibly site-python) to sys.path""" sitedirs = [] seen = [] for prefix in PREFIXES: if not prefix or prefix in seen: continue seen.append(prefix) if sys.platform in ('os2emx', 'riscos'): sitedirs.append(os.path.join(prefix, "Lib", "site-packages")) elif os.sep == '/': sitedirs.append(os.path.join(prefix, "lib", "python" + sys.version[:3], "site-packages")) sitedirs.append(os.path.join(prefix, "lib", "site-python")) else: sitedirs.append(prefix) sitedirs.append(os.path.join(prefix, "lib", "site-packages")) if sys.platform == "darwin": # for framework builds *only* we add the standard Apple # locations. Currently only per-user, but /Library and # /Network/Library could be added too if 'Python.framework' in prefix: sitedirs.append( os.path.expanduser( os.path.join("~", "Library", "Python", sys.version[:3], "site-packages"))) for sitedir in sitedirs: if os.path.isdir(sitedir): addsitedir(sitedir, known_paths) return known_paths def setBEGINLIBPATH(): """The OS/2 EMX port has optional extension modules that do double duty as DLLs (and must use the .DLL file extension) for other extensions. The library search path needs to be amended so these will be found during module import. Use BEGINLIBPATH so that these are at the start of the library search path. """ dllpath = os.path.join(sys.prefix, "Lib", "lib-dynload") libpath = os.environ['BEGINLIBPATH'].split(';') if libpath[-1]: libpath.append(dllpath) else: libpath[-1] = dllpath os.environ['BEGINLIBPATH'] = ';'.join(libpath) def setquit(): """Define new built-ins 'quit' and 'exit'. These are simply strings that display a hint on how to exit. """ if os.sep == ':': eof = 'Cmd-Q' elif os.sep == '\\': eof = 'Ctrl-Z plus Return' else: eof = 'Ctrl-D (i.e. EOF)' class Quitter(object): def __init__(self, name): self.name = name def __repr__(self): return 'Use %s() or %s to exit' % (self.name, eof) def __call__(self, code=None): # Shells like IDLE catch the SystemExit, but listen when their # stdin wrapper is closed. try: sys.stdin.close() except: pass raise SystemExit(code) __builtin__.quit = Quitter('quit') __builtin__.exit = Quitter('exit') class _Printer(object): """interactive prompt objects for printing the license text, a list of contributors and the copyright notice.""" MAXLINES = 23 def __init__(self, name, data, files=(), dirs=()): self.__name = name self.__data = data self.__files = files self.__dirs = dirs self.__lines = None def __setup(self): if self.__lines: return data = None for dir in self.__dirs: for filename in self.__files: filename = os.path.join(dir, filename) try: fp = file(filename, "rU") data = fp.read() fp.close() break except IOError: pass if data: break if not data: data = self.__data self.__lines = data.split('\n') self.__linecnt = len(self.__lines) def __repr__(self): self.__setup() if len(self.__lines) <= self.MAXLINES: return "\n".join(self.__lines) else: return "Type %s() to see the full %s text" % ((self.__name,)*2) def __call__(self): self.__setup() prompt = 'Hit Return for more, or q (and Return) to quit: ' lineno = 0 while 1: try: for i in range(lineno, lineno + self.MAXLINES): print self.__lines[i] except IndexError: break else: lineno += self.MAXLINES key = None while key is None: key = raw_input(prompt) if key not in ('', 'q'): key = None if key == 'q': break def setcopyright(): """Set 'copyright' and 'credits' in __builtin__""" __builtin__.copyright = _Printer("copyright", sys.copyright) if sys.platform[:4] == 'java': __builtin__.credits = _Printer( "credits", "Jython is maintained by the Jython developers (www.jython.org).") else: __builtin__.credits = _Printer("credits", """\ Thanks to CWI, CNRI, BeOpen.com, Zope Corporation and a cast of thousands for supporting Python development. See www.python.org for more information.""") here = os.path.dirname(os.__file__) __builtin__.license = _Printer( "license", "See http://www.python.org/%.3s/license.html" % sys.version, ["LICENSE.txt", "LICENSE"], [os.path.join(here, os.pardir), here, os.curdir]) class _Helper(object): """Define the built-in 'help'. This is a wrapper around pydoc.help (with a twist). """ def __repr__(self): return "Type help() for interactive help, " \ "or help(object) for help about object." def __call__(self, *args, **kwds): import pydoc return pydoc.help(*args, **kwds) def sethelper(): __builtin__.help = _Helper() def aliasmbcs(): """On Windows, some default encodings are not provided by Python, while they are always available as "mbcs" in each locale. Make them usable by aliasing to "mbcs" in such a case.""" if sys.platform == 'win32': import locale, codecs enc = locale.getdefaultlocale()[1] if enc.startswith('cp'): # "cp***" ? try: codecs.lookup(enc) except LookupError: import encodings encodings._cache[enc] = encodings._unknown encodings.aliases.aliases[enc] = 'mbcs' def setencoding(): """Set the string encoding used by the Unicode implementation. The default is 'ascii', but if you're willing to experiment, you can change this.""" encoding = "ascii" # Default value set by _PyUnicode_Init() if 0: # Enable to support locale aware default string encodings. import locale loc = locale.getdefaultlocale() if loc[1]: encoding = loc[1] if 0: # Enable to switch off string to Unicode coercion and implicit # Unicode to string conversion. encoding = "undefined" if encoding != "ascii": # On Non-Unicode builds this will raise an AttributeError... sys.setdefaultencoding(encoding) # Needs Python Unicode build ! def execsitecustomize(): """Run custom site specific code, if available.""" try: import sitecustomize except ImportError: pass except Exception: if sys.flags.verbose: sys.excepthook(*sys.exc_info()) else: print >>sys.stderr, \ "'import sitecustomize' failed; use -v for traceback" def execusercustomize(): """Run custom user specific code, if available.""" try: import usercustomize except ImportError: pass except Exception: if sys.flags.verbose: sys.excepthook(*sys.exc_info()) else: print>>sys.stderr, \ "'import usercustomize' failed; use -v for traceback" def main(): global ENABLE_USER_SITE abs__file__() known_paths = removeduppaths() if (os.name == "posix" and sys.path and os.path.basename(sys.path[-1]) == "Modules"): addbuilddir() if ENABLE_USER_SITE is None: ENABLE_USER_SITE = check_enableusersite() known_paths = addusersitepackages(known_paths) known_paths = addsitepackages(known_paths) if sys.platform == 'os2emx': setBEGINLIBPATH() setquit() setcopyright() sethelper() aliasmbcs() setencoding() execsitecustomize() if ENABLE_USER_SITE: execusercustomize() # Remove sys.setdefaultencoding() so that users cannot change the # encoding after initialization. The test for presence is needed when # this module is run as a script, because this code is executed twice. if hasattr(sys, "setdefaultencoding"): del sys.setdefaultencoding main() def _script(): help = """\ %s [--user-base] [--user-site] Without arguments print some useful information With arguments print the value of USER_BASE and/or USER_SITE separated by '%s'. Exit codes with --user-base or --user-site: 0 - user site directory is enabled 1 - user site directory is disabled by user 2 - uses site directory is disabled by super user or for security reasons >2 - unknown error """ args = sys.argv[1:] if not args: print "sys.path = [" for dir in sys.path: print " %r," % (dir,) print "]" print "USER_BASE: %r (%s)" % (USER_BASE, "exists" if os.path.isdir(USER_BASE) else "doesn't exist") print "USER_SITE: %r (%s)" % (USER_SITE, "exists" if os.path.isdir(USER_SITE) else "doesn't exist") print "ENABLE_USER_SITE: %r" % ENABLE_USER_SITE sys.exit(0) buffer = [] if '--user-base' in args: buffer.append(USER_BASE) if '--user-site' in args: buffer.append(USER_SITE) if buffer: print os.pathsep.join(buffer) if ENABLE_USER_SITE: sys.exit(0) elif ENABLE_USER_SITE is False: sys.exit(1) elif ENABLE_USER_SITE is None: sys.exit(2) else: sys.exit(3) else: import textwrap print textwrap.dedent(help % (sys.argv[0], os.pathsep)) sys.exit(10) if __name__ == '__main__': _script()
ffantast/magnum
refs/heads/master
magnum/objects/__init__.py
6
# Copyright 2013 IBM Corp. # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. from magnum.objects import bay from magnum.objects import baylock from magnum.objects import baymodel from magnum.objects import container from magnum.objects import node from magnum.objects import pod from magnum.objects import replicationcontroller as rc from magnum.objects import service from magnum.objects import x509keypair Container = container.Container Bay = bay.Bay BayLock = baylock.BayLock BayModel = baymodel.BayModel Node = node.Node Pod = pod.Pod ReplicationController = rc.ReplicationController Service = service.Service X509KeyPair = x509keypair.X509KeyPair __all__ = (Bay, BayLock, BayModel, Container, Node, Pod, ReplicationController, Service, X509KeyPair)
CompPhysics/ComputationalPhysicsMSU
refs/heads/master
doc/Programs/LecturePrograms/programs/NumericalIntegration/python/program1.py
3
# -*- coding: utf-8 -*- #Example of numerical integration with Gauss-Legendre quadrature #Translated to Python by Kyrre Ness Sjøbæk import sys import numpy from computationalLib import pylib #Read input if len(sys.argv) == 1: print "Number of integration points:" n = int(sys.stdin.readline()) print "Integration limits (a, b)" (a,b) = sys.stdin.readline().split(",") a = int(a) b = int(b) elif len(sys.argv) == 4: n = int(sys.argv[1]) a = int(sys.argv[2]) b = int(sys.argv[3]) else: print "Usage: python", sys.argv[0], "N a b" sys.exit(0) #Definitions m = pylib(inputcheck=False,cpp=False) def integrand(x): return 4./(1. + x*x) #Integrate with Gaus-legendre! (x,w) = m.gausLegendre(a,b,n) int_gaus = numpy.sum(w*integrand(x)) #Final output print "integration of f(x) 4/(1+x**2) from",a,"to",b,"with",n,"meshpoints:" print "Gaus-legendre:", int_gaus print "Trapezoidal: ", m.trapezoidal(a,b,n, integrand) print "Simpson: ", m.simpson(a,b,n,integrand)
ryanss/python-holidays
refs/heads/master
holidays/countries/india.py
2
# -*- coding: utf-8 -*- # python-holidays # --------------- # A fast, efficient Python library for generating country, province and state # specific sets of holidays on the fly. It aims to make determining whether a # specific date is a holiday as fast and flexible as possible. # # Author: ryanss <ryanssdev@icloud.com> (c) 2014-2017 # dr-prodigy <maurizio.montel@gmail.com> (c) 2017-2020 # Website: https://github.com/dr-prodigy/python-holidays # License: MIT (see LICENSE file) from datetime import date from holidays.constants import JAN, MAR, APR, MAY, JUN, AUG, OCT, \ NOV, DEC from holidays.holiday_base import HolidayBase class India(HolidayBase): # https://en.wikipedia.org/wiki/Public_holidays_in_India # https://www.calendarlabs.com/holidays/india/ # https://slusi.dacnet.nic.in/watershedatlas/list_of_state_abbreviation.htm PROVINCES = ['AS', 'CG', 'SK', 'KA', 'GJ', 'BR', 'RJ', 'OD', 'TN', 'AP', 'WB', 'KL', 'HR', 'MH', 'MP', 'UP', 'UK', 'TS'] def __init__(self, **kwargs): self.country = "IND" HolidayBase.__init__(self, **kwargs) def _populate(self, year): # Pongal/ Makar Sankranti self[date(year, JAN, 14)] = "Makar Sankranti / Pongal" if year >= 1950: # Republic Day self[date(year, JAN, 26)] = "Republic Day" if year >= 1947: # Independence Day self[date(year, AUG, 15)] = "Independence Day" # Gandhi Jayanti self[date(year, OCT, 2)] = "Gandhi Jayanti" # Labour Day self[date(year, MAY, 1)] = "Labour Day" # Christmas self[date(year, DEC, 25)] = "Christmas" # GJ: Gujarat if self.prov == "GJ": self[date(year, JAN, 14)] = "Uttarayan" self[date(year, MAY, 1)] = "Gujarat Day" self[date(year, OCT, 31)] = "Sardar Patel Jayanti" if self.prov == 'BR': self[date(year, MAR, 22)] = "Bihar Day" if self.prov == 'RJ': self[date(year, MAR, 30)] = "Rajasthan Day" self[date(year, JUN, 15)] = "Maharana Pratap Jayanti" if self.prov == 'OD': self[date(year, APR, 1)] = "Odisha Day (Utkala Dibasa)" self[date(year, APR, 15)] = "Maha Vishuva Sankranti / Pana" \ " Sankranti" if self.prov in ('OD', 'AP', 'BR', 'WB', 'KL', 'HR', 'MH', 'UP', 'UK', 'TN'): self[date(year, APR, 14)] = "Dr. B. R. Ambedkar's Jayanti" if self.prov == 'TN': self[date(year, APR, 14)] = "Puthandu (Tamil New Year)" self[date(year, APR, 15)] = "Puthandu (Tamil New Year)" if self.prov == 'WB': self[date(year, APR, 14)] = "Pohela Boishakh" self[date(year, APR, 15)] = "Pohela Boishakh" self[date(year, MAY, 9)] = "Rabindra Jayanti" if self.prov == 'AS': self[date(year, APR, 15)] = "Bihu (Assamese New Year)" if self.prov == 'MH': self[date(year, MAY, 1)] = "Maharashtra Day" if self.prov == 'SK': self[date(year, MAY, 16)] = "Annexation Day" if self.prov == 'KA': self[date(year, NOV, 1)] = "Karnataka Rajyotsava" if self.prov == 'AP': self[date(year, NOV, 1)] = "Andhra Pradesh Foundation Day" if self.prov == 'HR': self[date(year, NOV, 1)] = "Haryana Foundation Day" if self.prov == 'MP': self[date(year, NOV, 1)] = "Madhya Pradesh Foundation Day" if self.prov == 'KL': self[date(year, NOV, 1)] = "Kerala Foundation Day" if self.prov == 'CG': self[date(year, NOV, 1)] = "Chhattisgarh Foundation Day" # TS is Telangana State which was bifurcated in 2014 from AP # (AndhraPradesh) if self.prov == 'TS': self[date(year, OCT, 6)] = "Bathukamma Festival" self[date(year, APR, 6)] = "Eid al-Fitr" class IN(India): pass class IND(India): pass
auready/django
refs/heads/master
tests/gis_tests/geoadmin/models.py
21
from django.contrib.gis.db import models from ..admin import admin class City(models.Model): name = models.CharField(max_length=30) point = models.PointField() class Meta: app_label = 'geoadmin' required_db_features = ['gis_enabled'] def __str__(self): return self.name site = admin.AdminSite(name='admin_gis') site.register(City, admin.OSMGeoAdmin)
daviddoria/PointGraphsPhase1
refs/heads/PointGraphsPhase1
Examples/Infovis/Python/treering_view_simple.py
17
from vtk import * reader1 = vtkXMLTreeReader() reader1.SetFileName("treetest.xml") reader1.Update() view = vtkTreeRingView() view.SetRepresentationFromInput(reader1.GetOutput()) view.SetAreaSizeArrayName("size") view.SetAreaColorArrayName("level") view.SetAreaLabelArrayName("name") view.SetAreaLabelVisibility(True) view.SetAreaHoverArrayName("name") view.SetShrinkPercentage(0.05) view.Update() # Apply a theme to the views theme = vtkViewTheme.CreateMellowTheme() view.ApplyViewTheme(theme) theme.FastDelete() view.ResetCamera() view.Render() view.GetInteractor().Start()
dnaextrim/django_adminlte_x
refs/heads/master
adminlte/static/plugins/datatables/extensions/Responsive/examples/display-control/classes.html.py
1
XXXXXXXXX XXXXX XXXXXX XXXXXX XXXXX XXXXXXXXXXXXXXXX XXXXX XXXXXXXXXXXXX XXXXX XXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXX XXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXX XXXXXXX X XXXXX XXXXXXXXXXXXXXX XXXXX XXXXXXXXXXXXXXXX XXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXX XXXXXXXXXXXXXXXX XXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXX XXXXXXXXXXXXXXXX XXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXX XXXXXXXXXXXXXXXX XXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXX XXXXXXXXXXXXXXX XXXXXXXXXXXXX XXXXXXXX XXXXXXX XXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXX XXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXX XXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXX XXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXX XXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXX XXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXX X XXXXXXXXXXXXXXXXXXXXXXXX X XXXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXX X X XXXXXXX XXXXXX XX X XXXXXXX XXXXXXXXXX XX X XXXXXXX XXXXXXXX XX X XXXXXXX XXXXX XX X XXXXXXX XXXXXXXXXXXX XX X XXXXXXX XXXXXXXX XX X XXXXXXX XXXXXX X X X XX X XX XXXXXXXXX XXXXXXX XXXXX XXXXXXXXXXXXXXXXXXX XXXX XXXXXXXXXXXXXXXXXX XXXXXXXXX XXXXXXXXXXXXXX XXXXXXX XXXXXXXXXXX XXXXXXXXXXXXXXXXXXX XXXX XXXXXXXXXXXXX XXXXXX XXX XXXX XXXXXXXXXX XXXX XXXXXXX XX XXXX XX XX XXXXXXX XX XXXXXXXXX XXXXXXX XXXXXXX XXX XXX XX XXXXX XXXXX XX XXX XXXXXXXX XXX XXXXXXXXXXX XXX XXXXXXXXXX XXXXXX XXXXXXXXXXX XXX XXX XXXXXXXX XXX XXX XXX XXXXXX XXXXXXXXXXXX XXX XXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXXXXXX X XXXXX XXXXXXXXXXX XXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXX XXX XXXXX X XXXX XXXXXXXXX XXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXX XXX XXXXX X XXXX XXXXXXXX XXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXX XXX XXXXX X XXXX XXXXXXXX XXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXX X XXXX XXXXXXXX XXXXX XXXXXX XXX XXXXX XXX XXXXXXXXXXXXXXXXXX XXXXXX XXXX XXX XXX XX XXX XXXX XX XXXX XXXXXX XXX XXXXXX XX XXXXXX XXXXXXXX XXXXXXXXXXXX XX XXX XXX XXXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXX XX XXXXXXXXXXXXXXXXX XX X XXXXXX XX XXX XXXXX XXXX XX XXXXXXX XXXXX XXXXXXXXXXX XXX XXXXXXX XXXXXXXXXXXXXXXXXXXXXXX XXXXX XXXXX X XXXXXX XX XXXXXX XX XXXXXXX XX XXXXXXX XXX XXXXXX XXXXXXXX XXXXX XXXXXXXXXXXXXXXXXXXXXXXXX XXXXX XXXXXXX XX XXXXX X XXXXXXXXXX XXXXX XX XXX XXX XXX XXXXXXX XXXXXX XX XX XXXXXX XXX XX XXXXX XX XXX XXXXX XXXXXXXXXXX XXXXXXXXXXXXXXXX XXXXX XXX XXXXX XXXXXXX XXXXX XXXXXXXXXX XXX XXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXX X XXXXXX XXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXXX X XXXXX XXXXXXX XX X XXXXXXX XXX XXXX XX XXX XXXXX XXXXXXXX XXXXXXXXXXXXXXXXXXXXXX X XXXXX XXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXXXXXX X XXXX XXX XXX XXXXXXXXXXXXXXXXXXX XX XXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXXX XXXXXX XXXXXXXXXXX XXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXX XXXXX XXXXXXXXX XX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XX XXX XXXXXXXXXX XXXXXXXXXX XXX XXXXXXX XXXXXXX XX XXXXX XXXXXXXXXXXX XXXXXXX XXXXXXX XXXXX XXX XXXXXXXXXXXXXXXXXXX XXXXXX XXXXXXX XX X XXXXXXX XXXX X XXXXXXXXXXXXXXXXXXX XXX XXXXXXXXXXXXXXXX XXXXXXX X XXXXXXX XXXXX XXX XXXXXXXXXXXXXXXXXXXXX XXXXXX XXXXXXXX X XXXXX XX XXXXXXXXX XX XXXXXXX XXX XXXXXXXXXXXXXXXXX XXXXXX XX XXXXXX XXXXXXX XXX XXX XXXXXXXXXXX XXXXXXXXXXX XX XXXXX XXXXXXXXXXXX XXXXXXX XXX XX XXXXXX XX XXX XXXX XX XXXXXX XXX XXXXXX XX XXX XXXX XXXXX XX XXXXXXXXX XXXXXXXX XXX XXXXXXX XXXXX X XXXXXXXXXXX XX XXXXXXXXXXXXXXXXXXX XXX XXXXXXXXXXXXXXXXXXXXXXX XX XXX XXXXXXXXX XXXXXXX XXXXX XXXXX XXXXXXXXXXX XX XX XXXXXXXXX XXXXXXXX XXX XXXX XXXXXX XXXXXXXXX XXXXXX XXXX XXXXXXXXXXXXXXXXXXXXXX XXXXXX XXXXXXXXXXXX XXXXXXXXXXXXXX XXXXXXXXXXX XXXXXXXXXXXXX XXXXXXX XXXX XXX XXXXXXXXXXXXXXXXXXXXX XXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXX XXXXXXXXXXXXXXXXXXXXXXXXXXX XXX XXXXXXXXXXXXXXXXXXX XXXXXXXXX XXX XXXXXXXXXXXXXXXXXXXXXXXXXXX XXX XXXXXXXXXXXXXXXXXXXXXXX XXXXX XXXXXXXX XXXXXXX XXXX XXXXXXXXXXXXX XXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXX XXXXXXXXXXXX XXXXXXXXX XXXXXXXXX XXXXXXXXXXXXXXX XXXXXXXXXXXXXX XXXXX XXXXXXXX XXXXXXXX XXX XXXXXXXXXXXXX XXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXX XXXXXXXXXXXX XXXXXXXXXXXXX XXXXXXXXXXXXXXX XXXXXXXXXXX XXXXX XXXX XXXXXXXXXXXXX XXXX XXXXXXXXXXX XXXXXX XXXXXXXXXX XXXXX XXXXX XX XXXX XX XXXXXXXXXX XXX XXXXX XXXXX XX XXXX XXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX X XXXXXXXXXXXXXXXXXXXXXXXX X XXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXX X X XXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXX XX X XXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXX XX X XXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXX XX X XXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXX XX X XXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXXXX XX X XXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXX XX X XXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXX X X X XX X XXXXXXXXX XXXXX XXXXXXXX XX XXX XXXXX XXXXX XXX XXXXXXXXX XXXXXXXXXX XXXXXXX XXXXX XXX XXXXXX XXX XXX XX XXXX XXXXXXXXXXXX XXXX XXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXX XXXXXX XXXX XXXXXXXXXXXXXX XXXXXX XXXX XXXXX XXXXX XX XXX XXX XXXX XXXXX XXXXXXXX XXXXXX XX XXX XXXX XXXXXXXX XX XXXXXXXXXXXXXXX XXXXXX XXXX XXXXXXXXXXXX XXXXX XXXXXXX XXXXXXX XXXX X XXXXXX XXX XX XXXXXXXXXX XXX XXXXXX XXXX XX XXXXXX XXXX XXX XXXXXXX XXXXX XXXXXXXX XX XXXXX XX XXXXXXXXX XXXXXXX XXX XXXXXX XXX XXXXXXXXXX XXX XXXX XX XXXXX XXXXXXXXXXXXXXX XXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXXX XXXXXX XXXXXX XXXXXXXXX XXX XXXXXXX XXXXX XXX XXXXXX XXX XXX XX XXXX XXXXXXX XX XXXXXXX XXX XXXXXXX XX XXX XXXXXXXXXX XXXX XXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXX XXXXXX XXXX XXXXXXXXXXXXX XXXXXXX XXXXX XXXXX XXXX XX XXXXX XXX XXXXXX XXXX XXXX XXX XXXX XXXXXX XX XXXXX XXXXXX XXXX XXXX XXXX XXXXXX XXXXXXXXXXXXX XX XXX XXXXXXXXXX XXXX XX XXXXXXXXXXX XXXXXX XXXX XXXXXXXXXXXX XXXXXX XXXXXX XXXX XX XXXXXXX XXX XXXXXXXXXXX XXXXXXXXXX XXX XXXX XXXXX XX XXXXX XXXXXX XXXXXX XXXX XXXX XXXX XX XXXX XX XXXXXXX XXXXXX XXXXX XXXX XXXXXXXXXXX XXXXXXXXXX XXXXXXX XXX XX XXXXXXX XX XXX XXXXXXXXX XXXXX XX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXXX XXXXXXXXX XX XXX XXXXXXXXXX XXXXXXXXXXXXXXXXXXXXXX XXXXXX XXXXXX XXXXXXXXXX XXXXXX XXXXXXXXX XXXX XXXXXXXXXXXXXXX XXXX XXXXXXXXXXXXXXXXXXXXXXX XXXX XXXXXXXXXXXXXX XXXXXXXXX XXXXXXXXXXXXX XXXX XXXXXXXXXXXX XXXX XXXXXXXXXXXXXXXXXX XXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXXXXX XXX XXXXXXXXXXXX XXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXX XXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXX XXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXX XXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXX XXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXXXXX XXXXX XXXXXX XXXX XXXXXXXXXXXXXXXXXX XXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXX XXXXXXXXXXXX XXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXX XXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXX XXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXX XXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXX XXXXX XXXXXX XXXX XXXXXXXXXXXXXXXXXX XXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXX XXX XXXXXXXXXX XXXXXXXX XXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXX XXXXXXXXXXXXXXX XXX XXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXX XXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXX XXXXXXXXXXXXXXXX XXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXXXXX XXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXX XXXXXXXX X XXXXXXXXXXXXXXXXX XXXXX XXXXXX XXXX XXXXXXXXXXXXXXXXXX XXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXX XXX XXXXXXXXXXXX XXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXX XXXXXXXXXXXXX XXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXX XXXXX XXXXXXXXXXXXX XXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXX X XXXXXXXXXXXXXX XXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXX XXXXX XXX XXXXXXXXXXXXXXXX XXXXXX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXX XXX XXXXXXXXXXXXXXXXX XXXXX XXXXXX XXXXXX XXXX XXXXXXXXXXXXXXXXX XXXXXXXXX XXXXX XX XXX XX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXXX XXX XXXX XXXXXXXXXXX XXXXX XXX XXX XXXXXXXXXX XXX XXXXXXXXXXXX XXXXXXXXXXXXX XXXXX XXX X XXXX XXXXX XX XX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXX XX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXX XXXXXX XXX XXXXXXXXXXXX XX XXXXXXXXXXXXXXX XX XXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXXX XXX XXXXXXX XX XX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXX XXXXXX XXXXXXXXXXXXX XXXXXXXXXX XX XXXXXXXX XXXXX XXX XX XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXX XXXXXXXXXXXXXXXX XXXXXX XXXXXX XXXXXX XXXXXXXXXX XXXXXXX XXXXXXX
ulrichard/electrum
refs/heads/master
lib/plugins.py
1
#!/usr/bin/env python # # Electrum - lightweight Bitcoin client # Copyright (C) 2015 Thomas Voegtlin # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. import traceback import sys import os import imp import pkgutil import time from util import * from i18n import _ from util import profiler, PrintError, DaemonThread class Plugins(DaemonThread): @profiler def __init__(self, config, is_local, gui_name): DaemonThread.__init__(self) if is_local: find = imp.find_module('plugins') plugins = imp.load_module('electrum_plugins', *find) else: plugins = __import__('electrum_plugins') self.pkgpath = os.path.dirname(plugins.__file__) self.plugins = {} self.gui_name = gui_name self.descriptions = [] for loader, name, ispkg in pkgutil.iter_modules([self.pkgpath]): m = loader.find_module(name).load_module(name) d = m.__dict__ if gui_name not in d.get('available_for', []): continue self.descriptions.append(d) x = d.get('registers_wallet_type') if x: self.register_wallet_type(config, name, x) if not d.get('requires_wallet_type') and config.get('use_' + name): self.load_plugin(config, name) def get(self, name): return self.plugins.get(name) def count(self): return len(self.plugins) def load_plugin(self, config, name): full_name = 'electrum_plugins.' + name + '.' + self.gui_name try: p = pkgutil.find_loader(full_name).load_module(full_name) plugin = p.Plugin(self, config, name) self.add_jobs(plugin.thread_jobs()) self.plugins[name] = plugin self.print_error("loaded", name) return plugin except Exception: print_msg(_("Error: cannot initialize plugin"), name) traceback.print_exc(file=sys.stdout) return None def close_plugin(self, plugin): self.remove_jobs(plugin.thread_jobs()) def toggle_enabled(self, config, name): p = self.get(name) config.set_key('use_' + name, p is None, True) if p: self.plugins.pop(name) p.close() self.print_error("closed", name) return None return self.load_plugin(config, name) def is_available(self, name, w): for d in self.descriptions: if d.get('__name__') == name: break else: return False deps = d.get('requires', []) for dep, s in deps: try: __import__(dep) except ImportError: return False wallet_types = d.get('requires_wallet_type') if wallet_types: if w.wallet_type not in wallet_types: return False return True def wallet_plugin_loader(self, config, name): if self.plugins.get(name) is None: self.load_plugin(config, name) return self.plugins[name] def register_wallet_type(self, config, name, x): import wallet x += (lambda: self.wallet_plugin_loader(config, name),) wallet.wallet_types.append(x) def run(self): jobs = [job for plugin in self.plugins.values() for job in plugin.thread_jobs()] self.add_jobs(jobs) while self.is_running(): time.sleep(0.1) self.run_jobs() self.print_error("stopped") hook_names = set() hooks = {} def hook(func): hook_names.add(func.func_name) return func def run_hook(name, *args): return _run_hook(name, False, *args) def always_hook(name, *args): return _run_hook(name, True, *args) def _run_hook(name, always, *args): results = [] f_list = hooks.get(name, []) for p, f in f_list: if name == 'load_wallet': p.wallet = args[0] # For for p.is_enabled() below if always or p.is_enabled(): try: r = f(*args) except Exception: print_error("Plugin error") traceback.print_exc(file=sys.stdout) r = False if r: results.append(r) if results: assert len(results) == 1, results return results[0] class BasePlugin(PrintError): def __init__(self, parent, config, name): self.parent = parent # The plugins object self.name = name self.config = config self.wallet = None # add self to hooks for k in dir(self): if k in hook_names: l = hooks.get(k, []) l.append((self, getattr(self, k))) hooks[k] = l def diagnostic_name(self): return self.name def close(self): # remove self from hooks for k in dir(self): if k in hook_names: l = hooks.get(k, []) l.remove((self, getattr(self, k))) hooks[k] = l self.parent.close_plugin(self) self.on_close() def on_close(self): pass def requires_settings(self): return False def thread_jobs(self): return [] def is_enabled(self): return self.is_available() and self.config.get('use_'+self.name) is True def is_available(self): return True def settings_dialog(self): pass
jonboiser/content-curation
refs/heads/develop
contentcuration/contentcuration/migrations/0066_auto_20170412_0015.py
4
# -*- coding: utf-8 -*- # Generated by Django 1.9.12 on 2017-04-12 07:15 from __future__ import unicode_literals from django.db import migrations, models class Migration(migrations.Migration): dependencies = [ ('contentcuration', '0065_auto_20170411_1609'), ] operations = [ migrations.AlterField( model_name='user', name='preferences', field=models.TextField(default=b'{"auto_derive_video_thumbnail": true, "auto_randomize_questions": true, "copyright_holder": null, "auto_derive_document_thumbnail": true, "license_description": null, "mastery_model": "num_correct_in_a_row_5", "m_value": 5, "license": "CC BY", "language": null, "author": null, "auto_derive_exercise_thumbnail": true, "auto_derive_html5_thumbnail": true, "auto_derive_audio_thumbnail": true, "n_value": 5}'), ), ]
af1rst/bite-project
refs/heads/master
deps/gdata-python-client/samples/apps/marketplace_sample/gdata/apps/emailsettings/data.py
23
#!/usr/bin/python # # Copyright 2010 Google Inc. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Data model classes for the Email Settings API.""" __author__ = 'Claudio Cherubino <ccherubino@google.com>' import atom.data import gdata.apps import gdata.apps_property import gdata.data # This is required to work around a naming conflict between the Google # Spreadsheets API and Python's built-in property function pyproperty = property # The apps:property label of the label property LABEL_NAME = 'label' # The apps:property from of the filter property FILTER_FROM_NAME = 'from' # The apps:property to of the filter property FILTER_TO_NAME = 'to' # The apps:property subject of the filter property FILTER_SUBJECT_NAME = 'subject' # The apps:property hasTheWord of the filter property FILTER_HAS_THE_WORD_NAME = 'hasTheWord' # The apps:property doesNotHaveTheWord of the filter property FILTER_DOES_NOT_HAVE_THE_WORD_NAME = 'doesNotHaveTheWord' # The apps:property hasAttachment of the filter property FILTER_HAS_ATTACHMENTS_NAME = 'hasAttachment' # The apps:property label of the filter action property FILTER_LABEL = 'label' # The apps:property shouldMarkAsRead of the filter action property FILTER_MARK_AS_READ = 'shouldMarkAsRead' # The apps:property shouldArchive of the filter action propertylabel FILTER_ARCHIVE = 'shouldArchive' # The apps:property name of the send-as alias property SENDAS_ALIAS_NAME = 'name' # The apps:property address of theAPPS_TEMPLATE send-as alias property SENDAS_ALIAS_ADDRESS = 'address' # The apps:property replyTo of the send-as alias property SENDAS_ALIAS_REPLY_TO = 'replyTo' # The apps:property makeDefault of the send-as alias property SENDAS_ALIAS_MAKE_DEFAULT = 'makeDefault' # The apps:property enable of the webclip property WEBCLIP_ENABLE = 'enable' # The apps:property enable of the forwarding property FORWARDING_ENABLE = 'enable' # The apps:property forwardTo of the forwarding property FORWARDING_TO = 'forwardTo' # The apps:property action of the forwarding property FORWARDING_ACTION = 'action' # The apps:property enable of the POP property POP_ENABLE = 'enable' # The apps:property enableFor of the POP propertyACTION POP_ENABLE_FOR = 'enableFor' # The apps:property action of the POP property POP_ACTION = 'action' # The apps:property enable of the IMAP property IMAP_ENABLE = 'enable' # The apps:property enable of the vacation responder property VACATION_RESPONDER_ENABLE = 'enable' # The apps:property subject of the vacation responder property VACATION_RESPONDER_SUBJECT = 'subject' # The apps:property message of the vacation responder property VACATION_RESPONDER_MESSAGE = 'message' # The apps:property startDate of the vacation responder property VACATION_RESPONDER_STARTDATE = 'startDate' # The apps:property endDate of the vacation responder property VACATION_RESPONDER_ENDDATE = 'endDate' # The apps:property contactsOnly of the vacation responder property VACATION_RESPONDER_CONTACTS_ONLY = 'contactsOnly' # The apps:property domainOnly of the vacation responder property VACATION_RESPONDER_DOMAIN_ONLY = 'domainOnly' # The apps:property signature of the signature property SIGNATURE_VALUE = 'signature' # The apps:property language of the language property LANGUAGE_TAG = 'language' # The apps:property pageSize of the general settings property GENERAL_PAGE_SIZE = 'pageSize' # The apps:property shortcuts of the general settings property GENERAL_SHORTCUTS = 'shortcuts' # The apps:property arrows of the general settings property GENERAL_ARROWS = 'arrows' # The apps:prgdata.appsoperty snippets of the general settings property GENERAL_SNIPPETS = 'snippets' # The apps:property uniAppsProcode of the general settings property GENERAL_UNICODE = 'unicode' # The apps:property delegationId of the email delegation property DELEGATION_ID = 'delegationId' # The apps:property address of the email delegation property DELEGATION_ADDRESS = 'address' # The apps:property delegate of the email delegation property DELEGATION_DELEGATE = 'delegate' # The apps:property status of the email delegation property DELEGATION_STATUS = 'status' class EmailSettingsEntry(gdata.data.GDEntry): """Represents an Email Settings entry in object form.""" property = [gdata.apps_property.AppsProperty] def _GetProperty(self, name): """Get the apps:property value with the given name. Args: name: string Name of the apps:property value to get. Returns: The apps:property value with the given name, or None if the name was invalid. """ value = None for p in self.property: if p.name == name: value = p.value break return value def _SetProperty(self, name, value): """Set the apps:property value with the given name to the given value. Args: name: string Name of the apps:property value to set. value: string Value to give the apps:property value with the given name. """ found = False for i in range(len(self.property)): if self.property[i].name == name: self.property[i].value = value found = True break if not found: self.property.append(gdata.apps_property.AppsProperty(name=name, value=value)) def find_edit_link(self): return self.uri class EmailSettingsLabel(EmailSettingsEntry): """Represents a Label in object form.""" def GetName(self): """Get the name of the Label object. Returns: The name of this Label object as a string or None. """ return self._GetProperty(LABEL_NAME) def SetName(self, value): """Set the name of this Label object. Args: value: string The new label name to give this object. """ self._SetProperty(LABEL_NAME, value) name = pyproperty(GetName, SetName) def __init__(self, uri=None, name=None, *args, **kwargs): """Constructs a new EmailSettingsLabel object with the given arguments. Args: uri: string (optional) The uri of of this object for HTTP requests. name: string (optional) The name to give this new object. args: The other parameters to pass to gdata.entry.GDEntry constructor. kwargs: The other parameters to pass to gdata.entry.GDEntry constructor. """ super(EmailSettingsLabel, self).__init__(*args, **kwargs) if uri: self.uri = uri if name: self.name = name class EmailSettingsFilter(EmailSettingsEntry): """Represents an Email Settings Filter in object form.""" def GetFrom(self): """Get the From value of the Filter object. Returns: The From value of this Filter object as a string or None. """ return self._GetProperty(FILTER_FROM_NAME) def SetFrom(self, value): """Set the From value of this Filter object. Args: value: string The new From value to give this object. """ self._SetProperty(FILTER_FROM_NAME, value) from_address = pyproperty(GetFrom, SetFrom) def GetTo(self): """Get the To value of the Filter object. Returns: The To value of this Filter object as a string or None. """ return self._GetProperty(FILTER_TO_NAME) def SetTo(self, value): """Set the To value of this Filter object. Args: value: string The new To value to give this object. """ self._SetProperty(FILTER_TO_NAME, value) to_address = pyproperty(GetTo, SetTo) def GetSubject(self): """Get the Subject value of the Filter object. Returns: The Subject value of this Filter object as a string or None. """ return self._GetProperty(FILTER_SUBJECT_NAME) def SetSubject(self, value): """Set the Subject value of this Filter object. Args: value: string The new Subject value to give this object. """ self._SetProperty(FILTER_SUBJECT_NAME, value) subject = pyproperty(GetSubject, SetSubject) def GetHasTheWord(self): """Get the HasTheWord value of the Filter object. Returns: The HasTheWord value of this Filter object as a string or None. """ return self._GetProperty(FILTER_HAS_THE_WORD_NAME) def SetHasTheWord(self, value): """Set the HasTheWord value of this Filter object. Args: value: string The new HasTheWord value to give this object. """ self._SetProperty(FILTER_HAS_THE_WORD_NAME, value) has_the_word = pyproperty(GetHasTheWord, SetHasTheWord) def GetDoesNotHaveTheWord(self): """Get the DoesNotHaveTheWord value of the Filter object. Returns: The DoesNotHaveTheWord value of this Filter object as a string or None. """ return self._GetProperty(FILTER_DOES_NOT_HAVE_THE_WORD_NAME) def SetDoesNotHaveTheWord(self, value): """Set the DoesNotHaveTheWord value of this Filter object. Args: value: string The new DoesNotHaveTheWord value to give this object. """ self._SetProperty(FILTER_DOES_NOT_HAVE_THE_WORD_NAME, value) does_not_have_the_word = pyproperty(GetDoesNotHaveTheWord, SetDoesNotHaveTheWord) def GetHasAttachments(self): """Get the HasAttachments value of the Filter object. Returns: The HasAttachments value of this Filter object as a string or None. """ return self._GetProperty(FILTER_HAS_ATTACHMENTS_NAME) def SetHasAttachments(self, value): """Set the HasAttachments value of this Filter object. Args: value: string The new HasAttachments value to give this object. """ self._SetProperty(FILTER_HAS_ATTACHMENTS_NAME, value) has_attachments = pyproperty(GetHasAttachments, SetHasAttachments) def GetLabel(self): """Get the Label value of the Filter object. Returns: The Label value of this Filter object as a string or None. """ return self._GetProperty(FILTER_LABEL) def SetLabel(self, value): """Set the Label value of this Filter object. Args: value: string The new Label value to give this object. """ self._SetProperty(FILTER_LABEL, value) label = pyproperty(GetLabel, SetLabel) def GetMarkAsRead(self): """Get the MarkAsRead value of the Filter object. Returns: The MarkAsRead value of this Filter object as a string or None. """ return self._GetProperty(FILTER_MARK_AS_READ) def SetMarkAsRead(self, value): """Set the MarkAsRead value of this Filter object. Args: value: string The new MarkAsRead value to give this object. """ self._SetProperty(FILTER_MARK_AS_READ, value) mark_as_read = pyproperty(GetMarkAsRead, SetMarkAsRead) def GetArchive(self): """Get the Archive value of the Filter object. Returns: The Archive value of this Filter object as a string or None. """ return self._GetProperty(FILTER_ARCHIVE) def SetArchive(self, value): """Set the Archive value of this Filter object. Args: value: string The new Archive value to give this object. """ self._SetProperty(FILTER_ARCHIVE, value) archive = pyproperty(GetArchive, SetArchive) def __init__(self, uri=None, from_address=None, to_address=None, subject=None, has_the_word=None, does_not_have_the_word=None, has_attachments=None, label=None, mark_as_read=None, archive=None, *args, **kwargs): """Constructs a new EmailSettingsFilter object with the given arguments. Args: uri: string (optional) The uri of of this object for HTTP requests. from_address: string (optional) The source email address for the filter. to_address: string (optional) The destination email address for the filter. subject: string (optional) The value the email must have in its subject to be filtered. has_the_word: string (optional) The value the email must have in its subject or body to be filtered. does_not_have_the_word: string (optional) The value the email cannot have in its subject or body to be filtered. has_attachments: Boolean (optional) Whether or not the email must have an attachment to be filtered. label: string (optional) The name of the label to apply to messages matching the filter criteria. mark_as_read: Boolean (optional) Whether or not to mark messages matching the filter criteria as read. archive: Boolean (optional) Whether or not to move messages matching to Archived state. args: The other parameters to pass to gdata.entry.GDEntry constructor. kwargs: The other parameters to pass to gdata.entry.GDEntry constructor. """ super(EmailSettingsFilter, self).__init__(*args, **kwargs) if uri: self.uri = uri if from_address: self.from_address = from_address if to_address: self.to_address = to_address if subject: self.subject = subject if has_the_word: self.has_the_word = has_the_word if does_not_have_the_word: self.does_not_have_the_word = does_not_have_the_word if has_attachments is not None: self.has_attachments = str(has_attachments) if label: self.label = label if mark_as_read is not None: self.mark_as_read = str(mark_as_read) if archive is not None: self.archive = str(archive) class EmailSettingsSendAsAlias(EmailSettingsEntry): """Represents an Email Settings send-as Alias in object form.""" def GetName(self): """Get the Name of the send-as Alias object. Returns: The Name of this send-as Alias object as a string or None. """ return self._GetProperty(SENDAS_ALIAS_NAME) def SetName(self, value): """Set the Name of this send-as Alias object. Args: value: string The new Name to give this object. """ self._SetProperty(SENDAS_ALIAS_NAME, value) name = pyproperty(GetName, SetName) def GetAddress(self): """Get the Address of the send-as Alias object. Returns: The Address of this send-as Alias object as a string or None. """ return self._GetProperty(SENDAS_ALIAS_ADDRESS) def SetAddress(self, value): """Set the Address of this send-as Alias object. Args: value: string The new Address to give this object. """ self._SetProperty(SENDAS_ALIAS_ADDRESS, value) address = pyproperty(GetAddress, SetAddress) def GetReplyTo(self): """Get the ReplyTo address of the send-as Alias object. Returns: The ReplyTo address of this send-as Alias object as a string or None. """ return self._GetProperty(SENDAS_ALIAS_REPLY_TO) def SetReplyTo(self, value): """Set the ReplyTo address of this send-as Alias object. Args: value: string The new ReplyTo address to give this object. """ self._SetProperty(SENDAS_ALIAS_REPLY_TO, value) reply_to = pyproperty(GetReplyTo, SetReplyTo) def GetMakeDefault(self): """Get the MakeDefault value of the send-as Alias object. Returns: The MakeDefault value of this send-as Alias object as a string or None. """ return self._GetProperty(SENDAS_ALIAS_MAKE_DEFAULT) def SetMakeDefault(self, value): """Set the MakeDefault value of this send-as Alias object. Args: value: string The new MakeDefault valueto give this object.WebClip """ self._SetProperty(SENDAS_ALIAS_MAKE_DEFAULT, value) make_default = pyproperty(GetMakeDefault, SetMakeDefault) def __init__(self, uri=None, name=None, address=None, reply_to=None, make_default=None, *args, **kwargs): """Constructs a new EmailSettingsSendAsAlias object with the given arguments. Args: uri: string (optional) The uri of of this object for HTTP requests. name: string (optional) The name that will appear in the "From" field for this user. address: string (optional) The email address that appears as the origination address for emails sent by this user. reply_to: string (optional) The address to be used as the reply-to address in email sent using the alias. make_default: Boolean (optional) Whether or not this alias should become the default alias for this user. args: The other parameters to pass to gdata.entry.GDEntry constructor. kwargs: The other parameters to pass to gdata.entry.GDEntry constructor. """ super(EmailSettingsSendAsAlias, self).__init__(*args, **kwargs) if uri: self.uri = uri if name: self.name = name if address: self.address = address if reply_to: self.reply_to = reply_to if make_default is not None: self.make_default = str(make_default) class EmailSettingsWebClip(EmailSettingsEntry): """Represents a WebClip in object form.""" def GetEnable(self): """Get the Enable value of the WebClip object. Returns: The Enable value of this WebClip object as a string or None. """ return self._GetProperty(WEBCLIP_ENABLE) def SetEnable(self, value): """Set the Enable value of this WebClip object. Args: value: string The new Enable value to give this object. """ self._SetProperty(WEBCLIP_ENABLE, value) enable = pyproperty(GetEnable, SetEnable) def __init__(self, uri=None, enable=None, *args, **kwargs): """Constructs a new EmailSettingsWebClip object with the given arguments. Args: uri: string (optional) The uri of of this object for HTTP requests. enable: Boolean (optional) Whether to enable showing Web clips. args: The other parameters to pass to gdata.entry.GDEntry constructor. kwargs: The other parameters to pass to gdata.entry.GDEntry constructor. """ super(EmailSettingsWebClip, self).__init__(*args, **kwargs) if uri: self.uri = uri if enable is not None: self.enable = str(enable) class EmailSettingsForwarding(EmailSettingsEntry): """Represents Forwarding settings in object form.""" def GetEnable(self): """Get the Enable value of the Forwarding object. Returns: The Enable value of this Forwarding object as a string or None. """ return self._GetProperty(FORWARDING_ENABLE) def SetEnable(self, value): """Set the Enable value of this Forwarding object. Args: value: string The new Enable value to give this object. """ self._SetProperty(FORWARDING_ENABLE, value) enable = pyproperty(GetEnable, SetEnable) def GetForwardTo(self): """Get the ForwardTo value of the Forwarding object. Returns: The ForwardTo value of this Forwarding object as a string or None. """ return self._GetProperty(FORWARDING_TO) def SetForwardTo(self, value): """Set the ForwardTo value of this Forwarding object. Args: value: string The new ForwardTo value to give this object. """ self._SetProperty(FORWARDING_TO, value) forward_to = pyproperty(GetForwardTo, SetForwardTo) def GetAction(self): """Get the Action value of the Forwarding object. Returns: The Action value of this Forwarding object as a string or None. """ return self._GetProperty(FORWARDING_ACTION) def SetAction(self, value): """Set the Action value of this Forwarding object. Args: value: string The new Action value to give this object. """ self._SetProperty(FORWARDING_ACTION, value) action = pyproperty(GetAction, SetAction) def __init__(self, uri=None, enable=None, forward_to=None, action=None, *args, **kwargs): """Constructs a new EmailSettingsForwarding object with the given arguments. Args: uri: string (optional) The uri of of this object for HTTP requests. enable: Boolean (optional) Whether to enable incoming email forwarding. forward_to: string (optional) The address email will be forwarded to. action: string (optional) The action to perform after forwarding an email ("KEEP", "ARCHIVE", "DELETE"). args: The other parameters to pass to gdata.entry.GDEntry constructor. kwargs: The other parameters to pass to gdata.entry.GDEntry constructor. """ super(EmailSettingsForwarding, self).__init__(*args, **kwargs) if uri: self.uri = uri if enable is not None: self.enable = str(enable) if forward_to: self.forward_to = forward_to if action: self.action = action class EmailSettingsPop(EmailSettingsEntry): """Represents POP settings in object form.""" def GetEnable(self): """Get the Enable value of the POP object. Returns: The Enable value of this POP object as a string or None. """ return self._GetProperty(POP_ENABLE) def SetEnable(self, value): """Set the Enable value of this POP object. Args: value: string The new Enable value to give this object. """ self._SetProperty(POP_ENABLE, value) enable = pyproperty(GetEnable, SetEnable) def GetEnableFor(self): """Get the EnableFor value of the POP object. Returns: The EnableFor value of this POP object as a string or None. """ return self._GetProperty(POP_ENABLE_FOR) def SetEnableFor(self, value): """Set the EnableFor value of this POP object. Args: value: string The new EnableFor value to give this object. """ self._SetProperty(POP_ENABLE_FOR, value) enable_for = pyproperty(GetEnableFor, SetEnableFor) def GetPopAction(self): """Get the Action value of the POP object. Returns: The Action value of this POP object as a string or None. """ return self._GetProperty(POP_ACTION) def SetPopAction(self, value): """Set the Action value of this POP object. Args: value: string The new Action value to give this object. """ self._SetProperty(POP_ACTION, value) action = pyproperty(GetPopAction, SetPopAction) def __init__(self, uri=None, enable=None, enable_for=None, action=None, *args, **kwargs): """Constructs a new EmailSettingsPOP object with the given arguments. Args: uri: string (optional) The uri of of this object for HTTP requests. enable: Boolean (optional) Whether to enable incoming POP3 access. enable_for: string (optional) Whether to enable POP3 for all mail ("ALL_MAIL"), or mail from now on ("MAIL_FROM_NOW_ON"). action: string (optional) What Google Mail should do with its copy of the email after it is retrieved using POP ("KEEP", "ARCHIVE", or "DELETE"). args: The other parameters to pass to gdata.entry.GDEntry constructor. kwargs: The other parameters to pass to gdata.entry.GDEntry constructor. """ super(EmailSettingsPop, self).__init__(*args, **kwargs) if uri: self.uri = uri if enable is not None: self.enable = str(enable) if enable_for: self.enable_for = enable_for if action: self.action = action class EmailSettingsImap(EmailSettingsEntry): """Represents IMAP settings in object form.""" def GetEnable(self): """Get the Enable value of the IMAP object. Returns: The Enable value of this IMAP object as a string or None. """ return self._GetProperty(IMAP_ENABLE) def SetEnable(self, value): """Set the Enable value of this IMAP object. Args: value: string The new Enable value to give this object. """ self._SetProperty(IMAP_ENABLE, value) enable = pyproperty(GetEnable, SetEnable) def __init__(self, uri=None, enable=None, *args, **kwargs): """Constructs a new EmailSettingsImap object with the given arguments. Args: uri: string (optional) The uri of of this object for HTTP requests. enable: Boolean (optional) Whether to enable IMAP access. args: The other parameters to pass to gdata.entry.GDEntry constructor. kwargs: The other parameters to pass to gdata.entry.GDEntry constructor. """ super(EmailSettingsImap, self).__init__(*args, **kwargs) if uri: self.uri = uri if enable is not None: self.enable = str(enable) class EmailSettingsVacationResponder(EmailSettingsEntry): """Represents Vacation Responder settings in object form.""" def GetEnable(self): """Get the Enable value of the Vacation Responder object. Returns: The Enable value of this Vacation Responder object as a string or None. """ return self._GetProperty(VACATION_RESPONDER_ENABLE) def SetEnable(self, value): """Set the Enable value of this Vacation Responder object. Args: value: string The new Enable value to give this object. """ self._SetProperty(VACATION_RESPONDER_ENABLE, value) enable = pyproperty(GetEnable, SetEnable) def GetSubject(self): """Get the Subject value of the Vacation Responder object. Returns: The Subject value of this Vacation Responder object as a string or None. """ return self._GetProperty(VACATION_RESPONDER_SUBJECT) def SetSubject(self, value): """Set the Subject value of this Vacation Responder object. Args: value: string The new Subject value to give this object. """ self._SetProperty(VACATION_RESPONDER_SUBJECT, value) subject = pyproperty(GetSubject, SetSubject) def GetMessage(self): """Get the Message value of the Vacation Responder object. Returns: The Message value of this Vacation Responder object as a string or None. """ return self._GetProperty(VACATION_RESPONDER_MESSAGE) def SetMessage(self, value): """Set the Message value of this Vacation Responder object. Args: value: string The new Message value to give this object. """ self._SetProperty(VACATION_RESPONDER_MESSAGE, value) message = pyproperty(GetMessage, SetMessage) def GetStartDate(self): """Get the StartDate value of the Vacation Responder object. Returns: The StartDate value of this Vacation Responder object as a string(YYYY-MM-DD) or None. """ return self._GetProperty(VACATION_RESPONDER_STARTDATE) def SetStartDate(self, value): """Set the StartDate value of this Vacation Responder object. Args: value: string The new StartDate value to give this object. """ self._SetProperty(VACATION_RESPONDER_STARTDATE, value) start_date = pyproperty(GetStartDate, SetStartDate) def GetEndDate(self): """Get the EndDate value of the Vacation Responder object. Returns: The EndDate value of this Vacation Responder object as a string(YYYY-MM-DD) or None. """ return self._GetProperty(VACATION_RESPONDER_ENDDATE) def SetEndDate(self, value): """Set the EndDate value of this Vacation Responder object. Args: value: string The new EndDate value to give this object. """ self._SetProperty(VACATION_RESPONDER_ENDDATE, value) end_date = pyproperty(GetEndDate, SetEndDate) def GetContactsOnly(self): """Get the ContactsOnly value of the Vacation Responder object. Returns: The ContactsOnly value of this Vacation Responder object as a string or None. """ return self._GetProperty(VACATION_RESPONDER_CONTACTS_ONLY) def SetContactsOnly(self, value): """Set the ContactsOnly value of this Vacation Responder object. Args: value: string The new ContactsOnly value to give this object. """ self._SetProperty(VACATION_RESPONDER_CONTACTS_ONLY, value) contacts_only = pyproperty(GetContactsOnly, SetContactsOnly) def GetDomainOnly(self): """Get the DomainOnly value of the Vacation Responder object. Returns: The DomainOnly value of this Vacation Responder object as a string or None. """ return self._GetProperty(VACATION_RESPONDER_DOMAIN_ONLY) def SetDomainOnly(self, value): """Set the DomainOnly value of this Vacation Responder object. Args: value: string The new DomainOnly value to give this object. """ self._SetProperty(VACATION_RESPONDER_DOMAIN_ONLY, value) domain_only = pyproperty(GetDomainOnly, SetDomainOnly) def __init__(self, uri=None, enable=None, subject=None, message=None, start_date=None, end_date=None, contacts_only=None, domain_only=None, *args, **kwargs): """Constructs a new EmailSettingsVacationResponder object with the given arguments. Args: uri: string (optional) The uri of of this object for HTTP requests. enable: Boolean (optional) Whether to enable the vacation responder. subject: string (optional) The subject line of the vacation responder autoresponse. message: string (optional) The message body of the vacation responder autoresponse. start_date: string (optional) The start date of the vacation responder autoresponse end_date: string (optional) The end date of the vacation responder autoresponse contacts_only: Boolean (optional) Whether to only send autoresponses to known contacts. domain_only: Boolean (optional) Whether to only send autoresponses to users in the same primary domain . args: The other parameters to pass to gdata.entry.GDEntry constructor. kwargs: The other parameters to pass to gdata.entry.GDEntry constructor. """ super(EmailSettingsVacationResponder, self).__init__(*args, **kwargs) if uri: self.uri = uri if enable is not None: self.enable = str(enable) if subject: self.subject = subject if message: self.message = message if start_date: self.start_date = start_date if end_date: self.end_date = end_date if contacts_only is not None: self.contacts_only = str(contacts_only) if domain_only is not None: self.domain_only = str(domain_only) class EmailSettingsSignature(EmailSettingsEntry): """Represents a Signature in object form.""" def GetValue(self): """Get the value of the Signature object. Returns: The value of this Signature object as a string or None. """ value = self._GetProperty(SIGNATURE_VALUE) if value == ' ': # hack to support empty signature return '' else: return value def SetValue(self, value): """Set the name of this Signature object. Args: value: string The new signature value to give this object. """ if value == '': # hack to support empty signature value = ' ' self._SetProperty(SIGNATURE_VALUE, value) signature_value = pyproperty(GetValue, SetValue) def __init__(self, uri=None, signature=None, *args, **kwargs): """Constructs a new EmailSettingsSignature object with the given arguments. Args: uri: string (optional) The uri of of this object for HTTP requests. signature: string (optional) The signature to be appended to outgoing messages. args: The other parameters to pass to gdata.entry.GDEntry constructor. kwargs: The other parameters to pass to gdata.entry.GDEntry constructor. """ super(EmailSettingsSignature, self).__init__(*args, **kwargs) if uri: self.uri = uri if signature is not None: self.signature_value = signature class EmailSettingsLanguage(EmailSettingsEntry): """Represents Language Settings in object form.""" def GetLanguage(self): """Get the tag of the Language object. Returns: The tag of this Language object as a string or None. """ return self._GetProperty(LANGUAGE_TAG) def SetLanguage(self, value): """Set the tag of this Language object. Args: value: string The new tag value to give this object. """ self._SetProperty(LANGUAGE_TAG, value) language_tag = pyproperty(GetLanguage, SetLanguage) def __init__(self, uri=None, language=None, *args, **kwargs): """Constructs a new EmailSettingsLanguage object with the given arguments. Args: uri: string (optional) The uri of of this object for HTTP requests. language: string (optional) The language tag for Google Mail's display language. args: The other parameters to pass to gdata.entry.GDEntry constructor. kwargs: The other parameters to pass to gdata.entry.GDEntry constructor. """ super(EmailSettingsLanguage, self).__init__(*args, **kwargs) if uri: self.uri = uri if language: self.language_tag = language class EmailSettingsGeneral(EmailSettingsEntry): """Represents General Settings in object form.""" def GetPageSize(self): """Get the Page Size value of the General Settings object. Returns: The Page Size value of this General Settings object as a string or None. """ return self._GetProperty(GENERAL_PAGE_SIZE) def SetPageSize(self, value): """Set the Page Size value of this General Settings object. Args: value: string The new Page Size value to give this object. """ self._SetProperty(GENERAL_PAGE_SIZE, value) page_size = pyproperty(GetPageSize, SetPageSize) def GetShortcuts(self): """Get the Shortcuts value of the General Settings object. Returns: The Shortcuts value of this General Settings object as a string or None. """ return self._GetProperty(GENERAL_SHORTCUTS) def SetShortcuts(self, value): """Set the Shortcuts value of this General Settings object. Args: value: string The new Shortcuts value to give this object. """ self._SetProperty(GENERAL_SHORTCUTS, value) shortcuts = pyproperty(GetShortcuts, SetShortcuts) def GetArrows(self): """Get the Arrows value of the General Settings object. Returns: The Arrows value of this General Settings object as a string or None. """ return self._GetProperty(GENERAL_ARROWS) def SetArrows(self, value): """Set the Arrows value of this General Settings object. Args: value: string The new Arrows value to give this object. """ self._SetProperty(GENERAL_ARROWS, value) arrows = pyproperty(GetArrows, SetArrows) def GetSnippets(self): """Get the Snippets value of the General Settings object. Returns: The Snippets value of this General Settings object as a string or None. """ return self._GetProperty(GENERAL_SNIPPETS) def SetSnippets(self, value): """Set the Snippets value of this General Settings object. Args: value: string The new Snippets value to give this object. """ self._SetProperty(GENERAL_SNIPPETS, value) snippets = pyproperty(GetSnippets, SetSnippets) def GetUnicode(self): """Get the Unicode value of the General Settings object. Returns: The Unicode value of this General Settings object as a string or None. """ return self._GetProperty(GENERAL_UNICODE) def SetUnicode(self, value): """Set the Unicode value of this General Settings object. Args: value: string The new Unicode value to give this object. """ self._SetProperty(GENERAL_UNICODE, value) use_unicode = pyproperty(GetUnicode, SetUnicode) def __init__(self, uri=None, page_size=None, shortcuts=None, arrows=None, snippets=None, use_unicode=None, *args, **kwargs): """Constructs a new EmailSettingsGeneral object with the given arguments. Args: uri: string (optional) The uri of of this object for HTTP requests. page_size: int (optional) The number of conversations to be shown per page. shortcuts: Boolean (optional) Whether to enable keyboard shortcuts. arrows: Boolean (optional) Whether to display arrow-shaped personal indicators next to email sent specifically to the user. snippets: Boolean (optional) Whether to display snippets of the messages in the inbox and when searching. use_unicode: Boolean (optional) Whether to use UTF-8 (unicode) encoding for all outgoing messages. args: The other parameters to pass to gdata.entry.GDEntry constructor. kwargs: The other parameters to pass to gdata.entry.GDEntry constructor. """ super(EmailSettingsGeneral, self).__init__(*args, **kwargs) if uri: self.uri = uri if page_size is not None: self.page_size = str(page_size) if shortcuts is not None: self.shortcuts = str(shortcuts) if arrows is not None: self.arrows = str(arrows) if snippets is not None: self.snippets = str(snippets) if use_unicode is not None: self.use_unicode = str(use_unicode) class EmailSettingsDelegation(EmailSettingsEntry): """Represents an Email Settings delegation entry in object form.""" def GetAddress(self): """Get the email address of the delegated user. Returns: The email address of the delegated user as a string or None. """ return self._GetProperty(DELEGATION_ADDRESS) def SetAddress(self, value): """Set the email address of of the delegated user. Args: value: string The email address of another user on the same domain """ self._SetProperty(DELEGATION_ADDRESS, value) address = pyproperty(GetAddress, SetAddress) def __init__(self, uri=None, address=None, *args, **kwargs): """Constructs a new EmailSettingsDelegation object with the given arguments. Args: uri: string (optional) The uri of of this object for HTTP requests. address: string The email address of the delegated user. """ super(EmailSettingsDelegation, self).__init__(*args, **kwargs) if uri: self.uri = uri if address: self.address = address
PandaPayProject/PandaPay
refs/heads/master
qa/rpc-tests/fundrawtransaction.py
1
#!/usr/bin/env python2 # Copyright (c) 2014-2015 The Bitcoin Core developers # Distributed under the MIT software license, see the accompanying # file COPYING or http://www.opensource.org/licenses/mit-license.php. from test_framework.test_framework import BitcoinTestFramework from test_framework.util import * # Create one-input, one-output, no-fee transaction: class RawTransactionsTest(BitcoinTestFramework): def setup_chain(self): print("Initializing test directory "+self.options.tmpdir) initialize_chain_clean(self.options.tmpdir, 4) def setup_network(self, split=False): self.nodes = start_nodes(4, self.options.tmpdir) connect_nodes_bi(self.nodes,0,1) connect_nodes_bi(self.nodes,1,2) connect_nodes_bi(self.nodes,0,2) connect_nodes_bi(self.nodes,0,3) self.is_network_split=False self.sync_all() def run_test(self): print "Mining blocks..." min_relay_tx_fee = self.nodes[0].getnetworkinfo()['relayfee'] # This test is not meant to test fee estimation and we'd like # to be sure all txs are sent at a consistent desired feerate for node in self.nodes: node.settxfee(min_relay_tx_fee) # if the fee's positive delta is higher than this value tests will fail, # neg. delta always fail the tests. # The size of the signature of every input may be at most 2 bytes larger # than a minimum sized signature. # = 2 bytes * minRelayTxFeePerByte feeTolerance = 2 * min_relay_tx_fee/1000 self.nodes[2].generate(1) self.sync_all() self.nodes[0].generate(121) self.sync_all() watchonly_address = self.nodes[0].getnewaddress() watchonly_pubkey = self.nodes[0].validateaddress(watchonly_address)["pubkey"] watchonly_amount = Decimal(2000) self.nodes[3].importpubkey(watchonly_pubkey, "", True) watchonly_txid = self.nodes[0].sendtoaddress(watchonly_address, watchonly_amount) self.nodes[0].sendtoaddress(self.nodes[3].getnewaddress(), watchonly_amount / 10) self.nodes[0].sendtoaddress(self.nodes[2].getnewaddress(), 15) self.nodes[0].sendtoaddress(self.nodes[2].getnewaddress(), 10) self.nodes[0].sendtoaddress(self.nodes[2].getnewaddress(), 50) self.sync_all() self.nodes[0].generate(1) self.sync_all() ############### # simple test # ############### inputs = [ ] outputs = { self.nodes[0].getnewaddress() : 10 } rawtx = self.nodes[2].createrawtransaction(inputs, outputs) dec_tx = self.nodes[2].decoderawtransaction(rawtx) rawtxfund = self.nodes[2].fundrawtransaction(rawtx) fee = rawtxfund['fee'] dec_tx = self.nodes[2].decoderawtransaction(rawtxfund['hex']) assert(len(dec_tx['vin']) > 0) #test if we have enought inputs ############################## # simple test with two coins # ############################## inputs = [ ] outputs = { self.nodes[0].getnewaddress() : 22 } rawtx = self.nodes[2].createrawtransaction(inputs, outputs) dec_tx = self.nodes[2].decoderawtransaction(rawtx) rawtxfund = self.nodes[2].fundrawtransaction(rawtx) fee = rawtxfund['fee'] dec_tx = self.nodes[2].decoderawtransaction(rawtxfund['hex']) assert(len(dec_tx['vin']) > 0) #test if we have enough inputs ############################## # simple test with two coins # ############################## inputs = [ ] outputs = { self.nodes[0].getnewaddress() : 26 } rawtx = self.nodes[2].createrawtransaction(inputs, outputs) dec_tx = self.nodes[2].decoderawtransaction(rawtx) rawtxfund = self.nodes[2].fundrawtransaction(rawtx) fee = rawtxfund['fee'] dec_tx = self.nodes[2].decoderawtransaction(rawtxfund['hex']) assert(len(dec_tx['vin']) > 0) assert_equal(dec_tx['vin'][0]['scriptSig']['hex'], '') ################################ # simple test with two outputs # ################################ inputs = [ ] outputs = { self.nodes[0].getnewaddress() : 26, self.nodes[1].getnewaddress() : 25 } rawtx = self.nodes[2].createrawtransaction(inputs, outputs) dec_tx = self.nodes[2].decoderawtransaction(rawtx) rawtxfund = self.nodes[2].fundrawtransaction(rawtx) fee = rawtxfund['fee'] dec_tx = self.nodes[2].decoderawtransaction(rawtxfund['hex']) totalOut = 0 for out in dec_tx['vout']: totalOut += out['value'] assert(len(dec_tx['vin']) > 0) assert_equal(dec_tx['vin'][0]['scriptSig']['hex'], '') ######################################################################### # test a fundrawtransaction with a VIN greater than the required amount # ######################################################################### utx = False listunspent = self.nodes[2].listunspent() for aUtx in listunspent: if aUtx['amount'] == 50: utx = aUtx break assert(utx!=False) inputs = [ {'txid' : utx['txid'], 'vout' : utx['vout']}] outputs = { self.nodes[0].getnewaddress() : 10 } rawtx = self.nodes[2].createrawtransaction(inputs, outputs) dec_tx = self.nodes[2].decoderawtransaction(rawtx) assert_equal(utx['txid'], dec_tx['vin'][0]['txid']) rawtxfund = self.nodes[2].fundrawtransaction(rawtx) fee = rawtxfund['fee'] dec_tx = self.nodes[2].decoderawtransaction(rawtxfund['hex']) totalOut = 0 for out in dec_tx['vout']: totalOut += out['value'] assert_equal(fee + totalOut, utx['amount']) #compare vin total and totalout+fee ##################################################################### # test a fundrawtransaction with which will not get a change output # ##################################################################### utx = False listunspent = self.nodes[2].listunspent() for aUtx in listunspent: if aUtx['amount'] == 50: utx = aUtx break assert(utx!=False) inputs = [ {'txid' : utx['txid'], 'vout' : utx['vout']}] outputs = { self.nodes[0].getnewaddress() : Decimal(50) - fee - feeTolerance } rawtx = self.nodes[2].createrawtransaction(inputs, outputs) dec_tx = self.nodes[2].decoderawtransaction(rawtx) assert_equal(utx['txid'], dec_tx['vin'][0]['txid']) rawtxfund = self.nodes[2].fundrawtransaction(rawtx) fee = rawtxfund['fee'] dec_tx = self.nodes[2].decoderawtransaction(rawtxfund['hex']) totalOut = 0 for out in dec_tx['vout']: totalOut += out['value'] assert_equal(rawtxfund['changepos'], -1) assert_equal(fee + totalOut, utx['amount']) #compare vin total and totalout+fee ######################################################################### # test a fundrawtransaction with a VIN smaller than the required amount # ######################################################################### utx = False listunspent = self.nodes[2].listunspent() for aUtx in listunspent: if aUtx['amount'] == 10: utx = aUtx break assert(utx!=False) inputs = [ {'txid' : utx['txid'], 'vout' : utx['vout']}] outputs = { self.nodes[0].getnewaddress() : 10 } rawtx = self.nodes[2].createrawtransaction(inputs, outputs) # 4-byte version + 1-byte vin count + 36-byte prevout then script_len rawtx = rawtx[:82] + "0100" + rawtx[84:] dec_tx = self.nodes[2].decoderawtransaction(rawtx) assert_equal(utx['txid'], dec_tx['vin'][0]['txid']) assert_equal("00", dec_tx['vin'][0]['scriptSig']['hex']) rawtxfund = self.nodes[2].fundrawtransaction(rawtx) fee = rawtxfund['fee'] dec_tx = self.nodes[2].decoderawtransaction(rawtxfund['hex']) totalOut = 0 matchingOuts = 0 for i, out in enumerate(dec_tx['vout']): totalOut += out['value'] if out['scriptPubKey']['addresses'][0] in outputs: matchingOuts+=1 else: assert_equal(i, rawtxfund['changepos']) assert_equal(utx['txid'], dec_tx['vin'][0]['txid']) assert_equal("00", dec_tx['vin'][0]['scriptSig']['hex']) assert_equal(matchingOuts, 1) assert_equal(len(dec_tx['vout']), 2) ########################################### # test a fundrawtransaction with two VINs # ########################################### utx = False utx2 = False listunspent = self.nodes[2].listunspent() for aUtx in listunspent: if aUtx['amount'] == 10: utx = aUtx if aUtx['amount'] == 50: utx2 = aUtx assert(utx!=False) inputs = [ {'txid' : utx['txid'], 'vout' : utx['vout']},{'txid' : utx2['txid'], 'vout' : utx2['vout']} ] outputs = { self.nodes[0].getnewaddress() : 60 } rawtx = self.nodes[2].createrawtransaction(inputs, outputs) dec_tx = self.nodes[2].decoderawtransaction(rawtx) assert_equal(utx['txid'], dec_tx['vin'][0]['txid']) rawtxfund = self.nodes[2].fundrawtransaction(rawtx) fee = rawtxfund['fee'] dec_tx = self.nodes[2].decoderawtransaction(rawtxfund['hex']) totalOut = 0 matchingOuts = 0 for out in dec_tx['vout']: totalOut += out['value'] if out['scriptPubKey']['addresses'][0] in outputs: matchingOuts+=1 assert_equal(matchingOuts, 1) assert_equal(len(dec_tx['vout']), 2) matchingIns = 0 for vinOut in dec_tx['vin']: for vinIn in inputs: if vinIn['txid'] == vinOut['txid']: matchingIns+=1 assert_equal(matchingIns, 2) #we now must see two vins identical to vins given as params ######################################################### # test a fundrawtransaction with two VINs and two vOUTs # ######################################################### utx = False utx2 = False listunspent = self.nodes[2].listunspent() for aUtx in listunspent: if aUtx['amount'] == 10: utx = aUtx if aUtx['amount'] == 50: utx2 = aUtx assert(utx!=False) inputs = [ {'txid' : utx['txid'], 'vout' : utx['vout']},{'txid' : utx2['txid'], 'vout' : utx2['vout']} ] outputs = { self.nodes[0].getnewaddress() : 60, self.nodes[0].getnewaddress() : 10 } rawtx = self.nodes[2].createrawtransaction(inputs, outputs) dec_tx = self.nodes[2].decoderawtransaction(rawtx) assert_equal(utx['txid'], dec_tx['vin'][0]['txid']) rawtxfund = self.nodes[2].fundrawtransaction(rawtx) fee = rawtxfund['fee'] dec_tx = self.nodes[2].decoderawtransaction(rawtxfund['hex']) totalOut = 0 matchingOuts = 0 for out in dec_tx['vout']: totalOut += out['value'] if out['scriptPubKey']['addresses'][0] in outputs: matchingOuts+=1 assert_equal(matchingOuts, 2) assert_equal(len(dec_tx['vout']), 3) ############################################## # test a fundrawtransaction with invalid vin # ############################################## listunspent = self.nodes[2].listunspent() inputs = [ {'txid' : "1c7f966dab21119bac53213a2bc7532bff1fa844c124fd750a7d0b1332440bd1", 'vout' : 0} ] #invalid vin! outputs = { self.nodes[0].getnewaddress() : 10} rawtx = self.nodes[2].createrawtransaction(inputs, outputs) dec_tx = self.nodes[2].decoderawtransaction(rawtx) try: rawtxfund = self.nodes[2].fundrawtransaction(rawtx) raise AssertionError("Spent more than available") except JSONRPCException as e: assert("Insufficient" in e.error['message']) ############################################################ #compare fee of a standard pubkeyhash transaction inputs = [] outputs = {self.nodes[1].getnewaddress():11} rawTx = self.nodes[0].createrawtransaction(inputs, outputs) fundedTx = self.nodes[0].fundrawtransaction(rawTx) #create same transaction over sendtoaddress txId = self.nodes[0].sendtoaddress(self.nodes[1].getnewaddress(), 11) signedFee = self.nodes[0].getrawmempool(True)[txId]['fee'] #compare fee feeDelta = Decimal(fundedTx['fee']) - Decimal(signedFee) assert(feeDelta >= 0 and feeDelta <= feeTolerance) ############################################################ ############################################################ #compare fee of a standard pubkeyhash transaction with multiple outputs inputs = [] outputs = {self.nodes[1].getnewaddress():11,self.nodes[1].getnewaddress():12,self.nodes[1].getnewaddress():1,self.nodes[1].getnewaddress():13,self.nodes[1].getnewaddress():2,self.nodes[1].getnewaddress():3} rawTx = self.nodes[0].createrawtransaction(inputs, outputs) fundedTx = self.nodes[0].fundrawtransaction(rawTx) #create same transaction over sendtoaddress txId = self.nodes[0].sendmany("", outputs) signedFee = self.nodes[0].getrawmempool(True)[txId]['fee'] #compare fee feeDelta = Decimal(fundedTx['fee']) - Decimal(signedFee) assert(feeDelta >= 0 and feeDelta <= feeTolerance) ############################################################ ############################################################ #compare fee of a 2of2 multisig p2sh transaction # create 2of2 addr addr1 = self.nodes[1].getnewaddress() addr2 = self.nodes[1].getnewaddress() addr1Obj = self.nodes[1].validateaddress(addr1) addr2Obj = self.nodes[1].validateaddress(addr2) mSigObj = self.nodes[1].addmultisigaddress(2, [addr1Obj['pubkey'], addr2Obj['pubkey']]) inputs = [] outputs = {mSigObj:11} rawTx = self.nodes[0].createrawtransaction(inputs, outputs) fundedTx = self.nodes[0].fundrawtransaction(rawTx) #create same transaction over sendtoaddress txId = self.nodes[0].sendtoaddress(mSigObj, 11) signedFee = self.nodes[0].getrawmempool(True)[txId]['fee'] #compare fee feeDelta = Decimal(fundedTx['fee']) - Decimal(signedFee) assert(feeDelta >= 0 and feeDelta <= feeTolerance) ############################################################ ############################################################ #compare fee of a standard pubkeyhash transaction # create 4of5 addr addr1 = self.nodes[1].getnewaddress() addr2 = self.nodes[1].getnewaddress() addr3 = self.nodes[1].getnewaddress() addr4 = self.nodes[1].getnewaddress() addr5 = self.nodes[1].getnewaddress() addr1Obj = self.nodes[1].validateaddress(addr1) addr2Obj = self.nodes[1].validateaddress(addr2) addr3Obj = self.nodes[1].validateaddress(addr3) addr4Obj = self.nodes[1].validateaddress(addr4) addr5Obj = self.nodes[1].validateaddress(addr5) mSigObj = self.nodes[1].addmultisigaddress(4, [addr1Obj['pubkey'], addr2Obj['pubkey'], addr3Obj['pubkey'], addr4Obj['pubkey'], addr5Obj['pubkey']]) inputs = [] outputs = {mSigObj:11} rawTx = self.nodes[0].createrawtransaction(inputs, outputs) fundedTx = self.nodes[0].fundrawtransaction(rawTx) #create same transaction over sendtoaddress txId = self.nodes[0].sendtoaddress(mSigObj, 11) signedFee = self.nodes[0].getrawmempool(True)[txId]['fee'] #compare fee feeDelta = Decimal(fundedTx['fee']) - Decimal(signedFee) assert(feeDelta >= 0 and feeDelta <= feeTolerance) ############################################################ ############################################################ # spend a 2of2 multisig transaction over fundraw # create 2of2 addr addr1 = self.nodes[2].getnewaddress() addr2 = self.nodes[2].getnewaddress() addr1Obj = self.nodes[2].validateaddress(addr1) addr2Obj = self.nodes[2].validateaddress(addr2) mSigObj = self.nodes[2].addmultisigaddress(2, [addr1Obj['pubkey'], addr2Obj['pubkey']]) # send 12 PANDA to msig addr txId = self.nodes[0].sendtoaddress(mSigObj, 12) self.sync_all() self.nodes[1].generate(1) self.sync_all() oldBalance = self.nodes[1].getbalance() inputs = [] outputs = {self.nodes[1].getnewaddress():11} rawTx = self.nodes[2].createrawtransaction(inputs, outputs) fundedTx = self.nodes[2].fundrawtransaction(rawTx) signedTx = self.nodes[2].signrawtransaction(fundedTx['hex']) txId = self.nodes[2].sendrawtransaction(signedTx['hex']) self.sync_all() self.nodes[1].generate(1) self.sync_all() # make sure funds are received at node1 assert_equal(oldBalance+Decimal('11.0000000'), self.nodes[1].getbalance()) ############################################################ # locked wallet test self.nodes[1].encryptwallet("test") self.nodes.pop(1) stop_nodes(self.nodes) wait_bitcoinds() self.nodes = start_nodes(4, self.options.tmpdir) # This test is not meant to test fee estimation and we'd like # to be sure all txs are sent at a consistent desired feerate for node in self.nodes: node.settxfee(min_relay_tx_fee) connect_nodes_bi(self.nodes,0,1) connect_nodes_bi(self.nodes,1,2) connect_nodes_bi(self.nodes,0,2) connect_nodes_bi(self.nodes,0,3) self.is_network_split=False self.sync_all() try: self.nodes[1].sendtoaddress(self.nodes[0].getnewaddress(), 12) raise AssertionError("Wallet unlocked without passphrase") except JSONRPCException as e: assert('walletpassphrase' in e.error['message']) oldBalance = self.nodes[0].getbalance() inputs = [] outputs = {self.nodes[0].getnewaddress():11} rawTx = self.nodes[1].createrawtransaction(inputs, outputs) fundedTx = self.nodes[1].fundrawtransaction(rawTx) #now we need to unlock self.nodes[1].walletpassphrase("test", 100) signedTx = self.nodes[1].signrawtransaction(fundedTx['hex']) txId = self.nodes[1].sendrawtransaction(signedTx['hex']) self.sync_all() self.nodes[1].generate(1) self.sync_all() # make sure funds are received at node1 assert_equal(oldBalance+Decimal('511.0000000'), self.nodes[0].getbalance()) ############################################### # multiple (~19) inputs tx test | Compare fee # ############################################### #empty node1, send some small coins from node0 to node1 self.nodes[1].sendtoaddress(self.nodes[0].getnewaddress(), self.nodes[1].getbalance(), "", "", True) self.sync_all() self.nodes[0].generate(1) self.sync_all() for i in range(0,20): self.nodes[0].sendtoaddress(self.nodes[1].getnewaddress(), 0.01) self.sync_all() self.nodes[0].generate(1) self.sync_all() #fund a tx with ~20 small inputs inputs = [] outputs = {self.nodes[0].getnewaddress():0.15,self.nodes[0].getnewaddress():0.04} rawTx = self.nodes[1].createrawtransaction(inputs, outputs) fundedTx = self.nodes[1].fundrawtransaction(rawTx) #create same transaction over sendtoaddress txId = self.nodes[1].sendmany("", outputs) signedFee = self.nodes[1].getrawmempool(True)[txId]['fee'] #compare fee feeDelta = Decimal(fundedTx['fee']) - Decimal(signedFee) assert(feeDelta >= 0 and feeDelta <= feeTolerance*19) #~19 inputs ############################################# # multiple (~19) inputs tx test | sign/send # ############################################# #again, empty node1, send some small coins from node0 to node1 self.nodes[1].sendtoaddress(self.nodes[0].getnewaddress(), self.nodes[1].getbalance(), "", "", True) self.sync_all() self.nodes[0].generate(1) self.sync_all() for i in range(0,20): self.nodes[0].sendtoaddress(self.nodes[1].getnewaddress(), 0.01) self.sync_all() self.nodes[0].generate(1) self.sync_all() #fund a tx with ~20 small inputs oldBalance = self.nodes[0].getbalance() inputs = [] outputs = {self.nodes[0].getnewaddress():0.15,self.nodes[0].getnewaddress():0.04} rawTx = self.nodes[1].createrawtransaction(inputs, outputs) fundedTx = self.nodes[1].fundrawtransaction(rawTx) fundedAndSignedTx = self.nodes[1].signrawtransaction(fundedTx['hex']) txId = self.nodes[1].sendrawtransaction(fundedAndSignedTx['hex']) self.sync_all() self.nodes[0].generate(1) self.sync_all() assert_equal(oldBalance+Decimal('500.19000000'), self.nodes[0].getbalance()) #0.19+block reward ##################################################### # test fundrawtransaction with OP_RETURN and no vin # ##################################################### rawtx = "0100000000010000000000000000066a047465737400000000" dec_tx = self.nodes[2].decoderawtransaction(rawtx) assert_equal(len(dec_tx['vin']), 0) assert_equal(len(dec_tx['vout']), 1) rawtxfund = self.nodes[2].fundrawtransaction(rawtx) dec_tx = self.nodes[2].decoderawtransaction(rawtxfund['hex']) assert_greater_than(len(dec_tx['vin']), 0) # at least one vin assert_equal(len(dec_tx['vout']), 2) # one change output added ################################################## # test a fundrawtransaction using only watchonly # ################################################## inputs = [] outputs = {self.nodes[2].getnewaddress() : watchonly_amount / 2} rawtx = self.nodes[3].createrawtransaction(inputs, outputs) result = self.nodes[3].fundrawtransaction(rawtx, True) res_dec = self.nodes[0].decoderawtransaction(result["hex"]) assert_equal(len(res_dec["vin"]), 1) assert_equal(res_dec["vin"][0]["txid"], watchonly_txid) assert("fee" in result.keys()) assert_greater_than(result["changepos"], -1) ############################################################### # test fundrawtransaction using the entirety of watched funds # ############################################################### inputs = [] outputs = {self.nodes[2].getnewaddress() : watchonly_amount} rawtx = self.nodes[3].createrawtransaction(inputs, outputs) result = self.nodes[3].fundrawtransaction(rawtx, True) res_dec = self.nodes[0].decoderawtransaction(result["hex"]) assert_equal(len(res_dec["vin"]), 2) assert(res_dec["vin"][0]["txid"] == watchonly_txid or res_dec["vin"][1]["txid"] == watchonly_txid) assert_greater_than(result["fee"], 0) assert_greater_than(result["changepos"], -1) assert_equal(result["fee"] + res_dec["vout"][result["changepos"]]["value"], watchonly_amount / 10) signedtx = self.nodes[3].signrawtransaction(result["hex"]) assert(not signedtx["complete"]) signedtx = self.nodes[0].signrawtransaction(signedtx["hex"]) assert(signedtx["complete"]) self.nodes[0].sendrawtransaction(signedtx["hex"]) if __name__ == '__main__': RawTransactionsTest().main()
mvidalgarcia/indico
refs/heads/master
indico/legacy/common/cache.py
2
# This file is part of Indico. # Copyright (C) 2002 - 2019 CERN # # Indico is free software; you can redistribute it and/or # modify it under the terms of the MIT License; see the # LICENSE file for more details. import cPickle as pickle import datetime import hashlib import os import time from itertools import izip import redis from flask import g from indico.core.config import config from indico.core.logger import Logger from indico.legacy.common.utils import OSSpecific from indico.util.fs import silentremove from indico.util.string import truncate # To cache `None` we need to actually store something else since memcached # does not distinguish between a None value and a cache miss... class _NoneValue(object): @classmethod def replace(cls, value): """Replaces `None` with a `_NoneValue`""" return cls() if value is None else value @classmethod def restore(cls, value): """Replaces `_NoneValue` with `None`""" return None if isinstance(value, cls) else value class CacheClient(object): """This is an abstract class. A cache client provide a simple API to get/set/delete cache entries. Implementation must provide the following methods: - set(self, key, val, ttl) - get(self, key) - delete(self, key) The unit for the ttl arguments is a second. """ def set_multi(self, mapping, ttl=0): for key, val in mapping.iteritems(): self.set(key, val, ttl) def get_multi(self, keys): values = {} for key in keys: val = self.get(key) if val is not None: values[key] = val return values def delete_multi(self, keys): for key in keys: self.delete(key) def set(self, key, val, ttl=0): raise NotImplementedError def get(self, key): raise NotImplementedError def delete(self, key): raise NotImplementedError class NullCacheClient(CacheClient): """Does nothing""" def set(self, key, val, ttl=0): pass def get(self, key): return None def delete(self, key): pass class RedisCacheClient(CacheClient): """Redis-based cache client with a simple API""" key_prefix = 'cache/gen/' def __init__(self, url): self._client = redis.StrictRedis.from_url(url) self._client.connection_pool.connection_kwargs['socket_timeout'] = 1 def _unpickle(self, val): if val is None: return None return pickle.loads(val) def hash_key(self, key): # Redis keys are even binary-safe, no need to hash anything return key def set_multi(self, mapping, ttl=0): try: self._client.mset(dict((k, pickle.dumps(v)) for k, v in mapping.iteritems())) if ttl: for key in mapping: self._client.expire(key, ttl) except redis.RedisError: Logger.get('cache.redis').exception('set_multi(%r, %r) failed', mapping, ttl) def get_multi(self, keys): try: return dict(zip(keys, map(self._unpickle, self._client.mget(keys)))) except redis.RedisError: Logger.get('cache.redis').exception('get_multi(%r) failed', keys) def delete_multi(self, keys): try: self._client.delete(*keys) except redis.RedisError: Logger.get('cache.redis').exception('delete_multi(%r) failed', keys) def set(self, key, val, ttl=0): try: if ttl: self._client.setex(key, ttl, pickle.dumps(val)) else: self._client.set(key, pickle.dumps(val)) except redis.RedisError: val_repr = truncate(repr(val), 1000) Logger.get('cache.redis').exception('set(%r, %s, %r) failed', key, val_repr, ttl) def get(self, key): try: return self._unpickle(self._client.get(key)) except redis.RedisError: Logger.get('cache.redis').exception('get(%r) failed', key) def delete(self, key): try: self._client.delete(key) except redis.RedisError: Logger.get('cache.redis').exception('delete(%r) failed', key) class FileCacheClient(CacheClient): """File-based cache with a memcached-like API. Contains only features needed by GenericCache. """ def __init__(self, dir): self._dir = os.path.join(dir, 'generic_cache') def _getFilePath(self, key, mkdir=True): # We assume keys have a 'namespace.hashedKey' format parts = key.split('.', 1) if len(parts) == 1: namespace = '_' filename = parts[0] else: namespace, filename = parts dir = os.path.join(self._dir, namespace, filename[:4], filename[:8]) if mkdir and not os.path.exists(dir): try: os.makedirs(dir) except OSError: # Handle race condition if not os.path.exists(dir): raise return os.path.join(dir, filename) def set(self, key, val, ttl=0): try: f = open(self._getFilePath(key), 'wb') OSSpecific.lockFile(f, 'LOCK_EX') try: expiry = int(time.time()) + ttl if ttl else None data = (expiry, val) pickle.dump(data, f) finally: OSSpecific.lockFile(f, 'LOCK_UN') f.close() except (IOError, OSError): Logger.get('cache.files').exception('Error setting value in cache') return 0 return 1 def get(self, key): try: path = self._getFilePath(key, False) if not os.path.exists(path): return None f = open(path, 'rb') OSSpecific.lockFile(f, 'LOCK_SH') expiry = val = None try: expiry, val = pickle.load(f) finally: OSSpecific.lockFile(f, 'LOCK_UN') f.close() if expiry and time.time() > expiry: return None except (IOError, OSError): Logger.get('cache.files').exception('Error getting cached value') return None except (EOFError, pickle.UnpicklingError): Logger.get('cache.files').exception('Cached information seems corrupted. Overwriting it.') return None return val def delete(self, key): path = self._getFilePath(key, False) if os.path.exists(path): silentremove(path) return 1 class MemcachedCacheClient(CacheClient): """Memcached-based cache client""" @staticmethod def convert_ttl(ttl): """Convert a ttl in seconds to a timestamp for use with memcached.""" return (int(time.time()) + ttl) if ttl else 0 def __init__(self, servers): import memcache self._client = memcache.Client(servers) def set(self, key, val, ttl=0): return self._client.set(key, val, self.convert_ttl(ttl)) def get(self, key): return self._client.get(key) def delete(self, key): return self._client.delete(key) class GenericCache(object): """A simple cache interface that supports various backends. The backends are accessed through the CacheClient interface. """ def __init__(self, namespace): self._client = None self._namespace = namespace def __repr__(self): return 'GenericCache(%r)' % self._namespace def _connect(self): """Connect to the CacheClient. This method must be called before accessing ``self._client``. """ # Maybe we already have a client in this instance if self._client is not None: return # If not, we might have one from another instance self._client = g.get('generic_cache_client', None) if self._client is not None: return # If not, create a new one backend = config.CACHE_BACKEND if backend == 'memcached': self._client = MemcachedCacheClient(config.MEMCACHED_SERVERS) elif backend == 'redis': self._client = RedisCacheClient(config.REDIS_CACHE_URL) elif backend == 'files': self._client = FileCacheClient(config.CACHE_DIR) else: self._client = NullCacheClient() g.generic_cache_client = self._client def _hashKey(self, key): if hasattr(self._client, 'hash_key'): return self._client.hash_key(key) return hashlib.sha256(key).hexdigest() def _makeKey(self, key): if not isinstance(key, basestring): # In case we get something not a string (number, list, whatever) key = repr(key) # Hashlib doesn't allow unicode so let's ensure it's not! key = key.encode('utf-8') return '%s%s.%s' % (getattr(self._client, 'key_prefix', ''), self._namespace, self._hashKey(key)) def _processTime(self, ts): if isinstance(ts, datetime.timedelta): ts = ts.seconds + (ts.days * 24 * 3600) return ts def set(self, key, val, time=0): """Set key to val with optional validity time. :param key: the key of the cache entry :param val: any python object that can be pickled :param time: number of seconds or a datetime.timedelta """ self._connect() time = self._processTime(time) Logger.get('cache.generic').debug('SET %s %r (%d)', self._namespace, key, time) self._client.set(self._makeKey(key), _NoneValue.replace(val), time) def set_multi(self, mapping, time=0): self._connect() time = self._processTime(time) mapping = dict(((self._makeKey(key), _NoneValue.replace(val)) for key, val in mapping.iteritems())) self._client.set_multi(mapping, time) def get(self, key, default=None): self._connect() res = self._client.get(self._makeKey(key)) Logger.get('cache.generic').debug('GET %s %r (%s)', self._namespace, key, 'HIT' if res is not None else 'MISS') if res is None: return default return _NoneValue.restore(res) def get_multi(self, keys, default=None, asdict=True): self._connect() real_keys = map(self._makeKey, keys) data = self._client.get_multi(real_keys) # Add missing keys for real_key in real_keys: if real_key not in data: data[real_key] = default # Get data in the same order as our keys sorted_data = (default if data[rk] is None else _NoneValue.restore(data[rk]) for rk in real_keys) if asdict: return dict(izip(keys, sorted_data)) else: return list(sorted_data) def delete(self, key): self._connect() Logger.get('cache.generic').debug('DEL %s %r', self._namespace, key) self._client.delete(self._makeKey(key)) def delete_multi(self, keys): self._connect() keys = map(self._makeKey, keys) self._client.delete_multi(keys)
smart-developerr/my-first-blog
refs/heads/master
Lib/site-packages/setuptools/command/__init__.py
101
__all__ = [ 'alias', 'bdist_egg', 'bdist_rpm', 'build_ext', 'build_py', 'develop', 'easy_install', 'egg_info', 'install', 'install_lib', 'rotate', 'saveopts', 'sdist', 'setopt', 'test', 'install_egg_info', 'install_scripts', 'register', 'bdist_wininst', 'upload_docs', 'upload', ] from distutils.command.bdist import bdist import sys from setuptools.command import install_scripts if 'egg' not in bdist.format_commands: bdist.format_command['egg'] = ('bdist_egg', "Python .egg file") bdist.format_commands.append('egg') del bdist, sys
sdlBasic/sdlbrt
refs/heads/master
win32/mingw/opt/lib/python2.7/re.py
131
# # Secret Labs' Regular Expression Engine # # re-compatible interface for the sre matching engine # # Copyright (c) 1998-2001 by Secret Labs AB. All rights reserved. # # This version of the SRE library can be redistributed under CNRI's # Python 1.6 license. For any other use, please contact Secret Labs # AB (info@pythonware.com). # # Portions of this engine have been developed in cooperation with # CNRI. Hewlett-Packard provided funding for 1.6 integration and # other compatibility work. # r"""Support for regular expressions (RE). This module provides regular expression matching operations similar to those found in Perl. It supports both 8-bit and Unicode strings; both the pattern and the strings being processed can contain null bytes and characters outside the US ASCII range. Regular expressions can contain both special and ordinary characters. Most ordinary characters, like "A", "a", or "0", are the simplest regular expressions; they simply match themselves. You can concatenate ordinary characters, so last matches the string 'last'. The special characters are: "." Matches any character except a newline. "^" Matches the start of the string. "$" Matches the end of the string or just before the newline at the end of the string. "*" Matches 0 or more (greedy) repetitions of the preceding RE. Greedy means that it will match as many repetitions as possible. "+" Matches 1 or more (greedy) repetitions of the preceding RE. "?" Matches 0 or 1 (greedy) of the preceding RE. *?,+?,?? Non-greedy versions of the previous three special characters. {m,n} Matches from m to n repetitions of the preceding RE. {m,n}? Non-greedy version of the above. "\\" Either escapes special characters or signals a special sequence. [] Indicates a set of characters. A "^" as the first character indicates a complementing set. "|" A|B, creates an RE that will match either A or B. (...) Matches the RE inside the parentheses. The contents can be retrieved or matched later in the string. (?iLmsux) Set the I, L, M, S, U, or X flag for the RE (see below). (?:...) Non-grouping version of regular parentheses. (?P<name>...) The substring matched by the group is accessible by name. (?P=name) Matches the text matched earlier by the group named name. (?#...) A comment; ignored. (?=...) Matches if ... matches next, but doesn't consume the string. (?!...) Matches if ... doesn't match next. (?<=...) Matches if preceded by ... (must be fixed length). (?<!...) Matches if not preceded by ... (must be fixed length). (?(id/name)yes|no) Matches yes pattern if the group with id/name matched, the (optional) no pattern otherwise. The special sequences consist of "\\" and a character from the list below. If the ordinary character is not on the list, then the resulting RE will match the second character. \number Matches the contents of the group of the same number. \A Matches only at the start of the string. \Z Matches only at the end of the string. \b Matches the empty string, but only at the start or end of a word. \B Matches the empty string, but not at the start or end of a word. \d Matches any decimal digit; equivalent to the set [0-9]. \D Matches any non-digit character; equivalent to the set [^0-9]. \s Matches any whitespace character; equivalent to [ \t\n\r\f\v]. \S Matches any non-whitespace character; equiv. to [^ \t\n\r\f\v]. \w Matches any alphanumeric character; equivalent to [a-zA-Z0-9_]. With LOCALE, it will match the set [0-9_] plus characters defined as letters for the current locale. \W Matches the complement of \w. \\ Matches a literal backslash. This module exports the following functions: match Match a regular expression pattern to the beginning of a string. search Search a string for the presence of a pattern. sub Substitute occurrences of a pattern found in a string. subn Same as sub, but also return the number of substitutions made. split Split a string by the occurrences of a pattern. findall Find all occurrences of a pattern in a string. finditer Return an iterator yielding a match object for each match. compile Compile a pattern into a RegexObject. purge Clear the regular expression cache. escape Backslash all non-alphanumerics in a string. Some of the functions in this module takes flags as optional parameters: I IGNORECASE Perform case-insensitive matching. L LOCALE Make \w, \W, \b, \B, dependent on the current locale. M MULTILINE "^" matches the beginning of lines (after a newline) as well as the string. "$" matches the end of lines (before a newline) as well as the end of the string. S DOTALL "." matches any character at all, including the newline. X VERBOSE Ignore whitespace and comments for nicer looking RE's. U UNICODE Make \w, \W, \b, \B, dependent on the Unicode locale. This module also defines an exception 'error'. """ import sys import sre_compile import sre_parse try: import _locale except ImportError: _locale = None # public symbols __all__ = [ "match", "search", "sub", "subn", "split", "findall", "compile", "purge", "template", "escape", "I", "L", "M", "S", "X", "U", "IGNORECASE", "LOCALE", "MULTILINE", "DOTALL", "VERBOSE", "UNICODE", "error" ] __version__ = "2.2.1" # flags I = IGNORECASE = sre_compile.SRE_FLAG_IGNORECASE # ignore case L = LOCALE = sre_compile.SRE_FLAG_LOCALE # assume current 8-bit locale U = UNICODE = sre_compile.SRE_FLAG_UNICODE # assume unicode locale M = MULTILINE = sre_compile.SRE_FLAG_MULTILINE # make anchors look for newline S = DOTALL = sre_compile.SRE_FLAG_DOTALL # make dot match newline X = VERBOSE = sre_compile.SRE_FLAG_VERBOSE # ignore whitespace and comments # sre extensions (experimental, don't rely on these) T = TEMPLATE = sre_compile.SRE_FLAG_TEMPLATE # disable backtracking DEBUG = sre_compile.SRE_FLAG_DEBUG # dump pattern after compilation # sre exception error = sre_compile.error # -------------------------------------------------------------------- # public interface def match(pattern, string, flags=0): """Try to apply the pattern at the start of the string, returning a match object, or None if no match was found.""" return _compile(pattern, flags).match(string) def search(pattern, string, flags=0): """Scan through string looking for a match to the pattern, returning a match object, or None if no match was found.""" return _compile(pattern, flags).search(string) def sub(pattern, repl, string, count=0, flags=0): """Return the string obtained by replacing the leftmost non-overlapping occurrences of the pattern in string by the replacement repl. repl can be either a string or a callable; if a string, backslash escapes in it are processed. If it is a callable, it's passed the match object and must return a replacement string to be used.""" return _compile(pattern, flags).sub(repl, string, count) def subn(pattern, repl, string, count=0, flags=0): """Return a 2-tuple containing (new_string, number). new_string is the string obtained by replacing the leftmost non-overlapping occurrences of the pattern in the source string by the replacement repl. number is the number of substitutions that were made. repl can be either a string or a callable; if a string, backslash escapes in it are processed. If it is a callable, it's passed the match object and must return a replacement string to be used.""" return _compile(pattern, flags).subn(repl, string, count) def split(pattern, string, maxsplit=0, flags=0): """Split the source string by the occurrences of the pattern, returning a list containing the resulting substrings.""" return _compile(pattern, flags).split(string, maxsplit) def findall(pattern, string, flags=0): """Return a list of all non-overlapping matches in the string. If one or more groups are present in the pattern, return a list of groups; this will be a list of tuples if the pattern has more than one group. Empty matches are included in the result.""" return _compile(pattern, flags).findall(string) if sys.hexversion >= 0x02020000: __all__.append("finditer") def finditer(pattern, string, flags=0): """Return an iterator over all non-overlapping matches in the string. For each match, the iterator returns a match object. Empty matches are included in the result.""" return _compile(pattern, flags).finditer(string) def compile(pattern, flags=0): "Compile a regular expression pattern, returning a pattern object." return _compile(pattern, flags) def purge(): "Clear the regular expression cache" _cache.clear() _cache_repl.clear() def template(pattern, flags=0): "Compile a template pattern, returning a pattern object" return _compile(pattern, flags|T) _alphanum = frozenset( "abcdefghijklmnopqrstuvwxyzABCDEFGHIJKLMNOPQRSTUVWXYZ0123456789") def escape(pattern): "Escape all non-alphanumeric characters in pattern." s = list(pattern) alphanum = _alphanum for i, c in enumerate(pattern): if c not in alphanum: if c == "\000": s[i] = "\\000" else: s[i] = "\\" + c return pattern[:0].join(s) # -------------------------------------------------------------------- # internals _cache = {} _cache_repl = {} _pattern_type = type(sre_compile.compile("", 0)) _MAXCACHE = 100 def _compile(*key): # internal: compile pattern pattern, flags = key bypass_cache = flags & DEBUG if not bypass_cache: cachekey = (type(key[0]),) + key try: p, loc = _cache[cachekey] if loc is None or loc == _locale.setlocale(_locale.LC_CTYPE): return p except KeyError: pass if isinstance(pattern, _pattern_type): if flags: raise ValueError('Cannot process flags argument with a compiled pattern') return pattern if not sre_compile.isstring(pattern): raise TypeError, "first argument must be string or compiled pattern" try: p = sre_compile.compile(pattern, flags) except error, v: raise error, v # invalid expression if not bypass_cache: if len(_cache) >= _MAXCACHE: _cache.clear() if p.flags & LOCALE: if not _locale: return p loc = _locale.setlocale(_locale.LC_CTYPE) else: loc = None _cache[cachekey] = p, loc return p def _compile_repl(*key): # internal: compile replacement pattern p = _cache_repl.get(key) if p is not None: return p repl, pattern = key try: p = sre_parse.parse_template(repl, pattern) except error, v: raise error, v # invalid expression if len(_cache_repl) >= _MAXCACHE: _cache_repl.clear() _cache_repl[key] = p return p def _expand(pattern, match, template): # internal: match.expand implementation hook template = sre_parse.parse_template(template, pattern) return sre_parse.expand_template(template, match) def _subx(pattern, template): # internal: pattern.sub/subn implementation helper template = _compile_repl(template, pattern) if not template[0] and len(template[1]) == 1: # literal replacement return template[1][0] def filter(match, template=template): return sre_parse.expand_template(template, match) return filter # register myself for pickling import copy_reg def _pickle(p): return _compile, (p.pattern, p.flags) copy_reg.pickle(_pattern_type, _pickle, _compile) # -------------------------------------------------------------------- # experimental stuff (see python-dev discussions for details) class Scanner: def __init__(self, lexicon, flags=0): from sre_constants import BRANCH, SUBPATTERN self.lexicon = lexicon # combine phrases into a compound pattern p = [] s = sre_parse.Pattern() s.flags = flags for phrase, action in lexicon: p.append(sre_parse.SubPattern(s, [ (SUBPATTERN, (len(p)+1, sre_parse.parse(phrase, flags))), ])) s.groups = len(p)+1 p = sre_parse.SubPattern(s, [(BRANCH, (None, p))]) self.scanner = sre_compile.compile(p) def scan(self, string): result = [] append = result.append match = self.scanner.scanner(string).match i = 0 while 1: m = match() if not m: break j = m.end() if i == j: break action = self.lexicon[m.lastindex-1][1] if hasattr(action, '__call__'): self.match = m action = action(self, m.group()) if action is not None: append(action) i = j return result, string[i:]
williamroot/opps
refs/heads/master
opps/articles/admin.py
1
#!/usr/bin/env python # -*- coding: utf-8 -*- from django.contrib import admin from django.utils.translation import ugettext_lazy as _ from .models import Post, PostRelated, Album, Link from .forms import PostAdminForm, AlbumAdminForm, LinkAdminForm from opps.containers.admin import ( ContainerAdmin, ContainerImageInline, ContainerRelatedInline) from opps.core.admin import apply_opps_rules, HaystackModelAdmin from opps.core.permissions.admin import AdminViewPermission @apply_opps_rules('articles') class PostRelatedInline(admin.TabularInline): model = PostRelated fk_name = 'post' raw_id_fields = ['related'] actions = None ordering = ('order',) extra = 1 classes = ('collapse',) verbose_name = _(u'Related post') verbose_name_plural = _(u'Related posts') @apply_opps_rules('articles') class PostAdmin(HaystackModelAdmin, ContainerAdmin, AdminViewPermission): form = PostAdminForm inlines = [ContainerImageInline, ContainerRelatedInline] search_fields = ['title', 'headline', 'slug', 'channel_name'] raw_id_fields = ['main_image', 'channel', 'mirror_channel', 'albums'] ordering = ('-date_available',) fieldsets = ( (_(u'Identification'), { 'fields': ('site', 'title', 'slug', 'get_http_absolute_url', 'short_url')}), (_(u'Content'), { 'fields': ('hat', 'short_title', 'headline', 'content', ('main_image', 'main_image_caption', 'image_thumb'), 'source', 'tags')}), (_(u'Custom'), { 'fields': ('json',)}), (_(u'Relationships'), { 'fields': ('channel', 'mirror_channel', 'albums',)}), (_(u'Publication'), { 'classes': ('extrapretty'), 'fields': ('published', 'date_available', 'show_on_root_channel', 'in_containerboxes', 'feed_publishable') }), ) @apply_opps_rules('articles') class AlbumAdmin(HaystackModelAdmin, ContainerAdmin, AdminViewPermission): form = AlbumAdminForm inlines = [ContainerImageInline, ContainerRelatedInline] list_display = ['title', 'channel', 'images_count', 'date_available', 'published', 'preview_url'] fieldsets = ( (_(u'Identification'), { 'fields': ('site', 'title', 'slug', 'get_http_absolute_url', 'short_url',)}), (_(u'Content'), { 'fields': ('hat', 'short_title', 'headline', ('main_image', 'main_image_caption', 'image_thumb'), 'tags')}), (_(u'Custom'), { 'fields': ('json',)}), (_(u'Relationships'), { 'fields': ('channel', 'mirror_channel')}), (_(u'Publication'), { 'classes': ('extrapretty'), 'fields': ('published', 'date_available', 'show_on_root_channel', 'feed_publishable')}), ) @apply_opps_rules('articles') class LinkAdmin(ContainerAdmin, AdminViewPermission): form = LinkAdminForm raw_id_fields = ['container', 'channel', 'mirror_channel', 'main_image'] fieldsets = ( (_(u'Identification'), { 'fields': ('site', 'title', 'slug', 'get_http_absolute_url', 'short_url',)}), (_(u'Content'), { 'fields': ('hat', 'short_title', 'headline', 'url', 'container', ('main_image', 'main_image_caption', 'image_thumb'), 'tags')}), (_(u'Custom'), { 'fields': ('json',)}), (_(u'Relationships'), { 'fields': ('channel', 'mirror_channel')}), (_(u'Publication'), { 'classes': ('extrapretty'), 'fields': ('published', 'date_available', 'show_on_root_channel')}), ) admin.site.register(Post, PostAdmin) admin.site.register(Album, AlbumAdmin) admin.site.register(Link, LinkAdmin)
ar7z1/ansible
refs/heads/devel
test/units/modules/net_tools/nios/test_nios_srv_record.py
68
# This file is part of Ansible # # Ansible is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # Ansible is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with Ansible. If not, see <http://www.gnu.org/licenses/>. # Make coding more python3-ish from __future__ import (absolute_import, division, print_function) __metaclass__ = type from ansible.modules.net_tools.nios import nios_srv_record from ansible.module_utils.net_tools.nios import api from units.compat.mock import patch, MagicMock, Mock from .test_nios_module import TestNiosModule, load_fixture class TestNiosSRVRecordModule(TestNiosModule): module = nios_srv_record def setUp(self): super(TestNiosSRVRecordModule, self).setUp() self.module = MagicMock(name='ansible.modules.net_tools.nios.nios_srv_record.WapiModule') self.module.check_mode = False self.module.params = {'provider': None} self.mock_wapi = patch('ansible.modules.net_tools.nios.nios_srv_record.WapiModule') self.exec_command = self.mock_wapi.start() self.mock_wapi_run = patch('ansible.modules.net_tools.nios.nios_srv_record.WapiModule.run') self.mock_wapi_run.start() self.load_config = self.mock_wapi_run.start() def tearDown(self): super(TestNiosSRVRecordModule, self).tearDown() self.mock_wapi.stop() self.mock_wapi_run.stop() def _get_wapi(self, test_object): wapi = api.WapiModule(self.module) wapi.get_object = Mock(name='get_object', return_value=test_object) wapi.create_object = Mock(name='create_object') wapi.update_object = Mock(name='update_object') wapi.delete_object = Mock(name='delete_object') return wapi def load_fixtures(self, commands=None): self.exec_command.return_value = (0, load_fixture('nios_result.txt').strip(), None) self.load_config.return_value = dict(diff=None, session='session') def test_nios_srv_record_create(self): self.module.params = {'provider': None, 'state': 'present', 'name': '_sip._tcp.service.ansible.com', 'port': 5080, 'target': 'service1.ansible.com', 'priority': 10, 'weight': 10, 'comment': None, 'extattrs': None} test_object = None test_spec = { "name": {"ib_req": True}, "port": {"ib_req": True}, "target": {"ib_req": True}, "priority": {"ib_req": True}, "weight": {"ib_req": True}, "comment": {}, "extattrs": {} } wapi = self._get_wapi(test_object) print("WAPI: ", wapi) res = wapi.run('testobject', test_spec) self.assertTrue(res['changed']) wapi.create_object.assert_called_once_with('testobject', {'name': self.module._check_type_dict().__getitem__(), 'port': 5080, 'target': 'service1.ansible.com', 'priority': 10, 'weight': 10}) def test_nios_srv_record_update_comment(self): self.module.params = {'provider': None, 'state': 'present', 'name': '_sip._tcp.service.ansible.com', 'port': 5080, 'target': 'service1.ansible.com', 'priority': 10, 'weight': 10, 'comment': None, 'extattrs': None} test_object = [ { "comment": "test comment", "_ref": "srvrecord/ZG5zLm5ldHdvcmtfdmlldyQw:default/true", "name": "_sip._tcp.service.ansible.com", 'port': 5080, "target": "mailhost.ansible.com", "priority": 10, 'weight': 10, "extattrs": {} } ] test_spec = { "name": {"ib_req": True}, "port": {"ib_req": True}, "target": {"ib_req": True}, "priority": {"ib_req": True}, "weight": {"ib_req": True}, "comment": {}, "extattrs": {} } wapi = self._get_wapi(test_object) res = wapi.run('testobject', test_spec) self.assertTrue(res['changed']) def test_nios_srv_record_remove(self): self.module.params = {'provider': None, 'state': 'absent', 'name': '_sip._tcp.service.ansible.com', 'port': 5080, 'target': 'service1.ansible.com', 'priority': 10, 'weight': 10, 'comment': None, 'extattrs': None} ref = "srvrecord/ZG5zLm5ldHdvcmtfdmlldyQw:default/false" test_object = [ { "comment": "test comment", "_ref": ref, "name": "_sip._tcp.service.ansible.com", "port": 5080, "target": "mailhost.ansible.com", "priority": 10, "weight": 10, "extattrs": {'Site': {'value': 'test'}} } ] test_spec = { "name": {"ib_req": True}, "port": {"ib_req": True}, "target": {"ib_req": True}, "priority": {"ib_req": True}, "weight": {"ib_req": True}, "comment": {}, "extattrs": {} } wapi = self._get_wapi(test_object) res = wapi.run('testobject', test_spec) self.assertTrue(res['changed']) wapi.delete_object.assert_called_once_with(ref)
pavlovml/tensorflow
refs/heads/master
tensorflow/python/kernel_tests/scatter_ops_test.py
5
"""Tests for tensorflow.ops.tf.scatter.""" from __future__ import absolute_import from __future__ import division from __future__ import print_function import tensorflow.python.platform import numpy as np import tensorflow as tf class ScatterTest(tf.test.TestCase): def _VariableRankTest(self, np_scatter, tf_scatter): np.random.seed(8) with self.test_session(): for indices_shape in (), (2,), (2, 3), (2, 3, 4): for extra_shape in (), (5,), (5, 6): # Generate random indices with no duplicates for easy numpy comparison size = np.prod(indices_shape, dtype=np.int32) indices = np.arange(2 * size) np.random.shuffle(indices) indices = indices[:size].reshape(indices_shape) updates = np.random.randn(*(indices_shape + extra_shape)) old = np.random.randn(*((2 * size,) + extra_shape)) # Scatter via numpy new = old.copy() np_scatter(new, indices, updates) # Scatter via tensorflow ref = tf.Variable(old) ref.initializer.run() tf_scatter(ref, indices, updates).eval() # Compare self.assertAllClose(ref.eval(), new) def testVariableRankUpdate(self): def update(ref, indices, updates): ref[indices] = updates self._VariableRankTest(update, tf.scatter_update) def testVariableRankAdd(self): def add(ref, indices, updates): ref[indices] += updates self._VariableRankTest(add, tf.scatter_add) def testVariableRankSub(self): def sub(ref, indices, updates): ref[indices] -= updates self._VariableRankTest(sub, tf.scatter_sub) if __name__ == "__main__": tf.test.main()
dingmingliu/quanttrade
refs/heads/master
quanttrade/__init__.py
5
__author__ = 'tyler'
Relrin/aiorest-ws
refs/heads/master
tests/db/orm/django/test_validators.py
1
# -*- coding: utf-8 -*- from django.db import models from aiorest_ws.db.orm.exceptions import ValidationError from aiorest_ws.db.orm.django.serializers import ModelSerializer from aiorest_ws.db.orm.django.validators import qs_filter, qs_exists, \ UniqueValidator from tests.db.orm.django.base import DjangoUnitTest class TestQuerysetFunctions(DjangoUnitTest): class User(models.Model): name = models.CharField(max_length=30) class Meta: app_label = 'test_django_queryset_functions' def __str__(self): return '<User(%s)>' % self.name apps = ('test_django_queryset_functions',) models = (User, ) @classmethod def setUpClass(cls): super(TestQuerysetFunctions, cls).setUpClass() cls.User.objects.create(name='admin') def test_qs_exists_returns_true(self): queryset = self.User.objects.filter(name='admin') self.assertTrue(qs_exists(queryset)) def test_qs_exists_returns_false(self): queryset = self.User.objects.filter(name='nonexistent_user') self.assertFalse(qs_exists(queryset)) def test_qs_exists_returns_false_for_invalid_type(self): class InvalidType(object): def exists(self): raise TypeError() obj = InvalidType() self.assertFalse(qs_exists(obj)) def test_qs_filter_returns_object(self): user = self.User.objects.get(name='admin') queryset = self.User.objects.all() result = qs_filter(queryset, name='admin') self.assertEqual(len(result), 1) self.assertEqual(result[0], user) def test_qs_filter_returns_none_for_invalid_queryset(self): class InvalidQueryset(object): none_method_is_called = False def filter(self, **kwargs): raise ValueError() def none(self): self.none_method_is_called = True return None obj = InvalidQueryset() self.assertIsNone(qs_filter(obj)) self.assertTrue(obj.none_method_is_called) class TestUniqueValidator(DjangoUnitTest): class User(models.Model): name = models.CharField(max_length=30, unique=True) class Meta: app_label = 'test_django_unique_validator' def __str__(self): return '<User(%s)>' % self.name apps = ('test_django_queryset_functions', ) models = (User, ) @classmethod def setUpClass(cls): super(TestUniqueValidator, cls).setUpClass() cls.User.objects.create(name='admin') cls.User.objects.create(name='manager') cls.User.objects.create(name='author') class UserSerializer(ModelSerializer): class Meta: model = cls.User cls.user_serializer = UserSerializer def test_set_context(self): queryset = self.User.objects.all() serializer = self.user_serializer() validator = UniqueValidator(queryset) validator.set_context(serializer.fields['name']) self.assertEqual(validator.field_name, 'name') self.assertIsNone(validator.instance) def test_filter_queryset(self): queryset = self.User.objects.all() serializer = self.user_serializer() validator = UniqueValidator(queryset) validator.set_context(serializer.fields['name']) filtered_queryset = validator.filter_queryset('admin', queryset) self.assertEqual(len(filtered_queryset), 1) self.assertIsInstance(filtered_queryset[0], self.User) self.assertEqual(filtered_queryset[0].name, 'admin') def test_filter_queryset_returns_empty_queryset(self): queryset = self.User.objects.all() serializer = self.user_serializer() validator = UniqueValidator(queryset) validator.set_context(serializer.fields['name']) filtered_queryset = validator.filter_queryset('anonymous', queryset) self.assertEqual(len(filtered_queryset), 0) def test_exclude_current_instance_returns_queryset(self): queryset = self.User.objects.all() serializer = self.user_serializer() validator = UniqueValidator(queryset) validator.set_context(serializer.fields['name']) self.assertEqual( validator.exclude_current_instance(queryset), queryset ) def test_exclude_current_instance_returns_queryset_with_exclude(self): admin = self.User.objects.get(name='admin') queryset = self.User.objects.all() serializer = self.user_serializer() validator = UniqueValidator(queryset) serializer.fields['name'].parent.instance = admin validator.set_context(serializer.fields['name']) filtered_queryset = validator.exclude_current_instance(queryset) self.assertNotIn(queryset, filtered_queryset) def test_call_passed_successfully(self): queryset = self.User.objects.all() serializer = self.user_serializer() validator = UniqueValidator(queryset) validator.set_context(serializer.fields['name']) self.assertIsNone(validator('unnamed')) def test_call_raises_validation_error(self): queryset = self.User.objects.all() serializer = self.user_serializer() validator = UniqueValidator(queryset) validator.set_context(serializer.fields['name']) self.assertRaises(ValidationError, validator, 'admin') def test_repr(self): queryset = self.User.objects.all().order_by('id') serializer = self.user_serializer() validator = UniqueValidator(queryset) validator.set_context(serializer.fields['name']) self.assertEqual( validator.__repr__(), "'<UniqueValidator(queryset=<QuerySet [<User: <User(admin)>>, " "<User: <User(manager)>>, <User: <User(author)>>]>)>'" )
catapult-project/catapult
refs/heads/master
third_party/gsutil/third_party/apitools/apitools/base/py/testing/__init__.py
38
# # Copyright 2015 Google Inc. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Package marker file."""
ruchee/vimrc
refs/heads/master
vimfiles/bundle/vim-python/submodules/pylint/tests/functional/n/no/no_else_continue.py
2
""" Test that superfluous else return are detected. """ # pylint:disable=invalid-name,missing-docstring,unused-variable def foo1(x, y, z): for i in x: if i < y: # [no-else-continue] continue else: a = z def foo2(x, y, w, z): for i in x: if i < y: # [no-else-continue] continue elif i < w: continue else: a = z def foo3(x, y, z): for i in x: if i < y: a = 1 if z: # [no-else-continue] b = 2 continue else: c = 3 continue else: d = 4 continue def foo4(x, y): for i in x: if i < x: # [no-else-continue] if y: a = 4 else: b = 2 continue else: c = 3 continue def foo5(x, y, z): for i in x: if i < y: # [no-else-continue] if y: a = 4 else: b = 2 continue elif z: c = 2 else: c = 3 continue def foo6(x, y): for i in x: if i < x: if y: # [no-else-continue] a = 4 continue else: b = 2 else: c = 3 continue def bar1(x, y, z): for i in x: if i < y: continue return z def bar2(w, x, y, z): for i in w: if i < x: continue if z < i: a = y else: continue return def bar3(x, y, z): for i in x: if i < y: if z: continue else: return z return None def bar4(x): for i in range(10): if x: # [no-else-continue] continue else: try: return except ValueError: continue def nested1(x, y, z): for i in x: if i < x: for j in y: if j < z: continue else: a = z def nested2(x, y, z): for i in x: if i < x: for j in y: if j < z: continue continue else: a = z
ray-project/ray
refs/heads/master
rllib/agents/ppo/appo.py
2
""" Asynchronous Proximal Policy Optimization (APPO) ================================================ This file defines the distributed Trainer class for the asynchronous version of proximal policy optimization (APPO). See `appo_[tf|torch]_policy.py` for the definition of the policy loss. Detailed documentation: https://docs.ray.io/en/master/rllib-algorithms.html#appo """ from typing import Optional, Type from ray.rllib.agents.trainer import Trainer from ray.rllib.agents.impala.impala import validate_config from ray.rllib.agents.ppo.appo_tf_policy import AsyncPPOTFPolicy from ray.rllib.agents.ppo.ppo import UpdateKL from ray.rllib.agents import impala from ray.rllib.policy.policy import Policy from ray.rllib.execution.common import STEPS_SAMPLED_COUNTER, \ LAST_TARGET_UPDATE_TS, NUM_TARGET_UPDATES, _get_shared_metrics from ray.rllib.utils.typing import TrainerConfigDict # yapf: disable # __sphinx_doc_begin__ # Adds the following updates to the `IMPALATrainer` config in # rllib/agents/impala/impala.py. DEFAULT_CONFIG = impala.ImpalaTrainer.merge_trainer_configs( impala.DEFAULT_CONFIG, # See keys in impala.py, which are also supported. { # Whether to use V-trace weighted advantages. If false, PPO GAE # advantages will be used instead. "vtrace": True, # == These two options only apply if vtrace: False == # Should use a critic as a baseline (otherwise don't use value # baseline; required for using GAE). "use_critic": True, # If true, use the Generalized Advantage Estimator (GAE) # with a value function, see https://arxiv.org/pdf/1506.02438.pdf. "use_gae": True, # GAE(lambda) parameter "lambda": 1.0, # == PPO surrogate loss options == "clip_param": 0.4, # == PPO KL Loss options == "use_kl_loss": False, "kl_coeff": 1.0, "kl_target": 0.01, # == IMPALA optimizer params (see documentation in impala.py) == "rollout_fragment_length": 50, "train_batch_size": 500, "min_iter_time_s": 10, "num_workers": 2, "num_gpus": 0, "num_data_loader_buffers": 1, "minibatch_buffer_size": 1, "num_sgd_iter": 1, "replay_proportion": 0.0, "replay_buffer_num_slots": 100, "learner_queue_size": 16, "learner_queue_timeout": 300, "max_sample_requests_in_flight_per_worker": 2, "broadcast_interval": 1, "grad_clip": 40.0, "opt_type": "adam", "lr": 0.0005, "lr_schedule": None, "decay": 0.99, "momentum": 0.0, "epsilon": 0.1, "vf_loss_coeff": 0.5, "entropy_coeff": 0.01, "entropy_coeff_schedule": None, }, _allow_unknown_configs=True, ) # __sphinx_doc_end__ # yapf: enable class UpdateTargetAndKL: def __init__(self, workers, config): self.workers = workers self.config = config self.update_kl = UpdateKL(workers) self.target_update_freq = config["num_sgd_iter"] \ * config["minibatch_buffer_size"] def __call__(self, fetches): metrics = _get_shared_metrics() cur_ts = metrics.counters[STEPS_SAMPLED_COUNTER] last_update = metrics.counters[LAST_TARGET_UPDATE_TS] if cur_ts - last_update > self.target_update_freq: metrics.counters[NUM_TARGET_UPDATES] += 1 metrics.counters[LAST_TARGET_UPDATE_TS] = cur_ts # Update Target Network self.workers.local_worker().foreach_trainable_policy( lambda p, _: p.update_target()) # Also update KL Coeff if self.config["use_kl_loss"]: self.update_kl(fetches) def add_target_callback(config: TrainerConfigDict): """Add the update target and kl hook. This hook is called explicitly after each learner step in the execution setup for IMPALA. Args: config (TrainerConfigDict): The APPO config dict. """ config["after_train_step"] = UpdateTargetAndKL validate_config(config) def get_policy_class(config: TrainerConfigDict) -> Optional[Type[Policy]]: """Policy class picker function. Class is chosen based on DL-framework. Args: config (TrainerConfigDict): The trainer's configuration dict. Returns: Optional[Type[Policy]]: The Policy class to use with PPOTrainer. If None, use `default_policy` provided in build_trainer(). """ if config["framework"] == "torch": from ray.rllib.agents.ppo.appo_torch_policy import AsyncPPOTorchPolicy return AsyncPPOTorchPolicy def initialize_target(trainer: Trainer) -> None: """Updates target network on startup by synching it with the policy net. Args: trainer (Trainer): The Trainer object. """ trainer.workers.local_worker().foreach_trainable_policy( lambda p, _: p.update_target()) # Build a child class of `Trainer`, based on ImpalaTrainer's setup. # Note: The generated class is NOT a sub-class of ImpalaTrainer, but directly # of the `Trainer` class. APPOTrainer = impala.ImpalaTrainer.with_updates( name="APPO", default_config=DEFAULT_CONFIG, validate_config=add_target_callback, default_policy=AsyncPPOTFPolicy, get_policy_class=get_policy_class, after_init=initialize_target)
chaowyc/youtube-dl
refs/heads/master
youtube_dl/extractor/rtp.py
99
# coding: utf-8 from __future__ import unicode_literals import re from .common import InfoExtractor class RTPIE(InfoExtractor): _VALID_URL = r'https?://(?:www\.)?rtp\.pt/play/p(?P<program_id>[0-9]+)/(?P<id>[^/?#]+)/?' _TESTS = [{ 'url': 'http://www.rtp.pt/play/p405/e174042/paixoes-cruzadas', 'md5': 'e736ce0c665e459ddb818546220b4ef8', 'info_dict': { 'id': 'e174042', 'ext': 'mp3', 'title': 'Paixões Cruzadas', 'description': 'As paixões musicais de António Cartaxo e António Macedo', 'thumbnail': 're:^https?://.*\.jpg', }, 'params': { # rtmp download 'skip_download': True, }, }, { 'url': 'http://www.rtp.pt/play/p831/a-quimica-das-coisas', 'only_matching': True, }] def _real_extract(self, url): video_id = self._match_id(url) webpage = self._download_webpage(url, video_id) title = self._html_search_meta( 'twitter:title', webpage, display_name='title', fatal=True) description = self._html_search_meta('description', webpage) thumbnail = self._og_search_thumbnail(webpage) player_config = self._search_regex( r'(?s)RTPPLAY\.player\.newPlayer\(\s*(\{.*?\})\s*\)', webpage, 'player config') config = self._parse_json(player_config, video_id) path, ext = config.get('file').rsplit('.', 1) formats = [{ 'format_id': 'rtmp', 'ext': ext, 'vcodec': config.get('type') == 'audio' and 'none' or None, 'preference': -2, 'url': 'rtmp://{streamer:s}/{application:s}'.format(**config), 'app': config.get('application'), 'play_path': '{ext:s}:{path:s}'.format(ext=ext, path=path), 'page_url': url, 'rtmp_live': config.get('live', False), 'player_url': 'http://programas.rtp.pt/play/player.swf?v3', 'rtmp_real_time': True, }] # Construct regular HTTP download URLs replacements = { 'audio': { 'format_id': 'mp3', 'pattern': r'^nas2\.share/wavrss/', 'repl': 'http://rsspod.rtp.pt/podcasts/', 'vcodec': 'none', }, 'video': { 'format_id': 'mp4_h264', 'pattern': r'^nas2\.share/h264/', 'repl': 'http://rsspod.rtp.pt/videocasts/', 'vcodec': 'h264', }, } r = replacements[config['type']] if re.match(r['pattern'], config['file']) is not None: formats.append({ 'format_id': r['format_id'], 'url': re.sub(r['pattern'], r['repl'], config['file']), 'vcodec': r['vcodec'], }) self._sort_formats(formats) return { 'id': video_id, 'title': title, 'formats': formats, 'description': description, 'thumbnail': thumbnail, }
walkingmu/InstantVideoConverter
refs/heads/master
spatial-media-2.0/spatialmedia/mpeg/mpeg4_container.py
2
#! /usr/bin/env python # -*- coding: utf-8 -*- # Copyright 2016 Google Inc. All rights reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """MPEG4 processing classes. Functions for loading MP4/MOV files and manipulating boxes. """ from spatialmedia.mpeg import box from spatialmedia.mpeg import constants from spatialmedia.mpeg import container def load(fh): """Load the mpeg4 file structure of a file. Args: fh: file handle, input file handle. position: int, current file position. size: int, maximum size. This is used to ensure correct box sizes. return: mpeg4, the loaded mpeg4 structure. """ fh.seek(0, 2) size = fh.tell() contents = container.load_multiple(fh, 0, size) if not contents: print "Error, failed to load .mp4 file." return None elif len(contents) == 0: print ("Error, no boxes found.") return None loaded_mpeg4 = Mpeg4Container() loaded_mpeg4.contents = contents for element in loaded_mpeg4.contents: if (element.name == "moov"): loaded_mpeg4.moov_box = element if (element.name == "free"): loaded_mpeg4.free_box = element if (element.name == "mdat" and not loaded_mpeg4.first_mdat_box): loaded_mpeg4.first_mdat_box = element if (element.name == "ftyp"): loaded_mpeg4.ftyp_box = element if not loaded_mpeg4.moov_box: print ("Error, file does not contain moov box.") return None if not loaded_mpeg4.first_mdat_box: print ("Error, file does not contain mdat box.") return None loaded_mpeg4.first_mdat_position = \ loaded_mpeg4.first_mdat_box.position loaded_mpeg4.first_mdat_position += \ loaded_mpeg4.first_mdat_box.header_size loaded_mpeg4.content_size = 0 for element in loaded_mpeg4.contents: loaded_mpeg4.content_size += element.size() return loaded_mpeg4 class Mpeg4Container(container.Container): """Specialized behaviour for the root mpeg4 container.""" def __init__(self): self.contents = list() self.content_size = 0 self.header_size = 0 self.moov_box = None self.free_box = None self.first_mdat_box = None self.ftyp_box = None self.first_mdat_position = None self.padding = 0 def merge(self, element): """Mpeg4 containers do not support merging.""" print "Cannot merge mpeg4 files" exit(0) def print_structure(self): """Print mpeg4 file structure recursively.""" print "mpeg4 [", self.content_size, "]" size = len(self.contents) for i in range(size): next_indent = " ├──" if i == (size - 1): next_indent = " └──" self.contents[i].print_structure(next_indent) def save(self, in_fh, out_fh): """Save mpeg4 filecontent to file. Args: in_fh: file handle, source file handle for uncached contents. out_fh: file handle, destination file hand for saved file. """ self.resize() new_position = 0 for element in self.contents: if element.name == constants.TAG_MDAT: new_position += element.header_size break new_position += element.size() delta = new_position - self.first_mdat_position for element in self.contents: element.save(in_fh, out_fh, delta)
emceethemouth/kernel_msm
refs/heads/master
tools/perf/scripts/python/netdev-times.py
11271
# Display a process of packets and processed time. # It helps us to investigate networking or network device. # # options # tx: show only tx chart # rx: show only rx chart # dev=: show only thing related to specified device # debug: work with debug mode. It shows buffer status. import os import sys sys.path.append(os.environ['PERF_EXEC_PATH'] + \ '/scripts/python/Perf-Trace-Util/lib/Perf/Trace') from perf_trace_context import * from Core import * from Util import * all_event_list = []; # insert all tracepoint event related with this script irq_dic = {}; # key is cpu and value is a list which stacks irqs # which raise NET_RX softirq net_rx_dic = {}; # key is cpu and value include time of NET_RX softirq-entry # and a list which stacks receive receive_hunk_list = []; # a list which include a sequence of receive events rx_skb_list = []; # received packet list for matching # skb_copy_datagram_iovec buffer_budget = 65536; # the budget of rx_skb_list, tx_queue_list and # tx_xmit_list of_count_rx_skb_list = 0; # overflow count tx_queue_list = []; # list of packets which pass through dev_queue_xmit of_count_tx_queue_list = 0; # overflow count tx_xmit_list = []; # list of packets which pass through dev_hard_start_xmit of_count_tx_xmit_list = 0; # overflow count tx_free_list = []; # list of packets which is freed # options show_tx = 0; show_rx = 0; dev = 0; # store a name of device specified by option "dev=" debug = 0; # indices of event_info tuple EINFO_IDX_NAME= 0 EINFO_IDX_CONTEXT=1 EINFO_IDX_CPU= 2 EINFO_IDX_TIME= 3 EINFO_IDX_PID= 4 EINFO_IDX_COMM= 5 # Calculate a time interval(msec) from src(nsec) to dst(nsec) def diff_msec(src, dst): return (dst - src) / 1000000.0 # Display a process of transmitting a packet def print_transmit(hunk): if dev != 0 and hunk['dev'].find(dev) < 0: return print "%7s %5d %6d.%06dsec %12.3fmsec %12.3fmsec" % \ (hunk['dev'], hunk['len'], nsecs_secs(hunk['queue_t']), nsecs_nsecs(hunk['queue_t'])/1000, diff_msec(hunk['queue_t'], hunk['xmit_t']), diff_msec(hunk['xmit_t'], hunk['free_t'])) # Format for displaying rx packet processing PF_IRQ_ENTRY= " irq_entry(+%.3fmsec irq=%d:%s)" PF_SOFT_ENTRY=" softirq_entry(+%.3fmsec)" PF_NAPI_POLL= " napi_poll_exit(+%.3fmsec %s)" PF_JOINT= " |" PF_WJOINT= " | |" PF_NET_RECV= " |---netif_receive_skb(+%.3fmsec skb=%x len=%d)" PF_NET_RX= " |---netif_rx(+%.3fmsec skb=%x)" PF_CPY_DGRAM= " | skb_copy_datagram_iovec(+%.3fmsec %d:%s)" PF_KFREE_SKB= " | kfree_skb(+%.3fmsec location=%x)" PF_CONS_SKB= " | consume_skb(+%.3fmsec)" # Display a process of received packets and interrputs associated with # a NET_RX softirq def print_receive(hunk): show_hunk = 0 irq_list = hunk['irq_list'] cpu = irq_list[0]['cpu'] base_t = irq_list[0]['irq_ent_t'] # check if this hunk should be showed if dev != 0: for i in range(len(irq_list)): if irq_list[i]['name'].find(dev) >= 0: show_hunk = 1 break else: show_hunk = 1 if show_hunk == 0: return print "%d.%06dsec cpu=%d" % \ (nsecs_secs(base_t), nsecs_nsecs(base_t)/1000, cpu) for i in range(len(irq_list)): print PF_IRQ_ENTRY % \ (diff_msec(base_t, irq_list[i]['irq_ent_t']), irq_list[i]['irq'], irq_list[i]['name']) print PF_JOINT irq_event_list = irq_list[i]['event_list'] for j in range(len(irq_event_list)): irq_event = irq_event_list[j] if irq_event['event'] == 'netif_rx': print PF_NET_RX % \ (diff_msec(base_t, irq_event['time']), irq_event['skbaddr']) print PF_JOINT print PF_SOFT_ENTRY % \ diff_msec(base_t, hunk['sirq_ent_t']) print PF_JOINT event_list = hunk['event_list'] for i in range(len(event_list)): event = event_list[i] if event['event_name'] == 'napi_poll': print PF_NAPI_POLL % \ (diff_msec(base_t, event['event_t']), event['dev']) if i == len(event_list) - 1: print "" else: print PF_JOINT else: print PF_NET_RECV % \ (diff_msec(base_t, event['event_t']), event['skbaddr'], event['len']) if 'comm' in event.keys(): print PF_WJOINT print PF_CPY_DGRAM % \ (diff_msec(base_t, event['comm_t']), event['pid'], event['comm']) elif 'handle' in event.keys(): print PF_WJOINT if event['handle'] == "kfree_skb": print PF_KFREE_SKB % \ (diff_msec(base_t, event['comm_t']), event['location']) elif event['handle'] == "consume_skb": print PF_CONS_SKB % \ diff_msec(base_t, event['comm_t']) print PF_JOINT def trace_begin(): global show_tx global show_rx global dev global debug for i in range(len(sys.argv)): if i == 0: continue arg = sys.argv[i] if arg == 'tx': show_tx = 1 elif arg =='rx': show_rx = 1 elif arg.find('dev=',0, 4) >= 0: dev = arg[4:] elif arg == 'debug': debug = 1 if show_tx == 0 and show_rx == 0: show_tx = 1 show_rx = 1 def trace_end(): # order all events in time all_event_list.sort(lambda a,b :cmp(a[EINFO_IDX_TIME], b[EINFO_IDX_TIME])) # process all events for i in range(len(all_event_list)): event_info = all_event_list[i] name = event_info[EINFO_IDX_NAME] if name == 'irq__softirq_exit': handle_irq_softirq_exit(event_info) elif name == 'irq__softirq_entry': handle_irq_softirq_entry(event_info) elif name == 'irq__softirq_raise': handle_irq_softirq_raise(event_info) elif name == 'irq__irq_handler_entry': handle_irq_handler_entry(event_info) elif name == 'irq__irq_handler_exit': handle_irq_handler_exit(event_info) elif name == 'napi__napi_poll': handle_napi_poll(event_info) elif name == 'net__netif_receive_skb': handle_netif_receive_skb(event_info) elif name == 'net__netif_rx': handle_netif_rx(event_info) elif name == 'skb__skb_copy_datagram_iovec': handle_skb_copy_datagram_iovec(event_info) elif name == 'net__net_dev_queue': handle_net_dev_queue(event_info) elif name == 'net__net_dev_xmit': handle_net_dev_xmit(event_info) elif name == 'skb__kfree_skb': handle_kfree_skb(event_info) elif name == 'skb__consume_skb': handle_consume_skb(event_info) # display receive hunks if show_rx: for i in range(len(receive_hunk_list)): print_receive(receive_hunk_list[i]) # display transmit hunks if show_tx: print " dev len Qdisc " \ " netdevice free" for i in range(len(tx_free_list)): print_transmit(tx_free_list[i]) if debug: print "debug buffer status" print "----------------------------" print "xmit Qdisc:remain:%d overflow:%d" % \ (len(tx_queue_list), of_count_tx_queue_list) print "xmit netdevice:remain:%d overflow:%d" % \ (len(tx_xmit_list), of_count_tx_xmit_list) print "receive:remain:%d overflow:%d" % \ (len(rx_skb_list), of_count_rx_skb_list) # called from perf, when it finds a correspoinding event def irq__softirq_entry(name, context, cpu, sec, nsec, pid, comm, vec): if symbol_str("irq__softirq_entry", "vec", vec) != "NET_RX": return event_info = (name, context, cpu, nsecs(sec, nsec), pid, comm, vec) all_event_list.append(event_info) def irq__softirq_exit(name, context, cpu, sec, nsec, pid, comm, vec): if symbol_str("irq__softirq_entry", "vec", vec) != "NET_RX": return event_info = (name, context, cpu, nsecs(sec, nsec), pid, comm, vec) all_event_list.append(event_info) def irq__softirq_raise(name, context, cpu, sec, nsec, pid, comm, vec): if symbol_str("irq__softirq_entry", "vec", vec) != "NET_RX": return event_info = (name, context, cpu, nsecs(sec, nsec), pid, comm, vec) all_event_list.append(event_info) def irq__irq_handler_entry(name, context, cpu, sec, nsec, pid, comm, irq, irq_name): event_info = (name, context, cpu, nsecs(sec, nsec), pid, comm, irq, irq_name) all_event_list.append(event_info) def irq__irq_handler_exit(name, context, cpu, sec, nsec, pid, comm, irq, ret): event_info = (name, context, cpu, nsecs(sec, nsec), pid, comm, irq, ret) all_event_list.append(event_info) def napi__napi_poll(name, context, cpu, sec, nsec, pid, comm, napi, dev_name): event_info = (name, context, cpu, nsecs(sec, nsec), pid, comm, napi, dev_name) all_event_list.append(event_info) def net__netif_receive_skb(name, context, cpu, sec, nsec, pid, comm, skbaddr, skblen, dev_name): event_info = (name, context, cpu, nsecs(sec, nsec), pid, comm, skbaddr, skblen, dev_name) all_event_list.append(event_info) def net__netif_rx(name, context, cpu, sec, nsec, pid, comm, skbaddr, skblen, dev_name): event_info = (name, context, cpu, nsecs(sec, nsec), pid, comm, skbaddr, skblen, dev_name) all_event_list.append(event_info) def net__net_dev_queue(name, context, cpu, sec, nsec, pid, comm, skbaddr, skblen, dev_name): event_info = (name, context, cpu, nsecs(sec, nsec), pid, comm, skbaddr, skblen, dev_name) all_event_list.append(event_info) def net__net_dev_xmit(name, context, cpu, sec, nsec, pid, comm, skbaddr, skblen, rc, dev_name): event_info = (name, context, cpu, nsecs(sec, nsec), pid, comm, skbaddr, skblen, rc ,dev_name) all_event_list.append(event_info) def skb__kfree_skb(name, context, cpu, sec, nsec, pid, comm, skbaddr, protocol, location): event_info = (name, context, cpu, nsecs(sec, nsec), pid, comm, skbaddr, protocol, location) all_event_list.append(event_info) def skb__consume_skb(name, context, cpu, sec, nsec, pid, comm, skbaddr): event_info = (name, context, cpu, nsecs(sec, nsec), pid, comm, skbaddr) all_event_list.append(event_info) def skb__skb_copy_datagram_iovec(name, context, cpu, sec, nsec, pid, comm, skbaddr, skblen): event_info = (name, context, cpu, nsecs(sec, nsec), pid, comm, skbaddr, skblen) all_event_list.append(event_info) def handle_irq_handler_entry(event_info): (name, context, cpu, time, pid, comm, irq, irq_name) = event_info if cpu not in irq_dic.keys(): irq_dic[cpu] = [] irq_record = {'irq':irq, 'name':irq_name, 'cpu':cpu, 'irq_ent_t':time} irq_dic[cpu].append(irq_record) def handle_irq_handler_exit(event_info): (name, context, cpu, time, pid, comm, irq, ret) = event_info if cpu not in irq_dic.keys(): return irq_record = irq_dic[cpu].pop() if irq != irq_record['irq']: return irq_record.update({'irq_ext_t':time}) # if an irq doesn't include NET_RX softirq, drop. if 'event_list' in irq_record.keys(): irq_dic[cpu].append(irq_record) def handle_irq_softirq_raise(event_info): (name, context, cpu, time, pid, comm, vec) = event_info if cpu not in irq_dic.keys() \ or len(irq_dic[cpu]) == 0: return irq_record = irq_dic[cpu].pop() if 'event_list' in irq_record.keys(): irq_event_list = irq_record['event_list'] else: irq_event_list = [] irq_event_list.append({'time':time, 'event':'sirq_raise'}) irq_record.update({'event_list':irq_event_list}) irq_dic[cpu].append(irq_record) def handle_irq_softirq_entry(event_info): (name, context, cpu, time, pid, comm, vec) = event_info net_rx_dic[cpu] = {'sirq_ent_t':time, 'event_list':[]} def handle_irq_softirq_exit(event_info): (name, context, cpu, time, pid, comm, vec) = event_info irq_list = [] event_list = 0 if cpu in irq_dic.keys(): irq_list = irq_dic[cpu] del irq_dic[cpu] if cpu in net_rx_dic.keys(): sirq_ent_t = net_rx_dic[cpu]['sirq_ent_t'] event_list = net_rx_dic[cpu]['event_list'] del net_rx_dic[cpu] if irq_list == [] or event_list == 0: return rec_data = {'sirq_ent_t':sirq_ent_t, 'sirq_ext_t':time, 'irq_list':irq_list, 'event_list':event_list} # merge information realted to a NET_RX softirq receive_hunk_list.append(rec_data) def handle_napi_poll(event_info): (name, context, cpu, time, pid, comm, napi, dev_name) = event_info if cpu in net_rx_dic.keys(): event_list = net_rx_dic[cpu]['event_list'] rec_data = {'event_name':'napi_poll', 'dev':dev_name, 'event_t':time} event_list.append(rec_data) def handle_netif_rx(event_info): (name, context, cpu, time, pid, comm, skbaddr, skblen, dev_name) = event_info if cpu not in irq_dic.keys() \ or len(irq_dic[cpu]) == 0: return irq_record = irq_dic[cpu].pop() if 'event_list' in irq_record.keys(): irq_event_list = irq_record['event_list'] else: irq_event_list = [] irq_event_list.append({'time':time, 'event':'netif_rx', 'skbaddr':skbaddr, 'skblen':skblen, 'dev_name':dev_name}) irq_record.update({'event_list':irq_event_list}) irq_dic[cpu].append(irq_record) def handle_netif_receive_skb(event_info): global of_count_rx_skb_list (name, context, cpu, time, pid, comm, skbaddr, skblen, dev_name) = event_info if cpu in net_rx_dic.keys(): rec_data = {'event_name':'netif_receive_skb', 'event_t':time, 'skbaddr':skbaddr, 'len':skblen} event_list = net_rx_dic[cpu]['event_list'] event_list.append(rec_data) rx_skb_list.insert(0, rec_data) if len(rx_skb_list) > buffer_budget: rx_skb_list.pop() of_count_rx_skb_list += 1 def handle_net_dev_queue(event_info): global of_count_tx_queue_list (name, context, cpu, time, pid, comm, skbaddr, skblen, dev_name) = event_info skb = {'dev':dev_name, 'skbaddr':skbaddr, 'len':skblen, 'queue_t':time} tx_queue_list.insert(0, skb) if len(tx_queue_list) > buffer_budget: tx_queue_list.pop() of_count_tx_queue_list += 1 def handle_net_dev_xmit(event_info): global of_count_tx_xmit_list (name, context, cpu, time, pid, comm, skbaddr, skblen, rc, dev_name) = event_info if rc == 0: # NETDEV_TX_OK for i in range(len(tx_queue_list)): skb = tx_queue_list[i] if skb['skbaddr'] == skbaddr: skb['xmit_t'] = time tx_xmit_list.insert(0, skb) del tx_queue_list[i] if len(tx_xmit_list) > buffer_budget: tx_xmit_list.pop() of_count_tx_xmit_list += 1 return def handle_kfree_skb(event_info): (name, context, cpu, time, pid, comm, skbaddr, protocol, location) = event_info for i in range(len(tx_queue_list)): skb = tx_queue_list[i] if skb['skbaddr'] == skbaddr: del tx_queue_list[i] return for i in range(len(tx_xmit_list)): skb = tx_xmit_list[i] if skb['skbaddr'] == skbaddr: skb['free_t'] = time tx_free_list.append(skb) del tx_xmit_list[i] return for i in range(len(rx_skb_list)): rec_data = rx_skb_list[i] if rec_data['skbaddr'] == skbaddr: rec_data.update({'handle':"kfree_skb", 'comm':comm, 'pid':pid, 'comm_t':time}) del rx_skb_list[i] return def handle_consume_skb(event_info): (name, context, cpu, time, pid, comm, skbaddr) = event_info for i in range(len(tx_xmit_list)): skb = tx_xmit_list[i] if skb['skbaddr'] == skbaddr: skb['free_t'] = time tx_free_list.append(skb) del tx_xmit_list[i] return def handle_skb_copy_datagram_iovec(event_info): (name, context, cpu, time, pid, comm, skbaddr, skblen) = event_info for i in range(len(rx_skb_list)): rec_data = rx_skb_list[i] if skbaddr == rec_data['skbaddr']: rec_data.update({'handle':"skb_copy_datagram_iovec", 'comm':comm, 'pid':pid, 'comm_t':time}) del rx_skb_list[i] return
yamila-moreno/django
refs/heads/master
tests/i18n/tests.py
11
# -*- encoding: utf-8 -*- from __future__ import unicode_literals import datetime import decimal import gettext as gettext_module import os import pickle from contextlib import contextmanager from importlib import import_module from threading import local from unittest import skipUnless from django import forms from django.conf import settings from django.template import Context, Template, TemplateSyntaxError from django.test import ( RequestFactory, SimpleTestCase, TestCase, override_settings, ) from django.utils import six, translation from django.utils._os import upath from django.utils.formats import ( date_format, get_format, get_format_modules, iter_format_modules, localize, localize_input, reset_format_cache, sanitize_separators, time_format, ) from django.utils.numberformat import format as nformat from django.utils.safestring import SafeBytes, SafeString, SafeText, mark_safe from django.utils.six import PY3 from django.utils.translation import ( LANGUAGE_SESSION_KEY, activate, check_for_language, deactivate, get_language, get_language_bidi, get_language_from_request, get_language_info, gettext, gettext_lazy, ngettext_lazy, npgettext, npgettext_lazy, pgettext, pgettext_lazy, string_concat, to_locale, trans_real, ugettext, ugettext_lazy, ungettext, ungettext_lazy, ) from .forms import CompanyForm, I18nForm, SelectDateForm from .models import Company, TestModel here = os.path.dirname(os.path.abspath(upath(__file__))) extended_locale_paths = settings.LOCALE_PATHS + [ os.path.join(here, 'other', 'locale'), ] @contextmanager def patch_formats(lang, **settings): from django.utils.formats import _format_cache # Populate _format_cache with temporary values for key, value in settings.items(): _format_cache[(key, lang)] = value try: yield finally: reset_format_cache() class TranslationTests(SimpleTestCase): @translation.override('fr') def test_plural(self): """ Test plurals with ungettext. French differs from English in that 0 is singular. """ self.assertEqual(ungettext("%d year", "%d years", 0) % 0, "0 année") self.assertEqual(ungettext("%d year", "%d years", 2) % 2, "2 années") self.assertEqual(ungettext("%(size)d byte", "%(size)d bytes", 0) % {'size': 0}, "0 octet") self.assertEqual(ungettext("%(size)d byte", "%(size)d bytes", 2) % {'size': 2}, "2 octets") def test_override(self): activate('de') try: with translation.override('pl'): self.assertEqual(get_language(), 'pl') self.assertEqual(get_language(), 'de') with translation.override(None): self.assertEqual(get_language(), None) self.assertEqual(get_language(), 'de') finally: deactivate() def test_override_decorator(self): @translation.override('pl') def func_pl(): self.assertEqual(get_language(), 'pl') @translation.override(None) def func_none(): self.assertEqual(get_language(), None) try: activate('de') func_pl() self.assertEqual(get_language(), 'de') func_none() self.assertEqual(get_language(), 'de') finally: deactivate() def test_override_exit(self): """ Test that the language restored is the one used when the function was called, not the one used when the decorator was initialized. refs #23381 """ activate('fr') @translation.override('pl') def func_pl(): pass deactivate() try: activate('en') func_pl() self.assertEqual(get_language(), 'en') finally: deactivate() def test_lazy_objects(self): """ Format string interpolation should work with *_lazy objects. """ s = ugettext_lazy('Add %(name)s') d = {'name': 'Ringo'} self.assertEqual('Add Ringo', s % d) with translation.override('de', deactivate=True): self.assertEqual('Ringo hinzuf\xfcgen', s % d) with translation.override('pl'): self.assertEqual('Dodaj Ringo', s % d) # It should be possible to compare *_lazy objects. s1 = ugettext_lazy('Add %(name)s') self.assertEqual(s, s1) s2 = gettext_lazy('Add %(name)s') s3 = gettext_lazy('Add %(name)s') self.assertEqual(s2, s3) self.assertEqual(s, s2) s4 = ugettext_lazy('Some other string') self.assertNotEqual(s, s4) @skipUnless(six.PY2, "No more bytestring translations on PY3") def test_lazy_and_bytestrings(self): # On Python 2, (n)gettext_lazy should not transform a bytestring to unicode self.assertEqual(gettext_lazy(b"test").upper(), b"TEST") self.assertEqual((ngettext_lazy(b"%d test", b"%d tests") % 1).upper(), b"1 TEST") # Other versions of lazy functions always return unicode self.assertEqual(ugettext_lazy(b"test").upper(), "TEST") self.assertEqual((ungettext_lazy(b"%d test", b"%d tests") % 1).upper(), "1 TEST") self.assertEqual(pgettext_lazy(b"context", b"test").upper(), "TEST") self.assertEqual( (npgettext_lazy(b"context", b"%d test", b"%d tests") % 1).upper(), "1 TEST" ) def test_lazy_pickle(self): s1 = ugettext_lazy("test") self.assertEqual(six.text_type(s1), "test") s2 = pickle.loads(pickle.dumps(s1)) self.assertEqual(six.text_type(s2), "test") @override_settings(LOCALE_PATHS=extended_locale_paths) def test_ungettext_lazy(self): simple_with_format = ungettext_lazy('%d good result', '%d good results') simple_str_with_format = ngettext_lazy(str('%d good result'), str('%d good results')) simple_context_with_format = npgettext_lazy('Exclamation', '%d good result', '%d good results') simple_without_format = ungettext_lazy('good result', 'good results') with translation.override('de'): self.assertEqual(simple_with_format % 1, '1 gutes Resultat') self.assertEqual(simple_with_format % 4, '4 guten Resultate') self.assertEqual(simple_str_with_format % 1, str('1 gutes Resultat')) self.assertEqual(simple_str_with_format % 4, str('4 guten Resultate')) self.assertEqual(simple_context_with_format % 1, '1 gutes Resultat!') self.assertEqual(simple_context_with_format % 4, '4 guten Resultate!') self.assertEqual(simple_without_format % 1, 'gutes Resultat') self.assertEqual(simple_without_format % 4, 'guten Resultate') complex_nonlazy = ungettext_lazy('Hi %(name)s, %(num)d good result', 'Hi %(name)s, %(num)d good results', 4) complex_deferred = ungettext_lazy('Hi %(name)s, %(num)d good result', 'Hi %(name)s, %(num)d good results', 'num') complex_str_nonlazy = ngettext_lazy(str('Hi %(name)s, %(num)d good result'), str('Hi %(name)s, %(num)d good results'), 4) complex_str_deferred = ngettext_lazy(str('Hi %(name)s, %(num)d good result'), str('Hi %(name)s, %(num)d good results'), 'num') complex_context_nonlazy = npgettext_lazy('Greeting', 'Hi %(name)s, %(num)d good result', 'Hi %(name)s, %(num)d good results', 4) complex_context_deferred = npgettext_lazy('Greeting', 'Hi %(name)s, %(num)d good result', 'Hi %(name)s, %(num)d good results', 'num') with translation.override('de'): self.assertEqual(complex_nonlazy % {'num': 4, 'name': 'Jim'}, 'Hallo Jim, 4 guten Resultate') self.assertEqual(complex_deferred % {'name': 'Jim', 'num': 1}, 'Hallo Jim, 1 gutes Resultat') self.assertEqual(complex_deferred % {'name': 'Jim', 'num': 5}, 'Hallo Jim, 5 guten Resultate') with six.assertRaisesRegex(self, KeyError, 'Your dictionary lacks key.*'): complex_deferred % {'name': 'Jim'} self.assertEqual(complex_str_nonlazy % {'num': 4, 'name': 'Jim'}, str('Hallo Jim, 4 guten Resultate')) self.assertEqual(complex_str_deferred % {'name': 'Jim', 'num': 1}, str('Hallo Jim, 1 gutes Resultat')) self.assertEqual(complex_str_deferred % {'name': 'Jim', 'num': 5}, str('Hallo Jim, 5 guten Resultate')) with six.assertRaisesRegex(self, KeyError, 'Your dictionary lacks key.*'): complex_str_deferred % {'name': 'Jim'} self.assertEqual(complex_context_nonlazy % {'num': 4, 'name': 'Jim'}, 'Willkommen Jim, 4 guten Resultate') self.assertEqual(complex_context_deferred % {'name': 'Jim', 'num': 1}, 'Willkommen Jim, 1 gutes Resultat') self.assertEqual(complex_context_deferred % {'name': 'Jim', 'num': 5}, 'Willkommen Jim, 5 guten Resultate') with six.assertRaisesRegex(self, KeyError, 'Your dictionary lacks key.*'): complex_context_deferred % {'name': 'Jim'} @skipUnless(six.PY2, "PY3 doesn't have distinct int and long types") def test_ungettext_lazy_long(self): """ Regression test for #22820: int and long should be treated alike in ungettext_lazy. """ result = ungettext_lazy('%(name)s has %(num)d good result', '%(name)s has %(num)d good results', 4) self.assertEqual(result % {'name': 'Joe', 'num': 4}, "Joe has 4 good results") # Now with a long result = ungettext_lazy('%(name)s has %(num)d good result', '%(name)s has %(num)d good results', long(4)) self.assertEqual(result % {'name': 'Joe', 'num': 4}, "Joe has 4 good results") @override_settings(LOCALE_PATHS=extended_locale_paths) def test_pgettext(self): trans_real._active = local() trans_real._translations = {} with translation.override('de'): self.assertEqual(pgettext("unexisting", "May"), "May") self.assertEqual(pgettext("month name", "May"), "Mai") self.assertEqual(pgettext("verb", "May"), "Kann") self.assertEqual(npgettext("search", "%d result", "%d results", 4) % 4, "4 Resultate") @override_settings(LOCALE_PATHS=extended_locale_paths) def test_template_tags_pgettext(self): """ Ensure that message contexts are taken into account the {% trans %} and {% blocktrans %} template tags. Refs #14806. """ trans_real._active = local() trans_real._translations = {} with translation.override('de'): # {% trans %} ----------------------------------- # Inexisting context... t = Template('{% load i18n %}{% trans "May" context "unexisting" %}') rendered = t.render(Context()) self.assertEqual(rendered, 'May') # Existing context... # Using a literal t = Template('{% load i18n %}{% trans "May" context "month name" %}') rendered = t.render(Context()) self.assertEqual(rendered, 'Mai') t = Template('{% load i18n %}{% trans "May" context "verb" %}') rendered = t.render(Context()) self.assertEqual(rendered, 'Kann') # Using a variable t = Template('{% load i18n %}{% trans "May" context message_context %}') rendered = t.render(Context({'message_context': 'month name'})) self.assertEqual(rendered, 'Mai') t = Template('{% load i18n %}{% trans "May" context message_context %}') rendered = t.render(Context({'message_context': 'verb'})) self.assertEqual(rendered, 'Kann') # Using a filter t = Template('{% load i18n %}{% trans "May" context message_context|lower %}') rendered = t.render(Context({'message_context': 'MONTH NAME'})) self.assertEqual(rendered, 'Mai') t = Template('{% load i18n %}{% trans "May" context message_context|lower %}') rendered = t.render(Context({'message_context': 'VERB'})) self.assertEqual(rendered, 'Kann') # Using 'as' t = Template('{% load i18n %}{% trans "May" context "month name" as var %}Value: {{ var }}') rendered = t.render(Context()) self.assertEqual(rendered, 'Value: Mai') t = Template('{% load i18n %}{% trans "May" as var context "verb" %}Value: {{ var }}') rendered = t.render(Context()) self.assertEqual(rendered, 'Value: Kann') # {% blocktrans %} ------------------------------ # Inexisting context... t = Template('{% load i18n %}{% blocktrans context "unexisting" %}May{% endblocktrans %}') rendered = t.render(Context()) self.assertEqual(rendered, 'May') # Existing context... # Using a literal t = Template('{% load i18n %}{% blocktrans context "month name" %}May{% endblocktrans %}') rendered = t.render(Context()) self.assertEqual(rendered, 'Mai') t = Template('{% load i18n %}{% blocktrans context "verb" %}May{% endblocktrans %}') rendered = t.render(Context()) self.assertEqual(rendered, 'Kann') # Using a variable t = Template('{% load i18n %}{% blocktrans context message_context %}May{% endblocktrans %}') rendered = t.render(Context({'message_context': 'month name'})) self.assertEqual(rendered, 'Mai') t = Template('{% load i18n %}{% blocktrans context message_context %}May{% endblocktrans %}') rendered = t.render(Context({'message_context': 'verb'})) self.assertEqual(rendered, 'Kann') # Using a filter t = Template('{% load i18n %}{% blocktrans context message_context|lower %}May{% endblocktrans %}') rendered = t.render(Context({'message_context': 'MONTH NAME'})) self.assertEqual(rendered, 'Mai') t = Template('{% load i18n %}{% blocktrans context message_context|lower %}May{% endblocktrans %}') rendered = t.render(Context({'message_context': 'VERB'})) self.assertEqual(rendered, 'Kann') # Using 'count' t = Template('{% load i18n %}{% blocktrans count number=1 context "super search" %}{{ number }} super result{% plural %}{{ number }} super results{% endblocktrans %}') rendered = t.render(Context()) self.assertEqual(rendered, '1 Super-Ergebnis') t = Template('{% load i18n %}{% blocktrans count number=2 context "super search" %}{{ number }} super result{% plural %}{{ number }} super results{% endblocktrans %}') rendered = t.render(Context()) self.assertEqual(rendered, '2 Super-Ergebnisse') t = Template('{% load i18n %}{% blocktrans context "other super search" count number=1 %}{{ number }} super result{% plural %}{{ number }} super results{% endblocktrans %}') rendered = t.render(Context()) self.assertEqual(rendered, '1 anderen Super-Ergebnis') t = Template('{% load i18n %}{% blocktrans context "other super search" count number=2 %}{{ number }} super result{% plural %}{{ number }} super results{% endblocktrans %}') rendered = t.render(Context()) self.assertEqual(rendered, '2 andere Super-Ergebnisse') # Using 'with' t = Template('{% load i18n %}{% blocktrans with num_comments=5 context "comment count" %}There are {{ num_comments }} comments{% endblocktrans %}') rendered = t.render(Context()) self.assertEqual(rendered, 'Es gibt 5 Kommentare') t = Template('{% load i18n %}{% blocktrans with num_comments=5 context "other comment count" %}There are {{ num_comments }} comments{% endblocktrans %}') rendered = t.render(Context()) self.assertEqual(rendered, 'Andere: Es gibt 5 Kommentare') # Using trimmed t = Template('{% load i18n %}{% blocktrans trimmed %}\n\nThere\n\t are 5 \n\n comments\n{% endblocktrans %}') rendered = t.render(Context()) self.assertEqual(rendered, 'There are 5 comments') t = Template('{% load i18n %}{% blocktrans with num_comments=5 context "comment count" trimmed %}\n\nThere are \t\n \t {{ num_comments }} comments\n\n{% endblocktrans %}') rendered = t.render(Context()) self.assertEqual(rendered, 'Es gibt 5 Kommentare') t = Template('{% load i18n %}{% blocktrans context "other super search" count number=2 trimmed %}\n{{ number }} super \n result{% plural %}{{ number }} super results{% endblocktrans %}') rendered = t.render(Context()) self.assertEqual(rendered, '2 andere Super-Ergebnisse') # Mis-uses self.assertRaises(TemplateSyntaxError, Template, '{% load i18n %}{% blocktrans context with month="May" %}{{ month }}{% endblocktrans %}') self.assertRaises(TemplateSyntaxError, Template, '{% load i18n %}{% blocktrans context %}{% endblocktrans %}') self.assertRaises(TemplateSyntaxError, Template, '{% load i18n %}{% blocktrans count number=2 context %}{{ number }} super result{% plural %}{{ number }} super results{% endblocktrans %}') def test_string_concat(self): """ six.text_type(string_concat(...)) should not raise a TypeError - #4796 """ self.assertEqual('django', six.text_type(string_concat("dja", "ngo"))) def test_empty_value(self): """ Empty value must stay empty after being translated (#23196). """ with translation.override('de'): self.assertEqual("", ugettext("")) self.assertEqual(str(""), gettext(str(""))) s = mark_safe("") self.assertEqual(s, ugettext(s)) def test_safe_status(self): """ Translating a string requiring no auto-escaping shouldn't change the "safe" status. """ s = mark_safe(str('Password')) self.assertEqual(SafeString, type(s)) with translation.override('de', deactivate=True): self.assertEqual(SafeText, type(ugettext(s))) self.assertEqual('aPassword', SafeText('a') + s) self.assertEqual('Passworda', s + SafeText('a')) self.assertEqual('Passworda', s + mark_safe('a')) self.assertEqual('aPassword', mark_safe('a') + s) self.assertEqual('as', mark_safe('a') + mark_safe('s')) def test_maclines(self): """ Translations on files with mac or dos end of lines will be converted to unix eof in .po catalogs, and they have to match when retrieved """ ca_translation = trans_real.translation('ca') ca_translation._catalog['Mac\nEOF\n'] = 'Catalan Mac\nEOF\n' ca_translation._catalog['Win\nEOF\n'] = 'Catalan Win\nEOF\n' with translation.override('ca', deactivate=True): self.assertEqual('Catalan Mac\nEOF\n', ugettext('Mac\rEOF\r')) self.assertEqual('Catalan Win\nEOF\n', ugettext('Win\r\nEOF\r\n')) def test_to_locale(self): """ Tests the to_locale function and the special case of Serbian Latin (refs #12230 and r11299) """ self.assertEqual(to_locale('en-us'), 'en_US') self.assertEqual(to_locale('sr-lat'), 'sr_Lat') def test_to_language(self): """ Test the to_language function """ self.assertEqual(trans_real.to_language('en_US'), 'en-us') self.assertEqual(trans_real.to_language('sr_Lat'), 'sr-lat') def test_language_bidi(self): self.assertEqual(get_language_bidi(), False) with translation.override(None): self.assertEqual(get_language_bidi(), False) @override_settings(LOCALE_PATHS=[os.path.join(here, 'other', 'locale')]) def test_bad_placeholder_1(self): """ Error in translation file should not crash template rendering (%(person)s is translated as %(personne)s in fr.po) Refs #16516. """ with translation.override('fr'): t = Template('{% load i18n %}{% blocktrans %}My name is {{ person }}.{% endblocktrans %}') rendered = t.render(Context({'person': 'James'})) self.assertEqual(rendered, 'My name is James.') @override_settings(LOCALE_PATHS=[os.path.join(here, 'other', 'locale')]) def test_bad_placeholder_2(self): """ Error in translation file should not crash template rendering (%(person) misses a 's' in fr.po, causing the string formatting to fail) Refs #18393. """ with translation.override('fr'): t = Template('{% load i18n %}{% blocktrans %}My other name is {{ person }}.{% endblocktrans %}') rendered = t.render(Context({'person': 'James'})) self.assertEqual(rendered, 'My other name is James.') class TranslationThreadSafetyTests(SimpleTestCase): def setUp(self): self._old_language = get_language() self._translations = trans_real._translations # here we rely on .split() being called inside the _fetch() # in trans_real.translation() class sideeffect_str(str): def split(self, *args, **kwargs): res = str.split(self, *args, **kwargs) trans_real._translations['en-YY'] = None return res trans_real._translations = {sideeffect_str('en-XX'): None} def tearDown(self): trans_real._translations = self._translations activate(self._old_language) def test_bug14894_translation_activate_thread_safety(self): translation_count = len(trans_real._translations) try: translation.activate('pl') except RuntimeError: self.fail('translation.activate() is not thread-safe') # make sure sideeffect_str actually added a new translation self.assertLess(translation_count, len(trans_real._translations)) @override_settings(USE_L10N=True) class FormattingTests(SimpleTestCase): def setUp(self): super(FormattingTests, self).setUp() self.n = decimal.Decimal('66666.666') self.f = 99999.999 self.d = datetime.date(2009, 12, 31) self.dt = datetime.datetime(2009, 12, 31, 20, 50) self.t = datetime.time(10, 15, 48) self.l = 10000 if PY3 else long(10000) self.ctxt = Context({ 'n': self.n, 't': self.t, 'd': self.d, 'dt': self.dt, 'f': self.f, 'l': self.l, }) def test_locale_independent(self): """ Localization of numbers """ with self.settings(USE_THOUSAND_SEPARATOR=False): self.assertEqual('66666.66', nformat(self.n, decimal_sep='.', decimal_pos=2, grouping=3, thousand_sep=',')) self.assertEqual('66666A6', nformat(self.n, decimal_sep='A', decimal_pos=1, grouping=1, thousand_sep='B')) self.assertEqual('66666', nformat(self.n, decimal_sep='X', decimal_pos=0, grouping=1, thousand_sep='Y')) with self.settings(USE_THOUSAND_SEPARATOR=True): self.assertEqual('66,666.66', nformat(self.n, decimal_sep='.', decimal_pos=2, grouping=3, thousand_sep=',')) self.assertEqual('6B6B6B6B6A6', nformat(self.n, decimal_sep='A', decimal_pos=1, grouping=1, thousand_sep='B')) self.assertEqual('-66666.6', nformat(-66666.666, decimal_sep='.', decimal_pos=1)) self.assertEqual('-66666.0', nformat(int('-66666'), decimal_sep='.', decimal_pos=1)) self.assertEqual('10000.0', nformat(self.l, decimal_sep='.', decimal_pos=1)) # This unusual grouping/force_grouping combination may be triggered by the intcomma filter (#17414) self.assertEqual('10000', nformat(self.l, decimal_sep='.', decimal_pos=0, grouping=0, force_grouping=True)) # date filter self.assertEqual('31.12.2009 в 20:50', Template('{{ dt|date:"d.m.Y в H:i" }}').render(self.ctxt)) self.assertEqual('⌚ 10:15', Template('{{ t|time:"⌚ H:i" }}').render(self.ctxt)) @override_settings(USE_L10N=False) def test_l10n_disabled(self): """ Catalan locale with format i18n disabled translations will be used, but not formats """ with translation.override('ca', deactivate=True): self.maxDiff = 3000 self.assertEqual('N j, Y', get_format('DATE_FORMAT')) self.assertEqual(0, get_format('FIRST_DAY_OF_WEEK')) self.assertEqual('.', get_format('DECIMAL_SEPARATOR')) self.assertEqual('10:15 a.m.', time_format(self.t)) self.assertEqual('des. 31, 2009', date_format(self.d)) self.assertEqual('desembre 2009', date_format(self.d, 'YEAR_MONTH_FORMAT')) self.assertEqual('12/31/2009 8:50 p.m.', date_format(self.dt, 'SHORT_DATETIME_FORMAT')) self.assertEqual('No localizable', localize('No localizable')) self.assertEqual('66666.666', localize(self.n)) self.assertEqual('99999.999', localize(self.f)) self.assertEqual('10000', localize(self.l)) self.assertEqual('des. 31, 2009', localize(self.d)) self.assertEqual('des. 31, 2009, 8:50 p.m.', localize(self.dt)) self.assertEqual('66666.666', Template('{{ n }}').render(self.ctxt)) self.assertEqual('99999.999', Template('{{ f }}').render(self.ctxt)) self.assertEqual('des. 31, 2009', Template('{{ d }}').render(self.ctxt)) self.assertEqual('des. 31, 2009, 8:50 p.m.', Template('{{ dt }}').render(self.ctxt)) self.assertEqual('66666.67', Template('{{ n|floatformat:2 }}').render(self.ctxt)) self.assertEqual('100000.0', Template('{{ f|floatformat }}').render(self.ctxt)) self.assertEqual('10:15 a.m.', Template('{{ t|time:"TIME_FORMAT" }}').render(self.ctxt)) self.assertEqual('12/31/2009', Template('{{ d|date:"SHORT_DATE_FORMAT" }}').render(self.ctxt)) self.assertEqual('12/31/2009 8:50 p.m.', Template('{{ dt|date:"SHORT_DATETIME_FORMAT" }}').render(self.ctxt)) form = I18nForm({ 'decimal_field': '66666,666', 'float_field': '99999,999', 'date_field': '31/12/2009', 'datetime_field': '31/12/2009 20:50', 'time_field': '20:50', 'integer_field': '1.234', }) self.assertFalse(form.is_valid()) self.assertEqual(['Introdu\xefu un n\xfamero.'], form.errors['float_field']) self.assertEqual(['Introdu\xefu un n\xfamero.'], form.errors['decimal_field']) self.assertEqual(['Introdu\xefu una data v\xe0lida.'], form.errors['date_field']) self.assertEqual(['Introdu\xefu una data/hora v\xe0lides.'], form.errors['datetime_field']) self.assertEqual(['Introdu\xefu un n\xfamero sencer.'], form.errors['integer_field']) form2 = SelectDateForm({ 'date_field_month': '12', 'date_field_day': '31', 'date_field_year': '2009' }) self.assertTrue(form2.is_valid()) self.assertEqual(datetime.date(2009, 12, 31), form2.cleaned_data['date_field']) self.assertHTMLEqual( '<select name="mydate_month" id="id_mydate_month">\n<option value="0">---</option>\n<option value="1">gener</option>\n<option value="2">febrer</option>\n<option value="3">mar\xe7</option>\n<option value="4">abril</option>\n<option value="5">maig</option>\n<option value="6">juny</option>\n<option value="7">juliol</option>\n<option value="8">agost</option>\n<option value="9">setembre</option>\n<option value="10">octubre</option>\n<option value="11">novembre</option>\n<option value="12" selected="selected">desembre</option>\n</select>\n<select name="mydate_day" id="id_mydate_day">\n<option value="0">---</option>\n<option value="1">1</option>\n<option value="2">2</option>\n<option value="3">3</option>\n<option value="4">4</option>\n<option value="5">5</option>\n<option value="6">6</option>\n<option value="7">7</option>\n<option value="8">8</option>\n<option value="9">9</option>\n<option value="10">10</option>\n<option value="11">11</option>\n<option value="12">12</option>\n<option value="13">13</option>\n<option value="14">14</option>\n<option value="15">15</option>\n<option value="16">16</option>\n<option value="17">17</option>\n<option value="18">18</option>\n<option value="19">19</option>\n<option value="20">20</option>\n<option value="21">21</option>\n<option value="22">22</option>\n<option value="23">23</option>\n<option value="24">24</option>\n<option value="25">25</option>\n<option value="26">26</option>\n<option value="27">27</option>\n<option value="28">28</option>\n<option value="29">29</option>\n<option value="30">30</option>\n<option value="31" selected="selected">31</option>\n</select>\n<select name="mydate_year" id="id_mydate_year">\n<option value="0">---</option>\n<option value="2009" selected="selected">2009</option>\n<option value="2010">2010</option>\n<option value="2011">2011</option>\n<option value="2012">2012</option>\n<option value="2013">2013</option>\n<option value="2014">2014</option>\n<option value="2015">2015</option>\n<option value="2016">2016</option>\n<option value="2017">2017</option>\n<option value="2018">2018</option>\n</select>', forms.SelectDateWidget(years=range(2009, 2019)).render('mydate', datetime.date(2009, 12, 31)) ) # We shouldn't change the behavior of the floatformat filter re: # thousand separator and grouping when USE_L10N is False even # if the USE_THOUSAND_SEPARATOR, NUMBER_GROUPING and # THOUSAND_SEPARATOR settings are specified with self.settings(USE_THOUSAND_SEPARATOR=True, NUMBER_GROUPING=1, THOUSAND_SEPARATOR='!'): self.assertEqual('66666.67', Template('{{ n|floatformat:2 }}').render(self.ctxt)) self.assertEqual('100000.0', Template('{{ f|floatformat }}').render(self.ctxt)) def test_false_like_locale_formats(self): """ Ensure that the active locale's formats take precedence over the default settings even if they would be interpreted as False in a conditional test (e.g. 0 or empty string). Refs #16938. """ with patch_formats('fr', THOUSAND_SEPARATOR='', FIRST_DAY_OF_WEEK=0): with translation.override('fr'): with self.settings(USE_THOUSAND_SEPARATOR=True, THOUSAND_SEPARATOR='!'): self.assertEqual('', get_format('THOUSAND_SEPARATOR')) # Even a second time (after the format has been cached)... self.assertEqual('', get_format('THOUSAND_SEPARATOR')) with self.settings(FIRST_DAY_OF_WEEK=1): self.assertEqual(0, get_format('FIRST_DAY_OF_WEEK')) # Even a second time (after the format has been cached)... self.assertEqual(0, get_format('FIRST_DAY_OF_WEEK')) def test_l10n_enabled(self): self.maxDiff = 3000 # Catalan locale with translation.override('ca', deactivate=True): self.assertEqual('j \d\e F \d\e Y', get_format('DATE_FORMAT')) self.assertEqual(1, get_format('FIRST_DAY_OF_WEEK')) self.assertEqual(',', get_format('DECIMAL_SEPARATOR')) self.assertEqual('10:15', time_format(self.t)) self.assertEqual('31 de desembre de 2009', date_format(self.d)) self.assertEqual('desembre del 2009', date_format(self.d, 'YEAR_MONTH_FORMAT')) self.assertEqual('31/12/2009 20:50', date_format(self.dt, 'SHORT_DATETIME_FORMAT')) self.assertEqual('No localizable', localize('No localizable')) with self.settings(USE_THOUSAND_SEPARATOR=True): self.assertEqual('66.666,666', localize(self.n)) self.assertEqual('99.999,999', localize(self.f)) self.assertEqual('10.000', localize(self.l)) self.assertEqual('True', localize(True)) with self.settings(USE_THOUSAND_SEPARATOR=False): self.assertEqual('66666,666', localize(self.n)) self.assertEqual('99999,999', localize(self.f)) self.assertEqual('10000', localize(self.l)) self.assertEqual('31 de desembre de 2009', localize(self.d)) self.assertEqual('31 de desembre de 2009 a les 20:50', localize(self.dt)) with self.settings(USE_THOUSAND_SEPARATOR=True): self.assertEqual('66.666,666', Template('{{ n }}').render(self.ctxt)) self.assertEqual('99.999,999', Template('{{ f }}').render(self.ctxt)) self.assertEqual('10.000', Template('{{ l }}').render(self.ctxt)) with self.settings(USE_THOUSAND_SEPARATOR=True): form3 = I18nForm({ 'decimal_field': '66.666,666', 'float_field': '99.999,999', 'date_field': '31/12/2009', 'datetime_field': '31/12/2009 20:50', 'time_field': '20:50', 'integer_field': '1.234', }) self.assertTrue(form3.is_valid()) self.assertEqual(decimal.Decimal('66666.666'), form3.cleaned_data['decimal_field']) self.assertEqual(99999.999, form3.cleaned_data['float_field']) self.assertEqual(datetime.date(2009, 12, 31), form3.cleaned_data['date_field']) self.assertEqual(datetime.datetime(2009, 12, 31, 20, 50), form3.cleaned_data['datetime_field']) self.assertEqual(datetime.time(20, 50), form3.cleaned_data['time_field']) self.assertEqual(1234, form3.cleaned_data['integer_field']) with self.settings(USE_THOUSAND_SEPARATOR=False): self.assertEqual('66666,666', Template('{{ n }}').render(self.ctxt)) self.assertEqual('99999,999', Template('{{ f }}').render(self.ctxt)) self.assertEqual('31 de desembre de 2009', Template('{{ d }}').render(self.ctxt)) self.assertEqual('31 de desembre de 2009 a les 20:50', Template('{{ dt }}').render(self.ctxt)) self.assertEqual('66666,67', Template('{{ n|floatformat:2 }}').render(self.ctxt)) self.assertEqual('100000,0', Template('{{ f|floatformat }}').render(self.ctxt)) self.assertEqual('10:15', Template('{{ t|time:"TIME_FORMAT" }}').render(self.ctxt)) self.assertEqual('31/12/2009', Template('{{ d|date:"SHORT_DATE_FORMAT" }}').render(self.ctxt)) self.assertEqual('31/12/2009 20:50', Template('{{ dt|date:"SHORT_DATETIME_FORMAT" }}').render(self.ctxt)) self.assertEqual(date_format(datetime.datetime.now(), "DATE_FORMAT"), Template('{% now "DATE_FORMAT" %}').render(self.ctxt)) with self.settings(USE_THOUSAND_SEPARATOR=False): form4 = I18nForm({ 'decimal_field': '66666,666', 'float_field': '99999,999', 'date_field': '31/12/2009', 'datetime_field': '31/12/2009 20:50', 'time_field': '20:50', 'integer_field': '1234', }) self.assertTrue(form4.is_valid()) self.assertEqual(decimal.Decimal('66666.666'), form4.cleaned_data['decimal_field']) self.assertEqual(99999.999, form4.cleaned_data['float_field']) self.assertEqual(datetime.date(2009, 12, 31), form4.cleaned_data['date_field']) self.assertEqual(datetime.datetime(2009, 12, 31, 20, 50), form4.cleaned_data['datetime_field']) self.assertEqual(datetime.time(20, 50), form4.cleaned_data['time_field']) self.assertEqual(1234, form4.cleaned_data['integer_field']) form5 = SelectDateForm({ 'date_field_month': '12', 'date_field_day': '31', 'date_field_year': '2009' }) self.assertTrue(form5.is_valid()) self.assertEqual(datetime.date(2009, 12, 31), form5.cleaned_data['date_field']) self.assertHTMLEqual( '<select name="mydate_day" id="id_mydate_day">\n<option value="0">---</option>\n<option value="1">1</option>\n<option value="2">2</option>\n<option value="3">3</option>\n<option value="4">4</option>\n<option value="5">5</option>\n<option value="6">6</option>\n<option value="7">7</option>\n<option value="8">8</option>\n<option value="9">9</option>\n<option value="10">10</option>\n<option value="11">11</option>\n<option value="12">12</option>\n<option value="13">13</option>\n<option value="14">14</option>\n<option value="15">15</option>\n<option value="16">16</option>\n<option value="17">17</option>\n<option value="18">18</option>\n<option value="19">19</option>\n<option value="20">20</option>\n<option value="21">21</option>\n<option value="22">22</option>\n<option value="23">23</option>\n<option value="24">24</option>\n<option value="25">25</option>\n<option value="26">26</option>\n<option value="27">27</option>\n<option value="28">28</option>\n<option value="29">29</option>\n<option value="30">30</option>\n<option value="31" selected="selected">31</option>\n</select>\n<select name="mydate_month" id="id_mydate_month">\n<option value="0">---</option>\n<option value="1">gener</option>\n<option value="2">febrer</option>\n<option value="3">mar\xe7</option>\n<option value="4">abril</option>\n<option value="5">maig</option>\n<option value="6">juny</option>\n<option value="7">juliol</option>\n<option value="8">agost</option>\n<option value="9">setembre</option>\n<option value="10">octubre</option>\n<option value="11">novembre</option>\n<option value="12" selected="selected">desembre</option>\n</select>\n<select name="mydate_year" id="id_mydate_year">\n<option value="0">---</option>\n<option value="2009" selected="selected">2009</option>\n<option value="2010">2010</option>\n<option value="2011">2011</option>\n<option value="2012">2012</option>\n<option value="2013">2013</option>\n<option value="2014">2014</option>\n<option value="2015">2015</option>\n<option value="2016">2016</option>\n<option value="2017">2017</option>\n<option value="2018">2018</option>\n</select>', forms.SelectDateWidget(years=range(2009, 2019)).render('mydate', datetime.date(2009, 12, 31)) ) # Russian locale (with E as month) with translation.override('ru', deactivate=True): self.assertHTMLEqual( '<select name="mydate_day" id="id_mydate_day">\n<option value="0">---</option>\n<option value="1">1</option>\n<option value="2">2</option>\n<option value="3">3</option>\n<option value="4">4</option>\n<option value="5">5</option>\n<option value="6">6</option>\n<option value="7">7</option>\n<option value="8">8</option>\n<option value="9">9</option>\n<option value="10">10</option>\n<option value="11">11</option>\n<option value="12">12</option>\n<option value="13">13</option>\n<option value="14">14</option>\n<option value="15">15</option>\n<option value="16">16</option>\n<option value="17">17</option>\n<option value="18">18</option>\n<option value="19">19</option>\n<option value="20">20</option>\n<option value="21">21</option>\n<option value="22">22</option>\n<option value="23">23</option>\n<option value="24">24</option>\n<option value="25">25</option>\n<option value="26">26</option>\n<option value="27">27</option>\n<option value="28">28</option>\n<option value="29">29</option>\n<option value="30">30</option>\n<option value="31" selected="selected">31</option>\n</select>\n<select name="mydate_month" id="id_mydate_month">\n<option value="0">---</option>\n<option value="1">\u042f\u043d\u0432\u0430\u0440\u044c</option>\n<option value="2">\u0424\u0435\u0432\u0440\u0430\u043b\u044c</option>\n<option value="3">\u041c\u0430\u0440\u0442</option>\n<option value="4">\u0410\u043f\u0440\u0435\u043b\u044c</option>\n<option value="5">\u041c\u0430\u0439</option>\n<option value="6">\u0418\u044e\u043d\u044c</option>\n<option value="7">\u0418\u044e\u043b\u044c</option>\n<option value="8">\u0410\u0432\u0433\u0443\u0441\u0442</option>\n<option value="9">\u0421\u0435\u043d\u0442\u044f\u0431\u0440\u044c</option>\n<option value="10">\u041e\u043a\u0442\u044f\u0431\u0440\u044c</option>\n<option value="11">\u041d\u043e\u044f\u0431\u0440\u044c</option>\n<option value="12" selected="selected">\u0414\u0435\u043a\u0430\u0431\u0440\u044c</option>\n</select>\n<select name="mydate_year" id="id_mydate_year">\n<option value="0">---</option>\n<option value="2009" selected="selected">2009</option>\n<option value="2010">2010</option>\n<option value="2011">2011</option>\n<option value="2012">2012</option>\n<option value="2013">2013</option>\n<option value="2014">2014</option>\n<option value="2015">2015</option>\n<option value="2016">2016</option>\n<option value="2017">2017</option>\n<option value="2018">2018</option>\n</select>', forms.SelectDateWidget(years=range(2009, 2019)).render('mydate', datetime.date(2009, 12, 31)) ) # English locale with translation.override('en', deactivate=True): self.assertEqual('N j, Y', get_format('DATE_FORMAT')) self.assertEqual(0, get_format('FIRST_DAY_OF_WEEK')) self.assertEqual('.', get_format('DECIMAL_SEPARATOR')) self.assertEqual('Dec. 31, 2009', date_format(self.d)) self.assertEqual('December 2009', date_format(self.d, 'YEAR_MONTH_FORMAT')) self.assertEqual('12/31/2009 8:50 p.m.', date_format(self.dt, 'SHORT_DATETIME_FORMAT')) self.assertEqual('No localizable', localize('No localizable')) with self.settings(USE_THOUSAND_SEPARATOR=True): self.assertEqual('66,666.666', localize(self.n)) self.assertEqual('99,999.999', localize(self.f)) self.assertEqual('10,000', localize(self.l)) with self.settings(USE_THOUSAND_SEPARATOR=False): self.assertEqual('66666.666', localize(self.n)) self.assertEqual('99999.999', localize(self.f)) self.assertEqual('10000', localize(self.l)) self.assertEqual('Dec. 31, 2009', localize(self.d)) self.assertEqual('Dec. 31, 2009, 8:50 p.m.', localize(self.dt)) with self.settings(USE_THOUSAND_SEPARATOR=True): self.assertEqual('66,666.666', Template('{{ n }}').render(self.ctxt)) self.assertEqual('99,999.999', Template('{{ f }}').render(self.ctxt)) self.assertEqual('10,000', Template('{{ l }}').render(self.ctxt)) with self.settings(USE_THOUSAND_SEPARATOR=False): self.assertEqual('66666.666', Template('{{ n }}').render(self.ctxt)) self.assertEqual('99999.999', Template('{{ f }}').render(self.ctxt)) self.assertEqual('Dec. 31, 2009', Template('{{ d }}').render(self.ctxt)) self.assertEqual('Dec. 31, 2009, 8:50 p.m.', Template('{{ dt }}').render(self.ctxt)) self.assertEqual('66666.67', Template('{{ n|floatformat:2 }}').render(self.ctxt)) self.assertEqual('100000.0', Template('{{ f|floatformat }}').render(self.ctxt)) self.assertEqual('12/31/2009', Template('{{ d|date:"SHORT_DATE_FORMAT" }}').render(self.ctxt)) self.assertEqual('12/31/2009 8:50 p.m.', Template('{{ dt|date:"SHORT_DATETIME_FORMAT" }}').render(self.ctxt)) form5 = I18nForm({ 'decimal_field': '66666.666', 'float_field': '99999.999', 'date_field': '12/31/2009', 'datetime_field': '12/31/2009 20:50', 'time_field': '20:50', 'integer_field': '1234', }) self.assertTrue(form5.is_valid()) self.assertEqual(decimal.Decimal('66666.666'), form5.cleaned_data['decimal_field']) self.assertEqual(99999.999, form5.cleaned_data['float_field']) self.assertEqual(datetime.date(2009, 12, 31), form5.cleaned_data['date_field']) self.assertEqual(datetime.datetime(2009, 12, 31, 20, 50), form5.cleaned_data['datetime_field']) self.assertEqual(datetime.time(20, 50), form5.cleaned_data['time_field']) self.assertEqual(1234, form5.cleaned_data['integer_field']) form6 = SelectDateForm({ 'date_field_month': '12', 'date_field_day': '31', 'date_field_year': '2009' }) self.assertTrue(form6.is_valid()) self.assertEqual(datetime.date(2009, 12, 31), form6.cleaned_data['date_field']) self.assertHTMLEqual( '<select name="mydate_month" id="id_mydate_month">\n<option value="0">---</option>\n<option value="1">January</option>\n<option value="2">February</option>\n<option value="3">March</option>\n<option value="4">April</option>\n<option value="5">May</option>\n<option value="6">June</option>\n<option value="7">July</option>\n<option value="8">August</option>\n<option value="9">September</option>\n<option value="10">October</option>\n<option value="11">November</option>\n<option value="12" selected="selected">December</option>\n</select>\n<select name="mydate_day" id="id_mydate_day">\n<option value="0">---</option>\n<option value="1">1</option>\n<option value="2">2</option>\n<option value="3">3</option>\n<option value="4">4</option>\n<option value="5">5</option>\n<option value="6">6</option>\n<option value="7">7</option>\n<option value="8">8</option>\n<option value="9">9</option>\n<option value="10">10</option>\n<option value="11">11</option>\n<option value="12">12</option>\n<option value="13">13</option>\n<option value="14">14</option>\n<option value="15">15</option>\n<option value="16">16</option>\n<option value="17">17</option>\n<option value="18">18</option>\n<option value="19">19</option>\n<option value="20">20</option>\n<option value="21">21</option>\n<option value="22">22</option>\n<option value="23">23</option>\n<option value="24">24</option>\n<option value="25">25</option>\n<option value="26">26</option>\n<option value="27">27</option>\n<option value="28">28</option>\n<option value="29">29</option>\n<option value="30">30</option>\n<option value="31" selected="selected">31</option>\n</select>\n<select name="mydate_year" id="id_mydate_year">\n<option value="0">---</option>\n<option value="2009" selected="selected">2009</option>\n<option value="2010">2010</option>\n<option value="2011">2011</option>\n<option value="2012">2012</option>\n<option value="2013">2013</option>\n<option value="2014">2014</option>\n<option value="2015">2015</option>\n<option value="2016">2016</option>\n<option value="2017">2017</option>\n<option value="2018">2018</option>\n</select>', forms.SelectDateWidget(years=range(2009, 2019)).render('mydate', datetime.date(2009, 12, 31)) ) def test_sub_locales(self): """ Check if sublocales fall back to the main locale """ with self.settings(USE_THOUSAND_SEPARATOR=True): with translation.override('de-at', deactivate=True): self.assertEqual('66.666,666', Template('{{ n }}').render(self.ctxt)) with translation.override('es-us', deactivate=True): self.assertEqual('31 de Diciembre de 2009', date_format(self.d)) def test_localized_input(self): """ Tests if form input is correctly localized """ self.maxDiff = 1200 with translation.override('de-at', deactivate=True): form6 = CompanyForm({ 'name': 'acme', 'date_added': datetime.datetime(2009, 12, 31, 6, 0, 0), 'cents_paid': decimal.Decimal('59.47'), 'products_delivered': 12000, }) self.assertTrue(form6.is_valid()) self.assertHTMLEqual( form6.as_ul(), '<li><label for="id_name">Name:</label> <input id="id_name" type="text" name="name" value="acme" maxlength="50" /></li>\n<li><label for="id_date_added">Date added:</label> <input type="text" name="date_added" value="31.12.2009 06:00:00" id="id_date_added" /></li>\n<li><label for="id_cents_paid">Cents paid:</label> <input type="text" name="cents_paid" value="59,47" id="id_cents_paid" /></li>\n<li><label for="id_products_delivered">Products delivered:</label> <input type="text" name="products_delivered" value="12000" id="id_products_delivered" /></li>' ) self.assertEqual(localize_input(datetime.datetime(2009, 12, 31, 6, 0, 0)), '31.12.2009 06:00:00') self.assertEqual(datetime.datetime(2009, 12, 31, 6, 0, 0), form6.cleaned_data['date_added']) with self.settings(USE_THOUSAND_SEPARATOR=True): # Checking for the localized "products_delivered" field self.assertInHTML('<input type="text" name="products_delivered" value="12.000" id="id_products_delivered" />', form6.as_ul()) def test_sanitize_separators(self): """ Tests django.utils.formats.sanitize_separators. """ # Non-strings are untouched self.assertEqual(sanitize_separators(123), 123) with translation.override('ru', deactivate=True): # Russian locale has non-breaking space (\xa0) as thousand separator # Check that usual space is accepted too when sanitizing inputs with self.settings(USE_THOUSAND_SEPARATOR=True): self.assertEqual(sanitize_separators('1\xa0234\xa0567'), '1234567') self.assertEqual(sanitize_separators('77\xa0777,777'), '77777.777') self.assertEqual(sanitize_separators('12 345'), '12345') self.assertEqual(sanitize_separators('77 777,777'), '77777.777') with self.settings(USE_THOUSAND_SEPARATOR=True, USE_L10N=False): self.assertEqual(sanitize_separators('12\xa0345'), '12\xa0345') with patch_formats(get_language(), THOUSAND_SEPARATOR='.', DECIMAL_SEPARATOR=','): with self.settings(USE_THOUSAND_SEPARATOR=True): self.assertEqual(sanitize_separators('10.234'), '10234') # Suspicion that user entered dot as decimal separator (#22171) self.assertEqual(sanitize_separators('10.10'), '10.10') def test_iter_format_modules(self): """ Tests the iter_format_modules function. """ # Importing some format modules so that we can compare the returned # modules with these expected modules default_mod = import_module('django.conf.locale.de.formats') test_mod = import_module('i18n.other.locale.de.formats') test_mod2 = import_module('i18n.other2.locale.de.formats') with translation.override('de-at', deactivate=True): # Should return the correct default module when no setting is set self.assertEqual(list(iter_format_modules('de')), [default_mod]) # When the setting is a string, should return the given module and # the default module self.assertEqual( list(iter_format_modules('de', 'i18n.other.locale')), [test_mod, default_mod]) # When setting is a list of strings, should return the given # modules and the default module self.assertEqual( list(iter_format_modules('de', ['i18n.other.locale', 'i18n.other2.locale'])), [test_mod, test_mod2, default_mod]) def test_iter_format_modules_stability(self): """ Tests the iter_format_modules function always yields format modules in a stable and correct order in presence of both base ll and ll_CC formats. """ en_format_mod = import_module('django.conf.locale.en.formats') en_gb_format_mod = import_module('django.conf.locale.en_GB.formats') self.assertEqual(list(iter_format_modules('en-gb')), [en_gb_format_mod, en_format_mod]) def test_get_format_modules_lang(self): with translation.override('de', deactivate=True): self.assertEqual('.', get_format('DECIMAL_SEPARATOR', lang='en')) def test_get_format_modules_stability(self): with self.settings(FORMAT_MODULE_PATH='i18n.other.locale'): with translation.override('de', deactivate=True): old = str("%r") % get_format_modules(reverse=True) new = str("%r") % get_format_modules(reverse=True) # second try self.assertEqual(new, old, 'Value returned by get_formats_modules() must be preserved between calls.') def test_localize_templatetag_and_filter(self): """ Tests the {% localize %} templatetag """ context = Context({'value': 3.14}) template1 = Template("{% load l10n %}{% localize %}{{ value }}{% endlocalize %};{% localize on %}{{ value }}{% endlocalize %}") template2 = Template("{% load l10n %}{{ value }};{% localize off %}{{ value }};{% endlocalize %}{{ value }}") template3 = Template('{% load l10n %}{{ value }};{{ value|unlocalize }}') template4 = Template('{% load l10n %}{{ value }};{{ value|localize }}') output1 = '3,14;3,14' output2 = '3,14;3.14;3,14' output3 = '3,14;3.14' output4 = '3.14;3,14' with translation.override('de', deactivate=True): with self.settings(USE_L10N=False): self.assertEqual(template1.render(context), output1) self.assertEqual(template4.render(context), output4) with self.settings(USE_L10N=True): self.assertEqual(template1.render(context), output1) self.assertEqual(template2.render(context), output2) self.assertEqual(template3.render(context), output3) def test_localized_as_text_as_hidden_input(self): """ Tests if form input with 'as_hidden' or 'as_text' is correctly localized. Ticket #18777 """ self.maxDiff = 1200 with translation.override('de-at', deactivate=True): template = Template('{% load l10n %}{{ form.date_added }}; {{ form.cents_paid }}') template_as_text = Template('{% load l10n %}{{ form.date_added.as_text }}; {{ form.cents_paid.as_text }}') template_as_hidden = Template('{% load l10n %}{{ form.date_added.as_hidden }}; {{ form.cents_paid.as_hidden }}') form = CompanyForm({ 'name': 'acme', 'date_added': datetime.datetime(2009, 12, 31, 6, 0, 0), 'cents_paid': decimal.Decimal('59.47'), 'products_delivered': 12000, }) context = Context({'form': form}) self.assertTrue(form.is_valid()) self.assertHTMLEqual( template.render(context), '<input id="id_date_added" name="date_added" type="text" value="31.12.2009 06:00:00" />; <input id="id_cents_paid" name="cents_paid" type="text" value="59,47" />' ) self.assertHTMLEqual( template_as_text.render(context), '<input id="id_date_added" name="date_added" type="text" value="31.12.2009 06:00:00" />; <input id="id_cents_paid" name="cents_paid" type="text" value="59,47" />' ) self.assertHTMLEqual( template_as_hidden.render(context), '<input id="id_date_added" name="date_added" type="hidden" value="31.12.2009 06:00:00" />; <input id="id_cents_paid" name="cents_paid" type="hidden" value="59,47" />' ) class MiscTests(SimpleTestCase): def setUp(self): super(MiscTests, self).setUp() self.rf = RequestFactory() @override_settings(LANGUAGE_CODE='de') def test_english_fallback(self): """ With a non-English LANGUAGE_CODE and if the active language is English or one of its variants, the untranslated string should be returned (instead of falling back to LANGUAGE_CODE) (See #24413). """ self.assertEqual(ugettext("Image"), "Bild") with translation.override('en'): self.assertEqual(ugettext("Image"), "Image") with translation.override('en-us'): self.assertEqual(ugettext("Image"), "Image") with translation.override('en-ca'): self.assertEqual(ugettext("Image"), "Image") def test_parse_spec_http_header(self): """ Testing HTTP header parsing. First, we test that we can parse the values according to the spec (and that we extract all the pieces in the right order). """ p = trans_real.parse_accept_lang_header # Good headers. self.assertEqual([('de', 1.0)], p('de')) self.assertEqual([('en-au', 1.0)], p('en-AU')) self.assertEqual([('es-419', 1.0)], p('es-419')) self.assertEqual([('*', 1.0)], p('*;q=1.00')) self.assertEqual([('en-au', 0.123)], p('en-AU;q=0.123')) self.assertEqual([('en-au', 0.5)], p('en-au;q=0.5')) self.assertEqual([('en-au', 1.0)], p('en-au;q=1.0')) self.assertEqual([('da', 1.0), ('en', 0.5), ('en-gb', 0.25)], p('da, en-gb;q=0.25, en;q=0.5')) self.assertEqual([('en-au-xx', 1.0)], p('en-au-xx')) self.assertEqual([('de', 1.0), ('en-au', 0.75), ('en-us', 0.5), ('en', 0.25), ('es', 0.125), ('fa', 0.125)], p('de,en-au;q=0.75,en-us;q=0.5,en;q=0.25,es;q=0.125,fa;q=0.125')) self.assertEqual([('*', 1.0)], p('*')) self.assertEqual([('de', 1.0)], p('de;q=0.')) self.assertEqual([('en', 1.0), ('*', 0.5)], p('en; q=1.0, * ; q=0.5')) self.assertEqual([], p('')) # Bad headers; should always return []. self.assertEqual([], p('en-gb;q=1.0000')) self.assertEqual([], p('en;q=0.1234')) self.assertEqual([], p('en;q=.2')) self.assertEqual([], p('abcdefghi-au')) self.assertEqual([], p('**')) self.assertEqual([], p('en,,gb')) self.assertEqual([], p('en-au;q=0.1.0')) self.assertEqual([], p('XXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXXZ,en')) self.assertEqual([], p('da, en-gb;q=0.8, en;q=0.7,#')) self.assertEqual([], p('de;q=2.0')) self.assertEqual([], p('de;q=0.a')) self.assertEqual([], p('12-345')) self.assertEqual([], p('')) self.assertEqual([], p('en; q=1,')) def test_parse_literal_http_header(self): """ Now test that we parse a literal HTTP header correctly. """ g = get_language_from_request r = self.rf.get('/') r.COOKIES = {} r.META = {'HTTP_ACCEPT_LANGUAGE': 'pt-br'} self.assertEqual('pt-br', g(r)) r.META = {'HTTP_ACCEPT_LANGUAGE': 'pt'} self.assertEqual('pt', g(r)) r.META = {'HTTP_ACCEPT_LANGUAGE': 'es,de'} self.assertEqual('es', g(r)) r.META = {'HTTP_ACCEPT_LANGUAGE': 'es-ar,de'} self.assertEqual('es-ar', g(r)) # This test assumes there won't be a Django translation to a US # variation of the Spanish language, a safe assumption. When the # user sets it as the preferred language, the main 'es' # translation should be selected instead. r.META = {'HTTP_ACCEPT_LANGUAGE': 'es-us'} self.assertEqual(g(r), 'es') # This tests the following scenario: there isn't a main language (zh) # translation of Django but there is a translation to variation (zh-hans) # the user sets zh-hans as the preferred language, it should be selected # by Django without falling back nor ignoring it. r.META = {'HTTP_ACCEPT_LANGUAGE': 'zh-hans,de'} self.assertEqual(g(r), 'zh-hans') r.META = {'HTTP_ACCEPT_LANGUAGE': 'NL'} self.assertEqual('nl', g(r)) r.META = {'HTTP_ACCEPT_LANGUAGE': 'fy'} self.assertEqual('fy', g(r)) r.META = {'HTTP_ACCEPT_LANGUAGE': 'ia'} self.assertEqual('ia', g(r)) r.META = {'HTTP_ACCEPT_LANGUAGE': 'sr-latn'} self.assertEqual('sr-latn', g(r)) r.META = {'HTTP_ACCEPT_LANGUAGE': 'zh-hans'} self.assertEqual('zh-hans', g(r)) r.META = {'HTTP_ACCEPT_LANGUAGE': 'zh-hant'} self.assertEqual('zh-hant', g(r)) @override_settings( LANGUAGES=[ ('en', 'English'), ('zh-hans', 'Simplified Chinese'), ('zh-hant', 'Traditional Chinese'), ] ) def test_support_for_deprecated_chinese_language_codes(self): """ Some browsers (Firefox, IE etc) use deprecated language codes. As these language codes will be removed in Django 1.9, these will be incorrectly matched. For example zh-tw (traditional) will be interpreted as zh-hans (simplified), which is wrong. So we should also accept these deprecated language codes. refs #18419 -- this is explicitly for browser compatibility """ g = get_language_from_request r = self.rf.get('/') r.COOKIES = {} r.META = {'HTTP_ACCEPT_LANGUAGE': 'zh-cn,en'} self.assertEqual(g(r), 'zh-hans') r.META = {'HTTP_ACCEPT_LANGUAGE': 'zh-tw,en'} self.assertEqual(g(r), 'zh-hant') def test_special_fallback_language(self): """ Some languages may have special fallbacks that don't follow the simple 'fr-ca' -> 'fr' logic (notably Chinese codes). """ r = self.rf.get('/') r.COOKIES = {} r.META = {'HTTP_ACCEPT_LANGUAGE': 'zh-my,en'} self.assertEqual(get_language_from_request(r), 'zh-hans') def test_parse_language_cookie(self): """ Now test that we parse language preferences stored in a cookie correctly. """ g = get_language_from_request r = self.rf.get('/') r.COOKIES = {settings.LANGUAGE_COOKIE_NAME: 'pt-br'} r.META = {} self.assertEqual('pt-br', g(r)) r.COOKIES = {settings.LANGUAGE_COOKIE_NAME: 'pt'} r.META = {} self.assertEqual('pt', g(r)) r.COOKIES = {settings.LANGUAGE_COOKIE_NAME: 'es'} r.META = {'HTTP_ACCEPT_LANGUAGE': 'de'} self.assertEqual('es', g(r)) # This test assumes there won't be a Django translation to a US # variation of the Spanish language, a safe assumption. When the # user sets it as the preferred language, the main 'es' # translation should be selected instead. r.COOKIES = {settings.LANGUAGE_COOKIE_NAME: 'es-us'} r.META = {} self.assertEqual(g(r), 'es') # This tests the following scenario: there isn't a main language (zh) # translation of Django but there is a translation to variation (zh-hans) # the user sets zh-hans as the preferred language, it should be selected # by Django without falling back nor ignoring it. r.COOKIES = {settings.LANGUAGE_COOKIE_NAME: 'zh-hans'} r.META = {'HTTP_ACCEPT_LANGUAGE': 'de'} self.assertEqual(g(r), 'zh-hans') def test_get_language_from_path_real(self): g = trans_real.get_language_from_path self.assertEqual(g('/pl/'), 'pl') self.assertEqual(g('/pl'), 'pl') self.assertEqual(g('/xyz/'), None) def test_get_language_from_path_null(self): from django.utils.translation.trans_null import get_language_from_path as g self.assertEqual(g('/pl/'), None) self.assertEqual(g('/pl'), None) self.assertEqual(g('/xyz/'), None) @override_settings(LOCALE_PATHS=extended_locale_paths) def test_percent_in_translatable_block(self): t_sing = Template("{% load i18n %}{% blocktrans %}The result was {{ percent }}%{% endblocktrans %}") t_plur = Template("{% load i18n %}{% blocktrans count num as number %}{{ percent }}% represents {{ num }} object{% plural %}{{ percent }}% represents {{ num }} objects{% endblocktrans %}") with translation.override('de'): self.assertEqual(t_sing.render(Context({'percent': 42})), 'Das Ergebnis war 42%') self.assertEqual(t_plur.render(Context({'percent': 42, 'num': 1})), '42% stellt 1 Objekt dar') self.assertEqual(t_plur.render(Context({'percent': 42, 'num': 4})), '42% stellt 4 Objekte dar') @override_settings(LOCALE_PATHS=extended_locale_paths) def test_percent_formatting_in_blocktrans(self): """ Test that using Python's %-formatting is properly escaped in blocktrans, singular or plural """ t_sing = Template("{% load i18n %}{% blocktrans %}There are %(num_comments)s comments{% endblocktrans %}") t_plur = Template("{% load i18n %}{% blocktrans count num as number %}%(percent)s% represents {{ num }} object{% plural %}%(percent)s% represents {{ num }} objects{% endblocktrans %}") with translation.override('de'): # Strings won't get translated as they don't match after escaping % self.assertEqual(t_sing.render(Context({'num_comments': 42})), 'There are %(num_comments)s comments') self.assertEqual(t_plur.render(Context({'percent': 42, 'num': 1})), '%(percent)s% represents 1 object') self.assertEqual(t_plur.render(Context({'percent': 42, 'num': 4})), '%(percent)s% represents 4 objects') def test_cache_resetting(self): """ #14170 after setting LANGUAGE, cache should be cleared and languages previously valid should not be used. """ g = get_language_from_request r = self.rf.get('/') r.COOKIES = {} r.META = {'HTTP_ACCEPT_LANGUAGE': 'pt-br'} self.assertEqual('pt-br', g(r)) with self.settings(LANGUAGES=[('en', 'English')]): self.assertNotEqual('pt-br', g(r)) class ResolutionOrderI18NTests(SimpleTestCase): def setUp(self): super(ResolutionOrderI18NTests, self).setUp() activate('de') def tearDown(self): deactivate() super(ResolutionOrderI18NTests, self).tearDown() def assertUgettext(self, msgid, msgstr): result = ugettext(msgid) self.assertIn(msgstr, result, ("The string '%s' isn't in the " "translation of '%s'; the actual result is '%s'." % (msgstr, msgid, result))) class AppResolutionOrderI18NTests(ResolutionOrderI18NTests): @override_settings(LANGUAGE_CODE='de') def test_app_translation(self): # Original translation. self.assertUgettext('Date/time', 'Datum/Zeit') # Different translation. with self.modify_settings(INSTALLED_APPS={'append': 'i18n.resolution'}): # Force refreshing translations. activate('de') # Doesn't work because it's added later in the list. self.assertUgettext('Date/time', 'Datum/Zeit') with self.modify_settings(INSTALLED_APPS={'remove': 'django.contrib.admin.apps.SimpleAdminConfig'}): # Force refreshing translations. activate('de') # Unless the original is removed from the list. self.assertUgettext('Date/time', 'Datum/Zeit (APP)') @override_settings(LOCALE_PATHS=extended_locale_paths) class LocalePathsResolutionOrderI18NTests(ResolutionOrderI18NTests): def test_locale_paths_translation(self): self.assertUgettext('Time', 'LOCALE_PATHS') def test_locale_paths_override_app_translation(self): with self.settings(INSTALLED_APPS=['i18n.resolution']): self.assertUgettext('Time', 'LOCALE_PATHS') class DjangoFallbackResolutionOrderI18NTests(ResolutionOrderI18NTests): def test_django_fallback(self): self.assertEqual(ugettext('Date/time'), 'Datum/Zeit') class TestModels(TestCase): def test_lazy(self): tm = TestModel() tm.save() def test_safestr(self): c = Company(cents_paid=12, products_delivered=1) c.name = SafeText('Iñtërnâtiônàlizætiøn1') c.save() c.name = SafeBytes('Iñtërnâtiônàlizætiøn1'.encode('utf-8')) c.save() class TestLanguageInfo(SimpleTestCase): def test_localized_language_info(self): li = get_language_info('de') self.assertEqual(li['code'], 'de') self.assertEqual(li['name_local'], 'Deutsch') self.assertEqual(li['name'], 'German') self.assertEqual(li['bidi'], False) def test_unknown_language_code(self): six.assertRaisesRegex(self, KeyError, r"Unknown language code xx\.", get_language_info, 'xx') def test_unknown_only_country_code(self): li = get_language_info('de-xx') self.assertEqual(li['code'], 'de') self.assertEqual(li['name_local'], 'Deutsch') self.assertEqual(li['name'], 'German') self.assertEqual(li['bidi'], False) def test_unknown_language_code_and_country_code(self): six.assertRaisesRegex(self, KeyError, r"Unknown language code xx-xx and xx\.", get_language_info, 'xx-xx') def test_fallback_language_code(self): """ get_language_info return the first fallback language info if the lang_info struct does not contain the 'name' key. """ li = get_language_info('zh-my') self.assertEqual(li['code'], 'zh-hans') li = get_language_info('zh-hans') self.assertEqual(li['code'], 'zh-hans') class MultipleLocaleActivationTests(SimpleTestCase): """ Tests for template rendering behavior when multiple locales are activated during the lifetime of the same process. """ def setUp(self): super(MultipleLocaleActivationTests, self).setUp() self._old_language = get_language() def tearDown(self): super(MultipleLocaleActivationTests, self).tearDown() activate(self._old_language) def test_single_locale_activation(self): """ Simple baseline behavior with one locale for all the supported i18n constructs. """ with translation.override('fr'): self.assertEqual(Template("{{ _('Yes') }}").render(Context({})), 'Oui') self.assertEqual(Template("{% load i18n %}{% trans 'Yes' %}").render(Context({})), 'Oui') self.assertEqual(Template("{% load i18n %}{% blocktrans %}Yes{% endblocktrans %}").render(Context({})), 'Oui') # Literal marked up with _() in a filter expression def test_multiple_locale_filter(self): with translation.override('de'): t = Template("{% load i18n %}{{ 0|yesno:_('yes,no,maybe') }}") with translation.override(self._old_language), translation.override('nl'): self.assertEqual(t.render(Context({})), 'nee') def test_multiple_locale_filter_deactivate(self): with translation.override('de', deactivate=True): t = Template("{% load i18n %}{{ 0|yesno:_('yes,no,maybe') }}") with translation.override('nl'): self.assertEqual(t.render(Context({})), 'nee') def test_multiple_locale_filter_direct_switch(self): with translation.override('de'): t = Template("{% load i18n %}{{ 0|yesno:_('yes,no,maybe') }}") with translation.override('nl'): self.assertEqual(t.render(Context({})), 'nee') # Literal marked up with _() def test_multiple_locale(self): with translation.override('de'): t = Template("{{ _('No') }}") with translation.override(self._old_language), translation.override('nl'): self.assertEqual(t.render(Context({})), 'Nee') def test_multiple_locale_deactivate(self): with translation.override('de', deactivate=True): t = Template("{{ _('No') }}") with translation.override('nl'): self.assertEqual(t.render(Context({})), 'Nee') def test_multiple_locale_direct_switch(self): with translation.override('de'): t = Template("{{ _('No') }}") with translation.override('nl'): self.assertEqual(t.render(Context({})), 'Nee') # Literal marked up with _(), loading the i18n template tag library def test_multiple_locale_loadi18n(self): with translation.override('de'): t = Template("{% load i18n %}{{ _('No') }}") with translation.override(self._old_language), translation.override('nl'): self.assertEqual(t.render(Context({})), 'Nee') def test_multiple_locale_loadi18n_deactivate(self): with translation.override('de', deactivate=True): t = Template("{% load i18n %}{{ _('No') }}") with translation.override('nl'): self.assertEqual(t.render(Context({})), 'Nee') def test_multiple_locale_loadi18n_direct_switch(self): with translation.override('de'): t = Template("{% load i18n %}{{ _('No') }}") with translation.override('nl'): self.assertEqual(t.render(Context({})), 'Nee') # trans i18n tag def test_multiple_locale_trans(self): with translation.override('de'): t = Template("{% load i18n %}{% trans 'No' %}") with translation.override(self._old_language), translation.override('nl'): self.assertEqual(t.render(Context({})), 'Nee') def test_multiple_locale_deactivate_trans(self): with translation.override('de', deactivate=True): t = Template("{% load i18n %}{% trans 'No' %}") with translation.override('nl'): self.assertEqual(t.render(Context({})), 'Nee') def test_multiple_locale_direct_switch_trans(self): with translation.override('de'): t = Template("{% load i18n %}{% trans 'No' %}") with translation.override('nl'): self.assertEqual(t.render(Context({})), 'Nee') # blocktrans i18n tag def test_multiple_locale_btrans(self): with translation.override('de'): t = Template("{% load i18n %}{% blocktrans %}No{% endblocktrans %}") with translation.override(self._old_language), translation.override('nl'): self.assertEqual(t.render(Context({})), 'Nee') def test_multiple_locale_deactivate_btrans(self): with translation.override('de', deactivate=True): t = Template("{% load i18n %}{% blocktrans %}No{% endblocktrans %}") with translation.override('nl'): self.assertEqual(t.render(Context({})), 'Nee') def test_multiple_locale_direct_switch_btrans(self): with translation.override('de'): t = Template("{% load i18n %}{% blocktrans %}No{% endblocktrans %}") with translation.override('nl'): self.assertEqual(t.render(Context({})), 'Nee') @override_settings( USE_I18N=True, LANGUAGES=[ ('en', 'English'), ('fr', 'French'), ], MIDDLEWARE_CLASSES=[ 'django.middleware.locale.LocaleMiddleware', 'django.middleware.common.CommonMiddleware', ], ROOT_URLCONF='i18n.urls', ) class LocaleMiddlewareTests(TestCase): def test_streaming_response(self): # Regression test for #5241 response = self.client.get('/fr/streaming/') self.assertContains(response, "Oui/Non") response = self.client.get('/en/streaming/') self.assertContains(response, "Yes/No") @override_settings( MIDDLEWARE_CLASSES=[ 'django.contrib.sessions.middleware.SessionMiddleware', 'django.middleware.locale.LocaleMiddleware', 'django.middleware.common.CommonMiddleware', ], ) def test_language_not_saved_to_session(self): """Checks that current language is not automatically saved to session on every request.""" # Regression test for #21473 self.client.get('/fr/simple/') self.assertNotIn(LANGUAGE_SESSION_KEY, self.client.session) @override_settings( USE_I18N=True, LANGUAGES=[ ('bg', 'Bulgarian'), ('en-us', 'English'), ('pt-br', 'Portugese (Brazil)'), ], MIDDLEWARE_CLASSES=[ 'django.middleware.locale.LocaleMiddleware', 'django.middleware.common.CommonMiddleware', ], ROOT_URLCONF='i18n.urls' ) class CountrySpecificLanguageTests(SimpleTestCase): def setUp(self): super(CountrySpecificLanguageTests, self).setUp() self.rf = RequestFactory() def test_check_for_language(self): self.assertTrue(check_for_language('en')) self.assertTrue(check_for_language('en-us')) self.assertTrue(check_for_language('en-US')) self.assertTrue(check_for_language('be')) self.assertTrue(check_for_language('be@latin')) self.assertTrue(check_for_language('sr-RS@latin')) self.assertTrue(check_for_language('sr-RS@12345')) self.assertFalse(check_for_language('en-ü')) self.assertFalse(check_for_language('en\x00')) self.assertFalse(check_for_language(None)) self.assertFalse(check_for_language('be@ ')) # Specifying encoding is not supported (Django enforces UTF-8) self.assertFalse(check_for_language('tr-TR.UTF-8')) self.assertFalse(check_for_language('tr-TR.UTF8')) self.assertFalse(check_for_language('de-DE.utf-8')) def test_get_language_from_request(self): # issue 19919 r = self.rf.get('/') r.COOKIES = {} r.META = {'HTTP_ACCEPT_LANGUAGE': 'en-US,en;q=0.8,bg;q=0.6,ru;q=0.4'} lang = get_language_from_request(r) self.assertEqual('en-us', lang) r = self.rf.get('/') r.COOKIES = {} r.META = {'HTTP_ACCEPT_LANGUAGE': 'bg-bg,en-US;q=0.8,en;q=0.6,ru;q=0.4'} lang = get_language_from_request(r) self.assertEqual('bg', lang) def test_specific_language_codes(self): # issue 11915 r = self.rf.get('/') r.COOKIES = {} r.META = {'HTTP_ACCEPT_LANGUAGE': 'pt,en-US;q=0.8,en;q=0.6,ru;q=0.4'} lang = get_language_from_request(r) self.assertEqual('pt-br', lang) r = self.rf.get('/') r.COOKIES = {} r.META = {'HTTP_ACCEPT_LANGUAGE': 'pt-pt,en-US;q=0.8,en;q=0.6,ru;q=0.4'} lang = get_language_from_request(r) self.assertEqual('pt-br', lang) class TranslationFilesMissing(SimpleTestCase): def setUp(self): super(TranslationFilesMissing, self).setUp() self.gettext_find_builtin = gettext_module.find def tearDown(self): gettext_module.find = self.gettext_find_builtin super(TranslationFilesMissing, self).tearDown() def patchGettextFind(self): gettext_module.find = lambda *args, **kw: None def test_failure_finding_default_mo_files(self): ''' Ensure IOError is raised if the default language is unparseable. Refs: #18192 ''' self.patchGettextFind() trans_real._translations = {} self.assertRaises(IOError, activate, 'en')
bthirion/scikit-learn
refs/heads/master
examples/neighbors/plot_regression.py
349
""" ============================ Nearest Neighbors regression ============================ Demonstrate the resolution of a regression problem using a k-Nearest Neighbor and the interpolation of the target using both barycenter and constant weights. """ print(__doc__) # Author: Alexandre Gramfort <alexandre.gramfort@inria.fr> # Fabian Pedregosa <fabian.pedregosa@inria.fr> # # License: BSD 3 clause (C) INRIA ############################################################################### # Generate sample data import numpy as np import matplotlib.pyplot as plt from sklearn import neighbors np.random.seed(0) X = np.sort(5 * np.random.rand(40, 1), axis=0) T = np.linspace(0, 5, 500)[:, np.newaxis] y = np.sin(X).ravel() # Add noise to targets y[::5] += 1 * (0.5 - np.random.rand(8)) ############################################################################### # Fit regression model n_neighbors = 5 for i, weights in enumerate(['uniform', 'distance']): knn = neighbors.KNeighborsRegressor(n_neighbors, weights=weights) y_ = knn.fit(X, y).predict(T) plt.subplot(2, 1, i + 1) plt.scatter(X, y, c='k', label='data') plt.plot(T, y_, c='g', label='prediction') plt.axis('tight') plt.legend() plt.title("KNeighborsRegressor (k = %i, weights = '%s')" % (n_neighbors, weights)) plt.show()
rimbalinux/LMD3
refs/heads/master
pygments/styles/trac.py
75
# -*- coding: utf-8 -*- """ pygments.styles.trac ~~~~~~~~~~~~~~~~~~~~ Port of the default trac highlighter design. :copyright: Copyright 2006-2010 by the Pygments team, see AUTHORS. :license: BSD, see LICENSE for details. """ from pygments.style import Style from pygments.token import Keyword, Name, Comment, String, Error, \ Number, Operator, Generic, Whitespace class TracStyle(Style): """ Port of the default trac highlighter design. """ default_style = '' styles = { Whitespace: '#bbbbbb', Comment: 'italic #999988', Comment.Preproc: 'bold noitalic #999999', Comment.Special: 'bold #999999', Operator: 'bold', String: '#bb8844', String.Regex: '#808000', Number: '#009999', Keyword: 'bold', Keyword.Type: '#445588', Name.Builtin: '#999999', Name.Function: 'bold #990000', Name.Class: 'bold #445588', Name.Exception: 'bold #990000', Name.Namespace: '#555555', Name.Variable: '#008080', Name.Constant: '#008080', Name.Tag: '#000080', Name.Attribute: '#008080', Name.Entity: '#800080', Generic.Heading: '#999999', Generic.Subheading: '#aaaaaa', Generic.Deleted: 'bg:#ffdddd #000000', Generic.Inserted: 'bg:#ddffdd #000000', Generic.Error: '#aa0000', Generic.Emph: 'italic', Generic.Strong: 'bold', Generic.Prompt: '#555555', Generic.Output: '#888888', Generic.Traceback: '#aa0000', Error: 'bg:#e3d2d2 #a61717' }
saikat007/GNOME_glib
refs/heads/master
gio/gdbus-2.0/codegen/__init__.py
33
# -*- Mode: Python -*- # GDBus - GLib D-Bus Library # # Copyright (C) 2008-2011 Red Hat, Inc. # # This library is free software; you can redistribute it and/or # modify it under the terms of the GNU Lesser General Public # License as published by the Free Software Foundation; either # version 2 of the License, or (at your option) any later version. # # This library is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # Lesser General Public License for more details. # # You should have received a copy of the GNU Lesser General # Public License along with this library; if not, see <http://www.gnu.org/licenses/>. # # Author: David Zeuthen <davidz@redhat.com> import os builddir = os.environ.get('UNINSTALLED_GLIB_BUILDDIR') if builddir is not None: __path__.append(os.path.abspath(os.path.join(builddir, 'gio', 'gdbus-2.0', 'codegen')))
palerdot/calibre
refs/heads/master
manual/latex.py
5
#!/usr/bin/env python # vim:fileencoding=UTF-8:ts=4:sw=4:sta:et:sts=4:ai from __future__ import with_statement __license__ = 'GPL v3' __copyright__ = '2009, Kovid Goyal <kovid@kovidgoyal.net>' __docformat__ = 'restructuredtext en' import os from sphinx.builders.latex import LaTeXBuilder class LaTeXHelpBuilder(LaTeXBuilder): name = 'mylatex' def finish(self): LaTeXBuilder.finish(self) self.info('Fixing Cyrillic characters...') tex = os.path.join(self.outdir, 'calibre.tex') with open(tex, 'r+b') as f: raw = f.read() for x in (b'Михаил Горбачёв', b'Фёдор Миха́йлович Достоевский'): raw = raw.replace(x, br'{\fontencoding{T2A}\selectfont %s}' % (x.replace(b'а́', b'a'))) f.seek(0) f.write(raw)
phihag/youtube-dl
refs/heads/master
youtube_dl/extractor/viddler.py
46
from __future__ import unicode_literals from .common import InfoExtractor from ..compat import ( compat_urllib_parse_urlencode, compat_urlparse, ) from ..utils import ( float_or_none, int_or_none, sanitized_Request, ) class ViddlerIE(InfoExtractor): _VALID_URL = r'https?://(?:www\.)?viddler\.com/(?:v|embed|player)/(?P<id>[a-z0-9]+)' _TESTS = [{ 'url': 'http://www.viddler.com/v/43903784', 'md5': '9eee21161d2c7f5b39690c3e325fab2f', 'info_dict': { 'id': '43903784', 'ext': 'mov', 'title': 'Video Made Easy', 'description': 'md5:6a697ebd844ff3093bd2e82c37b409cd', 'uploader': 'viddler', 'timestamp': 1335371429, 'upload_date': '20120425', 'duration': 100.89, 'thumbnail': r're:^https?://.*\.jpg$', 'view_count': int, 'comment_count': int, 'categories': ['video content', 'high quality video', 'video made easy', 'how to produce video with limited resources', 'viddler'], } }, { 'url': 'http://www.viddler.com/v/4d03aad9/', 'md5': 'f12c5a7fa839c47a79363bfdf69404fb', 'info_dict': { 'id': '4d03aad9', 'ext': 'ts', 'title': 'WALL-TO-GORTAT', 'upload_date': '20150126', 'uploader': 'deadspin', 'timestamp': 1422285291, 'view_count': int, 'comment_count': int, } }, { 'url': 'http://www.viddler.com/player/221ebbbd/0/', 'md5': '740511f61d3d1bb71dc14a0fe01a1c10', 'info_dict': { 'id': '221ebbbd', 'ext': 'mov', 'title': 'LETeens-Grammar-snack-third-conditional', 'description': ' ', 'upload_date': '20140929', 'uploader': 'BCLETeens', 'timestamp': 1411997190, 'view_count': int, 'comment_count': int, } }, { # secret protected 'url': 'http://www.viddler.com/v/890c0985?secret=34051570', 'info_dict': { 'id': '890c0985', 'ext': 'mp4', 'title': 'Complete Property Training - Traineeships', 'description': ' ', 'upload_date': '20130606', 'uploader': 'TiffanyBowtell', 'timestamp': 1370496993, 'view_count': int, 'comment_count': int, }, 'params': { 'skip_download': True, }, }] def _real_extract(self, url): video_id = self._match_id(url) query = { 'video_id': video_id, 'key': 'v0vhrt7bg2xq1vyxhkct', } qs = compat_urlparse.parse_qs(compat_urlparse.urlparse(url).query) secret = qs.get('secret', [None])[0] if secret: query['secret'] = secret headers = {'Referer': 'http://static.cdn-ec.viddler.com/js/arpeggio/v2/embed.html'} request = sanitized_Request( 'http://api.viddler.com/api/v2/viddler.videos.getPlaybackDetails.json?%s' % compat_urllib_parse_urlencode(query), None, headers) data = self._download_json(request, video_id)['video'] formats = [] for filed in data['files']: if filed.get('status', 'ready') != 'ready': continue format_id = filed.get('profile_id') or filed['profile_name'] f = { 'format_id': format_id, 'format_note': filed['profile_name'], 'url': self._proto_relative_url(filed['url']), 'width': int_or_none(filed.get('width')), 'height': int_or_none(filed.get('height')), 'filesize': int_or_none(filed.get('size')), 'ext': filed.get('ext'), 'source_preference': -1, } formats.append(f) if filed.get('cdn_url'): f = f.copy() f['url'] = self._proto_relative_url(filed['cdn_url'], 'http:') f['format_id'] = format_id + '-cdn' f['source_preference'] = 1 formats.append(f) if filed.get('html5_video_source'): f = f.copy() f['url'] = self._proto_relative_url(filed['html5_video_source']) f['format_id'] = format_id + '-html5' f['source_preference'] = 0 formats.append(f) self._sort_formats(formats) categories = [ t.get('text') for t in data.get('tags', []) if 'text' in t] return { 'id': video_id, 'title': data['title'], 'formats': formats, 'description': data.get('description'), 'timestamp': int_or_none(data.get('upload_time')), 'thumbnail': self._proto_relative_url(data.get('thumbnail_url')), 'uploader': data.get('author'), 'duration': float_or_none(data.get('length')), 'view_count': int_or_none(data.get('view_count')), 'comment_count': int_or_none(data.get('comment_count')), 'categories': categories, }
mancoast/CPythonPyc_test
refs/heads/master
cpython/253_test_socket_ssl.py
9
# Test just the SSL support in the socket module, in a moderately bogus way. import sys from test import test_support import socket import errno # Optionally test SSL support. This requires the 'network' resource as given # on the regrtest command line. skip_expected = not (test_support.is_resource_enabled('network') and hasattr(socket, "ssl")) def test_basic(): test_support.requires('network') import urllib if test_support.verbose: print "test_basic ..." socket.RAND_status() try: socket.RAND_egd(1) except TypeError: pass else: print "didn't raise TypeError" socket.RAND_add("this is a random string", 75.0) try: f = urllib.urlopen('https://sf.net') except IOError, exc: if exc.errno == errno.ETIMEDOUT: raise test_support.ResourceDenied('HTTPS connection is timing out') else: raise buf = f.read() f.close() def test_timeout(): test_support.requires('network') def error_msg(extra_msg): print >> sys.stderr, """\ WARNING: an attempt to connect to %r %s, in test_timeout. That may be legitimate, but is not the outcome we hoped for. If this message is seen often, test_timeout should be changed to use a more reliable address.""" % (ADDR, extra_msg) if test_support.verbose: print "test_timeout ..." # A service which issues a welcome banner (without need to write # anything). ADDR = "pop.gmail.com", 995 s = socket.socket() s.settimeout(30.0) try: s.connect(ADDR) except socket.timeout: error_msg('timed out') return except socket.error, exc: # In case connection is refused. if exc.args[0] == errno.ECONNREFUSED: error_msg('was refused') return else: raise ss = socket.ssl(s) # Read part of return welcome banner twice. ss.read(1) ss.read(1) s.close() def test_rude_shutdown(): if test_support.verbose: print "test_rude_shutdown ..." try: import threading except ImportError: return # Some random port to connect to. PORT = [9934] listener_ready = threading.Event() listener_gone = threading.Event() # `listener` runs in a thread. It opens a socket listening on PORT, and # sits in an accept() until the main thread connects. Then it rudely # closes the socket, and sets Event `listener_gone` to let the main thread # know the socket is gone. def listener(): s = socket.socket() PORT[0] = test_support.bind_port(s, '', PORT[0]) s.listen(5) listener_ready.set() s.accept() s = None # reclaim the socket object, which also closes it listener_gone.set() def connector(): listener_ready.wait() s = socket.socket() s.connect(('localhost', PORT[0])) listener_gone.wait() try: ssl_sock = socket.ssl(s) except socket.sslerror: pass else: raise test_support.TestFailed( 'connecting to closed SSL socket should have failed') t = threading.Thread(target=listener) t.start() connector() t.join() def test_main(): if not hasattr(socket, "ssl"): raise test_support.TestSkipped("socket module has no ssl support") test_rude_shutdown() test_basic() test_timeout() if __name__ == "__main__": test_main()
acsone/sale-workflow
refs/heads/8.0
sale_reason_to_export/models/sale_order.py
33
# -*- encoding: utf-8 -*- ############################################################################## # # Odoo, Open Source Management Solution # This module copyright (C) 2015 JPJ # ( http://www.savoirfairelinux.com). # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as # published by the Free Software Foundation, either version 3 of the # License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. # ############################################################################## from openerp import models, fields class SaleOrder(models.Model): _inherit = "sale.order" reason_to_export_id = fields.Many2one( 'sale.reason_to_export', 'Reason to Export', )
chen2aaron/SnirteneCodes
refs/heads/master
PythonCookbookPractise/chapter8/create_managed_attributes.py
1
# 8.6. Creating Managed Attributes class Person: def __init__(self, first_name): self.first_name = first_name # Getter function @property def first_name(self): return self._first_name # Setter function @first_name.setter def first_name(self, value): if not isinstance(value, str): raise TypeError('Expected a string') self._first_name = value # Deleter function (optional) @first_name.deleter def first_name(self): raise AttributeError("Can't delete attribute") class Person2: def __init__(self, first_name): self.set_first_name(first_name) # Getter function def get_first_name(self): return self._first_name # Setter function def set_first_name(self, value): if not isinstance(value, str): raise TypeError('Expected a string') self._first_name = value # Deleter function (optional) def del_first_name(self): raise AttributeError("Can't delete attribute") # Make a property from existing get/set methods name = property(get_first_name, set_first_name, del_first_name) p = Person('Ver')
pyfa-org/eos
refs/heads/master
eos/stats_container/slots.py
1
# ============================================================================== # Copyright (C) 2011 Diego Duclos # Copyright (C) 2011-2018 Anton Vorobyov # # This file is part of Eos. # # Eos is free software: you can redistribute it and/or modify # it under the terms of the GNU Lesser General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # Eos is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Lesser General Public License for more details. # # You should have received a copy of the GNU Lesser General Public License # along with Eos. If not, see <http://www.gnu.org/licenses/>. # ============================================================================== from collections import namedtuple SlotStats = namedtuple('SlotStats', ('used', 'total'))
12520054/pybot
refs/heads/master
chatterbot/ext/django_chatterbot/management/commands/train.py
1
from django.core.management.base import BaseCommand class Command(BaseCommand): help = 'Trains the database used by the chat bot' can_import_settings = True def handle(self, *args, **options): from chatterbot import ChatBot from chatterbot.ext.django_chatterbot import settings chatterbot = ChatBot(**settings.CHATTERBOT) chatterbot.train(chatterbot.training_data) # Django 1.8 does not define SUCCESS if hasattr(self.style, 'SUCCESS'): style = self.style.SUCCESS else: style = self.style.NOTICE training_class = chatterbot.trainer.__class__.__name__ self.stdout.write(style('ChatterBot trained using "%s"' % training_class))
hkariti/mopidy
refs/heads/develop
tests/m3u/test_translator.py
9
# encoding: utf-8 from __future__ import absolute_import, unicode_literals import os import tempfile import unittest from mopidy.internal import path from mopidy.m3u import translator from mopidy.models import Track from tests import path_to_data_dir data_dir = path_to_data_dir('') song1_path = path_to_data_dir('song1.mp3') song2_path = path_to_data_dir('song2.mp3') song3_path = path_to_data_dir('φοο.mp3') encoded_path = path_to_data_dir('æøå.mp3') song1_uri = path.path_to_uri(song1_path) song2_uri = path.path_to_uri(song2_path) song3_uri = path.path_to_uri(song3_path) encoded_uri = path.path_to_uri(encoded_path) song1_track = Track(uri=song1_uri) song2_track = Track(uri=song2_uri) song3_track = Track(uri=song3_uri) encoded_track = Track(uri=encoded_uri) song1_ext_track = song1_track.replace(name='song1') song2_ext_track = song2_track.replace(name='song2', length=60000) encoded_ext_track = encoded_track.replace(name='æøå') # FIXME use mock instead of tempfile.NamedTemporaryFile class M3UToUriTest(unittest.TestCase): def parse(self, name): return translator.parse_m3u(name, data_dir) def test_empty_file(self): tracks = self.parse(path_to_data_dir('empty.m3u')) self.assertEqual([], tracks) def test_basic_file(self): tracks = self.parse(path_to_data_dir('one.m3u')) self.assertEqual([song1_track], tracks) def test_file_with_comment(self): tracks = self.parse(path_to_data_dir('comment.m3u')) self.assertEqual([song1_track], tracks) def test_file_is_relative_to_correct_dir(self): with tempfile.NamedTemporaryFile(delete=False) as tmp: tmp.write('song1.mp3') try: tracks = self.parse(tmp.name) self.assertEqual([song1_track], tracks) finally: if os.path.exists(tmp.name): os.remove(tmp.name) def test_file_with_absolute_files(self): with tempfile.NamedTemporaryFile(delete=False) as tmp: tmp.write(song1_path) try: tracks = self.parse(tmp.name) self.assertEqual([song1_track], tracks) finally: if os.path.exists(tmp.name): os.remove(tmp.name) def test_file_with_multiple_absolute_files(self): with tempfile.NamedTemporaryFile(delete=False) as tmp: tmp.write(song1_path + '\n') tmp.write('# comment \n') tmp.write(song2_path) try: tracks = self.parse(tmp.name) self.assertEqual([song1_track, song2_track], tracks) finally: if os.path.exists(tmp.name): os.remove(tmp.name) def test_file_with_uri(self): with tempfile.NamedTemporaryFile(delete=False) as tmp: tmp.write(song1_uri) try: tracks = self.parse(tmp.name) self.assertEqual([song1_track], tracks) finally: if os.path.exists(tmp.name): os.remove(tmp.name) def test_encoding_is_latin1(self): tracks = self.parse(path_to_data_dir('encoding.m3u')) self.assertEqual([encoded_track], tracks) def test_open_missing_file(self): tracks = self.parse(path_to_data_dir('non-existant.m3u')) self.assertEqual([], tracks) def test_empty_ext_file(self): tracks = self.parse(path_to_data_dir('empty-ext.m3u')) self.assertEqual([], tracks) def test_basic_ext_file(self): tracks = self.parse(path_to_data_dir('one-ext.m3u')) self.assertEqual([song1_ext_track], tracks) def test_multi_ext_file(self): tracks = self.parse(path_to_data_dir('two-ext.m3u')) self.assertEqual([song1_ext_track, song2_ext_track], tracks) def test_ext_file_with_comment(self): tracks = self.parse(path_to_data_dir('comment-ext.m3u')) self.assertEqual([song1_ext_track], tracks) def test_ext_encoding_is_latin1(self): tracks = self.parse(path_to_data_dir('encoding-ext.m3u')) self.assertEqual([encoded_ext_track], tracks) def test_m3u8_file(self): with tempfile.NamedTemporaryFile(suffix='.m3u8', delete=False) as tmp: tmp.write(song3_path) try: tracks = self.parse(tmp.name) self.assertEqual([song3_track], tracks) finally: if os.path.exists(tmp.name): os.remove(tmp.name) class URItoM3UTest(unittest.TestCase): pass
whip112/Whip112
refs/heads/master
vendor/packages/logilab/astng/setup.py
24
#!/usr/bin/env python # pylint: disable=W0404,W0622,W0704,W0613 # copyright 2003-2013 LOGILAB S.A. (Paris, FRANCE), all rights reserved. # contact http://www.logilab.fr/ -- mailto:contact@logilab.fr # # This file is part of logilab-astng. # # logilab-astng is free software: you can redistribute it and/or modify it under # the terms of the GNU Lesser General Public License as published by the Free # Software Foundation, either version 2.1 of the License, or (at your option) any # later version. # # logilab-astng is distributed in the hope that it will be useful, but WITHOUT # ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or FITNESS # FOR A PARTICULAR PURPOSE. See the GNU Lesser General Public License for more # details. # # You should have received a copy of the GNU Lesser General Public License along # with logilab-astng. If not, see <http://www.gnu.org/licenses/>. """Generic Setup script, takes package info from __pkginfo__.py file. """ __docformat__ = "restructuredtext en" import os import sys import shutil from os.path import isdir, exists, join try: if os.environ.get('NO_SETUPTOOLS'): raise ImportError() from setuptools import setup from setuptools.command import install_lib USE_SETUPTOOLS = 1 except ImportError: from distutils.core import setup from distutils.command import install_lib USE_SETUPTOOLS = 0 try: # python3 from distutils.command.build_py import build_py_2to3 as build_py except ImportError: # python2.x from distutils.command.build_py import build_py sys.modules.pop('__pkginfo__', None) # import optional features __pkginfo__ = __import__("__pkginfo__") # import required features from __pkginfo__ import modname, version, license, description, \ web, author, author_email distname = getattr(__pkginfo__, 'distname', modname) scripts = getattr(__pkginfo__, 'scripts', []) data_files = getattr(__pkginfo__, 'data_files', None) subpackage_of = getattr(__pkginfo__, 'subpackage_of', None) include_dirs = getattr(__pkginfo__, 'include_dirs', []) ext_modules = getattr(__pkginfo__, 'ext_modules', None) install_requires = getattr(__pkginfo__, 'install_requires', None) dependency_links = getattr(__pkginfo__, 'dependency_links', []) classifiers = getattr(__pkginfo__, 'classifiers', []) STD_BLACKLIST = ('CVS', '.svn', '.hg', 'debian', 'dist', 'build') IGNORED_EXTENSIONS = ('.pyc', '.pyo', '.elc', '~') if exists('README'): long_description = open('README').read() else: long_description = '' def ensure_scripts(linux_scripts): """Creates the proper script names required for each platform (taken from 4Suite) """ from distutils import util if util.get_platform()[:3] == 'win': scripts_ = [script + '.bat' for script in linux_scripts] else: scripts_ = linux_scripts return scripts_ def get_packages(directory, prefix): """return a list of subpackages for the given directory""" result = [] for package in os.listdir(directory): absfile = join(directory, package) if isdir(absfile): if exists(join(absfile, '__init__.py')) or \ package in ('test', 'tests'): if prefix: result.append('%s.%s' % (prefix, package)) else: result.append(package) result += get_packages(absfile, result[-1]) return result EMPTY_FILE = '''"""generated file, don't modify or your data will be lost""" try: __import__('pkg_resources').declare_namespace(__name__) except ImportError: pass ''' class MyInstallLib(install_lib.install_lib): """extend install_lib command to handle package __init__.py if necessary """ def run(self): """overridden from install_lib class""" install_lib.install_lib.run(self) # create Products.__init__.py if needed if subpackage_of: product_init = join(self.install_dir, subpackage_of, '__init__.py') if not exists(product_init): self.announce('creating %s' % product_init) stream = open(product_init, 'w') stream.write(EMPTY_FILE) stream.close() class MyBuildPy(build_py): """extend build_by command to handle include_dirs variable if necessary """ def run(self): """overridden from install_lib class""" build_py.run(self) # manually install included directories if any if include_dirs: if subpackage_of: base = join(subpackage_of, modname) else: base = modname basedir = os.path.join(self.build_lib, base) for directory in include_dirs: dest = join(basedir, directory) shutil.rmtree(dest, ignore_errors=True) shutil.copytree(directory, dest) if sys.version_info >= (3, 0): # process manually python file in include_dirs (test data) from subprocess import check_call print('running 2to3 on', dest) # parens are NOT optional here for py3k compat check_call(['2to3', '-wn', dest]) def install(**kwargs): """setup entry point""" if USE_SETUPTOOLS: if '--force-manifest' in sys.argv: sys.argv.remove('--force-manifest') # install-layout option was introduced in 2.5.3-1~exp1 elif sys.version_info < (2, 5, 4) and '--install-layout=deb' in sys.argv: sys.argv.remove('--install-layout=deb') if subpackage_of: package = subpackage_of + '.' + modname kwargs['package_dir'] = {package : '.'} packages = [package] + get_packages(os.getcwd(), package) if USE_SETUPTOOLS: kwargs['namespace_packages'] = [subpackage_of] else: kwargs['package_dir'] = {modname : '.'} packages = [modname] + get_packages(os.getcwd(), modname) if USE_SETUPTOOLS and install_requires: kwargs['install_requires'] = install_requires kwargs['dependency_links'] = dependency_links kwargs['packages'] = packages return setup(name = distname, version = version, license = license, description = description, long_description = long_description, classifiers = classifiers, author = author, author_email = author_email, url = web, scripts = ensure_scripts(scripts), data_files = data_files, ext_modules = ext_modules, cmdclass = {'install_lib': MyInstallLib, 'build_py': MyBuildPy}, **kwargs ) if __name__ == '__main__' : install()
doismellburning/edx-platform
refs/heads/master
cms/djangoapps/contentstore/features/discussion-editor.py
153
# disable missing docstring # pylint: disable=missing-docstring from lettuce import world, step @step('I have created a Discussion Tag$') def i_created_discussion_tag(step): step.given('I am in Studio editing a new unit') world.create_component_instance( step=step, category='discussion', ) @step('I see three alphabetized settings and their expected values$') def i_see_only_the_settings_and_values(step): world.verify_all_setting_entries( [ ['Category', "Week 1", False], ['Display Name', "Discussion", False], ['Subcategory', "Topic-Level Student-Visible Label", False] ]) @step('I edit the component$') def i_edit_and_select_settings(_step): world.edit_component()
KaranToor/MA450
refs/heads/master
google-cloud-sdk/lib/surface/genomics/variantsets/export.py
3
# Copyright 2015 Google Inc. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Implementation of gcloud genomics variantsets export. """ from googlecloudsdk.api_lib.genomics import genomics_util from googlecloudsdk.calliope import arg_parsers from googlecloudsdk.calliope import base class Export(base.Command): """Exports data from a variant set to an external destination. """ @staticmethod def Args(parser): """Register flags for this command.""" parser.add_argument( 'id', type=str, help='The ID of the variant set to export.') parser.add_argument( 'table', type=str, help=('The BigQuery table to which data is exported. If this table ' 'does not exist, it will be created. If it already exists, it ' 'will be overwritten.')) parser.add_argument( '--call-set-ids', type=arg_parsers.ArgList(), help=('If provided, only variant call information ' 'from the specified call sets will be exported. ' 'By default all variant call sets are exported. ')) # TODO(user) Use resource parser for beta b/22357346 parser.add_argument( '--bigquery-project', type=str, help=('The Google Cloud project ID that owns the destination BigQuery ' 'dataset. The caller must have WRITE access to this project. ' 'This project will also own the resulting export job.' 'If not supplied, defaults to the gcloud project ID.')) parser.add_argument( '--bigquery-dataset', type=str, required=True, help=('The BigQuery dataset to which data is exported. This dataset ' 'must already exist. Note that this is distinct from the ' 'Genomics concept of "dataset."')) def Run(self, args): """This is what gets called when the user runs this command. Args: args: an argparse namespace, All the arguments that were provided to this command invocation. Raises: HttpException: An http error response was received while executing api request. Returns: an ExportVariantSetResponse """ apitools_client = genomics_util.GetGenomicsClient() genomics_messages = genomics_util.GetGenomicsMessages() enum = genomics_messages.ExportVariantSetRequest.FormatValueValuesEnum call_set_ids = args.call_set_ids if args.call_set_ids else [] project_id = args.bigquery_project if not project_id: project_id = genomics_util.GetProjectId() request = genomics_messages.GenomicsVariantsetsExportRequest( variantSetId=args.id, exportVariantSetRequest=genomics_messages.ExportVariantSetRequest( callSetIds=call_set_ids, projectId=project_id, format=enum.FORMAT_BIGQUERY, bigqueryDataset=args.bigquery_dataset, bigqueryTable=args.table)) return apitools_client.variantsets.Export(request)
stephen144/odoo
refs/heads/9.0
addons/mrp_repair/__openerp__.py
23
# -*- coding: utf-8 -*- # Part of Odoo. See LICENSE file for full copyright and licensing details. { 'name': 'Repairs Management', 'version': '1.0', 'sequence': 200, 'category': 'Manufacturing', 'summary': 'Repair broken or damaged products', 'description': """, The aim is to have a complete module to manage all products repairs. ==================================================================== The following topics should be covered by this module: ------------------------------------------------------ * Add/remove products in the reparation * Impact for stocks * Invoicing (products and/or services) * Warranty concept * Repair quotation report * Notes for the technician and for the final customer """, 'depends': ['stock', 'sale', 'account'], 'website': 'https://www.odoo.com/page/manufacturing', 'data': [ 'security/ir.model.access.csv', 'security/mrp_repair_security.xml', 'mrp_repair_sequence.xml', 'wizard/mrp_repair_cancel_view.xml', 'wizard/mrp_repair_make_invoice_view.xml', 'mrp_repair_view.xml', 'mrp_repair_workflow.xml', 'mrp_repair_report.xml', 'views/report_mrprepairorder.xml', ], 'demo': ['mrp_repair_demo.yml'], 'test': ['../account/test/account_minimal_test.xml', 'test/mrp_repair_users.yml', 'test/test_mrp_repair_noneinv.yml', 'test/test_mrp_repair_b4inv.yml', 'test/test_mrp_repair_afterinv.yml', 'test/test_mrp_repair_cancel.yml', 'test/test_mrp_repair_fee.yml', ], 'installable': True, 'auto_install': False, 'application': True, }
berycoin-project/berycoin
refs/heads/master
contrib/devtools/test-security-check.py
119
#!/usr/bin/env python2 # Copyright (c) 2015-2016 The Bitcoin Core developers # Distributed under the MIT software license, see the accompanying # file COPYING or http://www.opensource.org/licenses/mit-license.php. ''' Test script for security-check.py ''' from __future__ import division,print_function import subprocess import unittest def write_testcode(filename): with open(filename, 'w') as f: f.write(''' #include <stdio.h> int main() { printf("the quick brown fox jumps over the lazy god\\n"); return 0; } ''') def call_security_check(cc, source, executable, options): subprocess.check_call([cc,source,'-o',executable] + options) p = subprocess.Popen(['./security-check.py',executable], stdout=subprocess.PIPE, stderr=subprocess.PIPE, stdin=subprocess.PIPE) (stdout, stderr) = p.communicate() return (p.returncode, stdout.rstrip()) class TestSecurityChecks(unittest.TestCase): def test_ELF(self): source = 'test1.c' executable = 'test1' cc = 'gcc' write_testcode(source) self.assertEqual(call_security_check(cc, source, executable, ['-Wl,-zexecstack','-fno-stack-protector','-Wl,-znorelro']), (1, executable+': failed PIE NX RELRO Canary')) self.assertEqual(call_security_check(cc, source, executable, ['-Wl,-znoexecstack','-fno-stack-protector','-Wl,-znorelro']), (1, executable+': failed PIE RELRO Canary')) self.assertEqual(call_security_check(cc, source, executable, ['-Wl,-znoexecstack','-fstack-protector-all','-Wl,-znorelro']), (1, executable+': failed PIE RELRO')) self.assertEqual(call_security_check(cc, source, executable, ['-Wl,-znoexecstack','-fstack-protector-all','-Wl,-znorelro','-pie','-fPIE']), (1, executable+': failed RELRO')) self.assertEqual(call_security_check(cc, source, executable, ['-Wl,-znoexecstack','-fstack-protector-all','-Wl,-zrelro','-Wl,-z,now','-pie','-fPIE']), (0, '')) def test_PE(self): source = 'test1.c' executable = 'test1.exe' cc = 'i686-w64-mingw32-gcc' write_testcode(source) self.assertEqual(call_security_check(cc, source, executable, []), (1, executable+': failed PIE NX')) self.assertEqual(call_security_check(cc, source, executable, ['-Wl,--nxcompat']), (1, executable+': failed PIE')) self.assertEqual(call_security_check(cc, source, executable, ['-Wl,--nxcompat','-Wl,--dynamicbase']), (0, '')) if __name__ == '__main__': unittest.main()
Aninstance/invoicing
refs/heads/master
invoicing/signals.py
1
from django.dispatch import receiver from django.db.models.signals import post_save, pre_save from invoicing.models import * from haystack.management.commands import update_index from django.utils import timezone import invoicing.views as invoicing from invoicing import schedules from django.forms import ValidationError from django.utils.translation import ugettext_lazy as _ """ Note: if doing a save in post_save ensure signal is disconnected to avoid an infinite loop of saving: e.g.: post_save.disconnect(my_method, sender=sender) instance.save() post_save.connect(my_method, sender=Invoice) """ @receiver(post_save, sender=Invoice) def invoice_post_save(sender, instance, created, **kwargs): """ post save receiver for Invoice model """ save_updated = False post_save.disconnect(invoice_post_save, sender=sender) if created: invoice_id = model_functions.generate_invoice_random_id() # generate random id instance.invoice_number = '{}-{}'.format(invoice_id, instance.id) save_updated = True # recurring invoice stuff if not getattr(instance, 'save_pdf', False): # if save not being called again just to update DB after gen of PDF # set the invoice as start/stop recurring recurring(instance=instance) # dispatch email if client email notifications set to true and the save is being called when saving a new PDF if instance.client.email_notifications and getattr(instance, 'save_pdf', False): if instance.invoice_status == Invoice.INVOICE_STATUS[5][0]: # receipt (paid in full) # send email if not getattr(instance, 'receipt_emailed', False): # if hasn't already been sent invoicing.pdf_gen_or_fetch_or_email(invoice_number=instance.invoice_number, type=invoicing.PDF_TYPES.get('receipt'), email=True, regenerate=False) # mark as sent setattr(instance, 'receipt_emailed', True) save_updated = True elif instance.invoice_status == Invoice.INVOICE_STATUS[1][0]: # invoice (unpaid) # send email if not getattr(instance, 'invoice_emailed', False): # if hasn't already been sent invoicing.pdf_gen_or_fetch_or_email(invoice_number=instance.invoice_number, type=invoicing.PDF_TYPES.get('invoice'), email=True, regenerate=False) # mark as sent setattr(instance, 'invoice_emailed', True) # change status from issued to sent setattr(instance, 'invoice_status', Invoice.INVOICE_STATUS[2][0]) save_updated = True elif instance.invoice_status == Invoice.INVOICE_STATUS[4][0]: # invoice (partially paid) # send email invoicing.pdf_gen_or_fetch_or_email(invoice_number=instance.invoice_number, type=invoicing.PDF_TYPES.get('invoice_update'), email=True) # save the instance if something's been called ... if save_updated: # disable pre_save signal, as not required to be run again for the second save! pre_save.disconnect(invoice_pre_save, sender=Invoice) # save the instance instance.save() # re-enable pre_save signal pre_save.connect(invoice_pre_save, sender=Invoice) # re-enable post_save signal post_save.connect(invoice_post_save, sender=Invoice) @receiver(pre_save, sender=Invoice) def invoice_pre_save(sender, instance, **kwargs): """ Also to populate mark_as_paid field with datetime when a status is changed to 'PAID_IN_FULL' Also populates the "amount_paid" field with total. Also updates the invoice status to partially paid when an amount is paid """ try: # existing invoice to be modified inv = Invoice.objects.get(invoice_number=instance.invoice_number) except Invoice.DoesNotExist: # new invoice inv = Invoice() # generate an empty reference Invoice instance if no existing (i.e. A NEW INVOICE) # IF NOT SAVING PDF (Most stuff goes in here!) if not getattr(instance, 'save_pdf', False): # avoid running this pre_save if 'save_pdf' param added to instance # PRE-SAVE AMENDMENT STUFF instance_dict = invoicing.invoice_instance_to_dict(instance) # get instance as dict + sums # if invoice issued, save the time if getattr(instance, 'invoice_status') in dict(Invoice.INVOICE_STATUS[1:6]) and \ not getattr(inv, 'datetime_issued'): setattr(instance, 'datetime_issued', timezone.now()) # ensure invoice_status is upper case setattr(instance, 'invoice_status', instance.invoice_status.upper()) # # enter marked_as_paid datetime into database if status changed to marked as paid if instance.invoice_status == Invoice.INVOICE_STATUS[5][0]: if not inv.marked_as_paid: # if not originally marked as paid instance.marked_as_paid = timezone.now() # set as marked as paid # set paid_amount to total owed if status is set to paid in full instance.paid_amount = Decimal(instance_dict.get('Total after tax')) # change status if paid_amount is submitted if inv.paid_amount or instance.paid_amount: # if total paid >= total owed, set status to paid in full if inv.paid_amount >= Decimal(instance_dict.get('Total after tax')) or instance.paid_amount >= \ Decimal(instance_dict.get('Total after tax')): instance.invoice_status = Invoice.INVOICE_STATUS[5][0] # enter marked_as_paid datetime into database instance.marked_as_paid = timezone.now() else: # else set status to partially paid instance.invoice_status = Invoice.INVOICE_STATUS[4][0] # check for overdue status todo: move this to an automated django-q later date_due = getattr(instance, 'date_due', None) or getattr(inv, 'date_due') if date_due < timezone.now().date() and \ instance.invoice_status in dict(Invoice.INVOICE_STATUS[:5]): instance.overdue = True # set overdue to True if date_due < now else: instance.overdue = False # ensure set back to False if paid # todo: new feature - if amount paid exceeds amount owed, store client credit note ... @receiver(post_save, sender=Account) def change_uploaded_file_permissions(sender, instance, **kwargs): """ Changes the file permissions of uploaded media files to something sensible """ if getattr(instance, 'logo', False): os.chmod('{}'.format(instance.logo.file.name), 0o664) @receiver(post_save) # don't specify sender, so it is fired on all model saves def update_search_index(sender, instance, created, **kwargs): """ receiver to update the search index whenever a model is saved """ watched_models = [Invoice, Account] # if sender in watched but do NOT trigger when saving the model when saving PDF (no point) if sender in watched_models and not getattr(instance, 'save_pdf', False): update_index.Command().handle(interactive=False) def recurring(instance=None): """ Function to handle the creation of child invoices """ if instance.invoice_number: # ensure child recurring invoice does not produce its own child if int(getattr(instance, 'recurring', False)) not in dict(Invoice.RECURRING[:1]) and getattr(instance, 'parent_invoice', False): raise ValidationError(_('A child invoice cannot itself be recurring')) # ensure invoice is not its own parent try: if getattr(instance, 'invoice_number', False) == getattr(instance, 'parent_invoice', False).invoice_number: raise ValidationError(_('An invoice cannot be it\'s own parent ... stop messing around!')) except AttributeError: pass # thrown if no invoice_number for parent, so can discard as if no parent, no worries! # if status >= ISSUED, call the scheduler to start/stop recurring if getattr(instance, 'invoice_status', False) in dict(Invoice.INVOICE_STATUS[1:]): schedules.recurring_invoice_scheduler(instance=instance) return True return None
dgies/incubator-airflow
refs/heads/master
airflow/utils/compression.py
44
# -*- coding: utf-8 -*- # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. from tempfile import NamedTemporaryFile import shutil import gzip import bz2 def uncompress_file(input_file_name, file_extension, dest_dir): """ Uncompress gz and bz2 files """ if file_extension.lower() not in ('.gz', '.bz2'): raise NotImplementedError("Received {} format. Only gz and bz2 " "files can currently be uncompressed." .format(file_extension)) if file_extension.lower() == '.gz': fmodule = gzip.GzipFile elif file_extension.lower() == '.bz2': fmodule = bz2.BZ2File with fmodule(input_file_name, mode='rb') as f_compressed,\ NamedTemporaryFile(dir=dest_dir, mode='wb', delete=False) as f_uncompressed: shutil.copyfileobj(f_compressed, f_uncompressed) return f_uncompressed.name
toolmacher/micropython
refs/heads/master
tools/gen-cpydiff.py
7
# This file is part of the MicroPython project, http://micropython.org/ # # The MIT License (MIT) # # Copyright (c) 2016 Rami Ali # # Permission is hereby granted, free of charge, to any person obtaining a copy # of this software and associated documentation files (the "Software"), to deal # in the Software without restriction, including without limitation the rights # to use, copy, modify, merge, publish, distribute, sublicense, and/or sell # copies of the Software, and to permit persons to whom the Software is # furnished to do so, subject to the following conditions: # # The above copyright notice and this permission notice shall be included in # all copies or substantial portions of the Software. # # THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR # IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, # FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE # AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER # LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, # OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN # THE SOFTWARE. """ gen-cpydiff generates documentation which outlines operations that differ between MicroPython and CPython. This script is called by the docs Makefile for html and Latex and may be run manually using the command make gen-cpydiff. """ import os import errno import subprocess import time import re from collections import namedtuple # Micropython supports syntax of CPython 3.4 with some features from 3.5, and # such version should be used to test for differences. If your default python3 # executable is of lower version, you can point MICROPY_CPYTHON3 environment var # to the correct executable. if os.name == 'nt': CPYTHON3 = os.getenv('MICROPY_CPYTHON3', 'python3.exe') MICROPYTHON = os.getenv('MICROPY_MICROPYTHON', '../windows/micropython.exe') else: CPYTHON3 = os.getenv('MICROPY_CPYTHON3', 'python3') MICROPYTHON = os.getenv('MICROPY_MICROPYTHON', '../unix/micropython') TESTPATH = '../tests/cpydiff/' DOCPATH = '../docs/genrst/' INDEXTEMPLATE = '../docs/differences/index_template.txt' INDEX = 'index.rst' HEADER = '.. This document was generated by tools/gen-cpydiff.py\n\n' UIMPORTLIST = {'struct', 'collections', 'json'} CLASSMAP = {'Core': 'Core Language', 'Types': 'Builtin Types'} INDEXPRIORITY = ['syntax', 'core_language', 'builtin_types', 'modules'] RSTCHARS = ['=', '-', '~', '`', ':'] SPLIT = '"""\n|categories: |description: |cause: |workaround: ' TAB = ' ' Output = namedtuple('output', ['name', 'class_', 'desc', 'cause', 'workaround', 'code', 'output_cpy', 'output_upy', 'status']) def readfiles(): """ Reads test files """ tests = list(filter(lambda x: x.endswith('.py'), os.listdir(TESTPATH))) tests.sort() files = [] for test in tests: text = open(TESTPATH + test, 'r').read() try: class_, desc, cause, workaround, code = [x.rstrip() for x in \ list(filter(None, re.split(SPLIT, text)))] output = Output(test, class_, desc, cause, workaround, code, '', '', '') files.append(output) except IndexError: print('Incorrect format in file ' + TESTPATH + test) return files def uimports(code): """ converts CPython module names into MicroPython equivalents """ for uimport in UIMPORTLIST: uimport = bytes(uimport, 'utf8') code = code.replace(uimport, b'u' + uimport) return code def run_tests(tests): """ executes all tests """ results = [] for test in tests: with open(TESTPATH + test.name, 'rb') as f: input_cpy = f.read() input_upy = uimports(input_cpy) process = subprocess.Popen(CPYTHON3, shell=True, stdout=subprocess.PIPE, stdin=subprocess.PIPE, stderr=subprocess.PIPE) output_cpy = [com.decode('utf8') for com in process.communicate(input_cpy)] process = subprocess.Popen(MICROPYTHON, shell=True, stdout=subprocess.PIPE, stdin=subprocess.PIPE, stderr=subprocess.PIPE) output_upy = [com.decode('utf8') for com in process.communicate(input_upy)] if output_cpy[0] == output_upy[0] and output_cpy[1] == output_upy[1]: status = 'Supported' print('Supported operation!\nFile: ' + TESTPATH + test.name) else: status = 'Unsupported' output = Output(test.name, test.class_, test.desc, test.cause, test.workaround, test.code, output_cpy, output_upy, status) results.append(output) results.sort(key=lambda x: x.class_) return results def indent(block, spaces): """ indents paragraphs of text for rst formatting """ new_block = '' for line in block.split('\n'): new_block += spaces + line + '\n' return new_block def gen_table(contents): """ creates a table given any set of columns """ xlengths = [] ylengths = [] for column in contents: col_len = 0 for entry in column: lines = entry.split('\n') for line in lines: col_len = max(len(line) + 2, col_len) xlengths.append(col_len) for i in range(len(contents[0])): ymax = 0 for j in range(len(contents)): ymax = max(ymax, len(contents[j][i].split('\n'))) ylengths.append(ymax) table_divider = '+' + ''.join(['-' * i + '+' for i in xlengths]) + '\n' table = table_divider for i in range(len(ylengths)): row = [column[i] for column in contents] row = [entry + '\n' * (ylengths[i]-len(entry.split('\n'))) for entry in row] row = [entry.split('\n') for entry in row] for j in range(ylengths[i]): k = 0 for entry in row: width = xlengths[k] table += ''.join(['| {:{}}'.format(entry[j], width - 1)]) k += 1 table += '|\n' table += table_divider return table + '\n' def gen_rst(results): """ creates restructured text documents to display tests """ # make sure the destination directory exists try: os.mkdir(DOCPATH) except OSError as e: if e.args[0] != errno.EEXIST and e.args[0] != errno.EISDIR: raise toctree = [] class_ = [] for output in results: section = output.class_.split(',') for i in range(len(section)): section[i] = section[i].rstrip() if section[i] in CLASSMAP: section[i] = CLASSMAP[section[i]] if i >= len(class_) or section[i] != class_[i]: if i == 0: filename = section[i].replace(' ', '_').lower() rst = open(DOCPATH + filename + '.rst', 'w') rst.write(HEADER) rst.write(section[i] + '\n') rst.write(RSTCHARS[0] * len(section[i])) rst.write(time.strftime("\nGenerated %a %d %b %Y %X UTC\n\n", time.gmtime())) toctree.append(filename) else: rst.write(section[i] + '\n') rst.write(RSTCHARS[min(i, len(RSTCHARS)-1)] * len(section[i])) rst.write('\n\n') class_ = section rst.write('**' + output.desc + '**\n\n') if output.cause != 'Unknown': rst.write('**Cause:** ' + output.cause + '\n\n') if output.workaround != 'Unknown': rst.write('**Workaround:** ' + output.workaround + '\n\n') rst.write('Sample code::\n\n' + indent(output.code, TAB) + '\n') output_cpy = indent(''.join(output.output_cpy[0:2]), TAB).rstrip() output_cpy = ('::\n\n' if output_cpy != '' else '') + output_cpy output_upy = indent(''.join(output.output_upy[0:2]), TAB).rstrip() output_upy = ('::\n\n' if output_upy != '' else '') + output_upy table = gen_table([['CPy output:', output_cpy], ['uPy output:', output_upy]]) rst.write(table) template = open(INDEXTEMPLATE, 'r') index = open(DOCPATH + INDEX, 'w') index.write(HEADER) index.write(template.read()) for section in INDEXPRIORITY: if section in toctree: index.write(indent(section + '.rst', TAB)) toctree.remove(section) for section in toctree: index.write(indent(section + '.rst', TAB)) def main(): """ Main function """ # set search path so that test scripts find the test modules (and no other ones) os.environ['PYTHONPATH'] = TESTPATH os.environ['MICROPYPATH'] = TESTPATH files = readfiles() results = run_tests(files) gen_rst(results) main()
dxxb/micropython
refs/heads/upstream-tracking
tests/basics/dict_intern.py
117
# check that interned strings are compared against non-interned strings di = {"key1": "value"} # lookup interned string k = "key1" print(k in di) # lookup non-interned string k2 = "key" + "1" print(k == k2) print(k2 in di) # lookup non-interned string print("".join(['k', 'e', 'y', '1']) in di)
ol-loginov/intellij-community
refs/heads/master
python/testData/formatter/blankLineBeforeFunction_after.py
79
class C: x = 1 def foo(self): pass
stackforge/networking-mlnx
refs/heads/master
networking_mlnx/plugins/ml2/drivers/sdn/utils.py
2
# Copyright 2016 Mellanox Technologies, Ltd # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. def strings_to_url(*args): return "/".join(filter(None, args))
wolverineav/horizon
refs/heads/master
openstack_dashboard/dashboards/project/access_and_security/views.py
66
# Copyright 2012 United States Government as represented by the # Administrator of the National Aeronautics and Space Administration. # All Rights Reserved. # # Copyright 2012 Nebula, Inc. # Copyright 2012 OpenStack Foundation # # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. """ Views for Instances and Volumes. """ from django.utils.translation import ugettext_lazy as _ from horizon import tabs from openstack_dashboard.dashboards.project.access_and_security \ import tabs as project_tabs class IndexView(tabs.TabbedTableView): tab_group_class = project_tabs.AccessAndSecurityTabs template_name = 'project/access_and_security/index.html' page_title = _("Access & Security")
alexston/calibre-webserver
refs/heads/master
src/calibre/libwand.py
19
__license__ = 'GPL v3' __copyright__ = '2008, Kovid Goyal <kovid at kovidgoyal.net>' import ctypes, os, sys from calibre import iswindows, isosx class WandException(Exception): pass _lib_name = 'CORE_RL_wand_.dll' if iswindows else 'libWand.dylib' if isosx else 'libWand.so' if iswindows and hasattr(sys, 'frozen'): im_dir = os.path.join(os.path.dirname(sys.executable), 'ImageMagick') os.putenv('PATH', im_dir + ';' + os.environ['PATH']) _libwand = None try: _libwand = ctypes.cdll.LoadLibrary(_lib_name) except: pass class Severity(ctypes.c_long): pass class String(ctypes.c_char_p): def __del__(self): _libwand.MagickRelinquishMemory(self) def __str__(self): return self.value if _libwand is not None: _libwand.MagickGetException.argtypes = [ctypes.c_void_p, ctypes.POINTER(Severity)] _libwand.MagickGetException.restype = String def get_exception(wand): severity = Severity() desc = _libwand.MagickGetException(wand, ctypes.byref(severity)) return str(desc) def convert(source, dest): if _libwand is None: raise WandException('Could not find ImageMagick library') if not _libwand.MagickWandGenesis(): raise WandException('Unable to initialize Image Magick') wand = _libwand.NewMagickWand() if wand <= 0: raise WandException('Unable to initialize Image Magick. Cannot create wand.') if not _libwand.MagickReadImage(wand, source): raise WandException('Cannot read image %s: %s'%(source, get_exception(wand))) if not _libwand.MagickWriteImage(wand, dest): raise WandException('Cannot write image to file %s: %s'%(source, get_exception(wand))) _libwand.DestroyMagickWand(wand) _libwand.MagickWandTerminus()
strummerTFIU/TFG-IsometricMaps
refs/heads/master
LAStools/ArcGIS_toolbox/scripts_production/las2demPro.py
1
# # las2demPro.py # # (c) 2013, martin isenburg - http://rapidlasso.com # rapidlasso GmbH - fast tools to catch reality # # uses las2dem.exe to raster a folder of LiDAR files # # LiDAR input: LAS/LAZ/BIN/TXT/SHP/BIL/ASC/DTM # raster output: BIL/ASC/IMG/TIF/DTM/PNG/JPG # # for licensing see http://lastools.org/LICENSE.txt # import sys, os, arcgisscripting, subprocess def check_output(command,console): if console == True: process = subprocess.Popen(command) else: process = subprocess.Popen(command, shell=True, stdout=subprocess.PIPE, stderr=subprocess.STDOUT, universal_newlines=True) output,error = process.communicate() returncode = process.poll() return returncode,output ### create the geoprocessor object gp = arcgisscripting.create(9.3) ### report that something is happening gp.AddMessage("Starting las2dem production ...") ### get number of arguments argc = len(sys.argv) ### report arguments (for debug) #gp.AddMessage("Arguments:") #for i in range(0, argc): # gp.AddMessage("[" + str(i) + "]" + sys.argv[i]) ### get the path to LAStools lastools_path = os.path.dirname(os.path.dirname(os.path.dirname(sys.argv[0]))) ### make sure the path does not contain spaces if lastools_path.count(" ") > 0: gp.AddMessage("Error. Path to .\\lastools installation contains spaces.") gp.AddMessage("This does not work: " + lastools_path) gp.AddMessage("This would work: C:\\software\\lastools") sys.exit(1) ### complete the path to where the LAStools executables are lastools_path = lastools_path + "\\bin" ### check if path exists if os.path.exists(lastools_path) == False: gp.AddMessage("Cannot find .\\lastools\\bin at " + lastools_path) sys.exit(1) else: gp.AddMessage("Found " + lastools_path + " ...") ### create the full path to the las2dem executable las2dem_path = lastools_path+"\\las2dem.exe" ### check if executable exists if os.path.exists(lastools_path) == False: gp.AddMessage("Cannot find las2dem.exe at " + las2dem_path) sys.exit(1) else: gp.AddMessage("Found " + las2dem_path + " ...") ### create the command string for las2dem.exe command = ['"'+las2dem_path+'"'] ### maybe use '-verbose' option if sys.argv[argc-1] == "true": command.append("-v") ### counting up the arguments c = 1 ### add input LiDAR wildcards = sys.argv[c+1].split() for wildcard in wildcards: command.append("-i") command.append('"' + sys.argv[c] + "\\" + wildcard + '"') c = c + 2 ### maybe use a user-defined step size if sys.argv[c] != "1": command.append("-step") command.append(sys.argv[c].replace(",",".")) c = c + 1 ### maybe use a user-defined kill if sys.argv[c] != "100": command.append("-kill") command.append(sys.argv[c].replace(",",".")) c = c + 1 ### what should we raster if sys.argv[c] == "slope": command.append("-slope") elif sys.argv[c] == "intensity": command.append("-intensity") elif sys.argv[c] == "rgb": command.append("-rgb") c = c + 1 ### what should we output if sys.argv[c] == "hillshade": command.append("-hillshade") elif sys.argv[c] == "gray ramp": command.append("-gray") elif sys.argv[c] == "false colors": command.append("-false") ### do we have special lighting for hillshade if sys.argv[c] == "hillshade": if (sys.argv[c+1] != "north east") or (sys.argv[c+2] != "1 pm"): command.append("-light") if sys.argv[c+1] == "north": command.append("0") command.append("1.41421") elif sys.argv[c+1] == "south": command.append("0") command.append("-1.41421") elif sys.argv[c+1] == "east": command.append("1.41421") command.append("0") elif sys.argv[c+1] == "west": command.append("-1.41421") command.append("0") elif sys.argv[c+1] == "north east": command.append("1") command.append("1") elif sys.argv[c+1] == "south east": command.append("1") command.append("-1") elif sys.argv[c+1] == "north west": command.append("-1") command.append("1") else: ### if sys.argv[c+1] == "south west" command.append("-1") command.append("-1") if sys.argv[c+2] == "noon": command.append("100") elif sys.argv[c+2] == "1 pm": command.append("2") elif sys.argv[c+2] == "3 pm": command.append("1") elif sys.argv[c+2] == "6 pm": command.append("0.5") else: ### if sys.argv[c+2] == "9 pm" command.append("0.1") ### do we have a min max value for colors if (sys.argv[c] == "gray ramp") or (sys.argv[c] == "false colors"): if (sys.argv[c+3] != "#") and (sys.argv[c+4] != "#"): command.append("-set_min_max") command.append(sys.argv[c+3].replace(",",".")) command.append(sys.argv[c+4].replace(",",".")) c = c + 5 ### what should we triangulate if sys.argv[c] == "ground points only": command.append("-keep_class") command.append("2") command.append("-extra_pass") elif sys.argv[c] == "ground and keypoints": command.append("-keep_class") command.append("2") command.append("8") command.append("-extra_pass") elif sys.argv[c] == "ground and buildings": command.append("-keep_class") command.append("2") command.append("6") command.append("-extra_pass") elif sys.argv[c] == "ground and vegetation": command.append("-keep_class") command.append("2") command.append("3") command.append("4") command.append("5") command.append("-extra_pass") elif sys.argv[c] == "ground and objects": command.append("-keep_class") command.append("2") command.append("3") command.append("4") command.append("5") command.append("6") command.append("-extra_pass") elif sys.argv[c] == "last return only": command.append("-last_only") command.append("-extra_pass") elif sys.argv[c] == "first return only": command.append("-first_only") command.append("-extra_pass") c = c + 1 ### should we use the tile bounding box if sys.argv[c] == "true": command.append("-use_tile_bb") c = c + 1 ### do we have lakes if sys.argv[c] != "#": command.append("-lakes") command.append('"'+sys.argv[c]+'"') c = c + 1 ### do we have creeks if sys.argv[c] != "#": command.append("-creeks") command.append('"'+sys.argv[c]+'"') c = c + 1 ### maybe an output format was selected if sys.argv[c] != "#": command.append("-o" + sys.argv[c]) c = c + 1 ### maybe an output directory was selected if sys.argv[c] != "#": command.append("-odir") command.append('"'+sys.argv[c]+'"') c = c + 1 ### maybe an output appendix was selected if sys.argv[c] != "#": command.append("-odix") command.append('"'+sys.argv[c]+'"') c = c + 1 ### maybe we should run on multiple cores if sys.argv[c] != "1": command.append("-cores") command.append(sys.argv[c]) c = c + 1 ### maybe there are additional input options if sys.argv[c] != "#": additional_options = sys.argv[c].split() for option in additional_options: command.append(option) ### report command string gp.AddMessage("LAStools command line:") command_length = len(command) command_string = str(command[0]) command[0] = command[0].strip('"') for i in range(1, command_length): command_string = command_string + " " + str(command[i]) command[i] = command[i].strip('"') gp.AddMessage(command_string) ### run command returncode,output = check_output(command, False) ### report output of las2dem gp.AddMessage(str(output)) ### check return code if returncode != 0: gp.AddMessage("Error. las2dem failed.") sys.exit(1) ### report happy end gp.AddMessage("Success. las2dem done.")
onyb/mooca
refs/heads/master
Udacity/UD032_Data_Wrangling_with_MongoDB/Lesson_2/Problem_Set/02-Airport_List/airports.py
1
#!/usr/bin/env python # -*- coding: utf-8 -*- # All your changes should be in the 'extract_airports' function # It should return a list of airport codes, excluding any combinations like "All" from bs4 import BeautifulSoup html_page = "options.html" def extract_airports(page): data = [] with open(page, "r") as html: # do something here to find the necessary values soup = BeautifulSoup(html) airport_list = soup.find(id="AirportList") for option in airport_list.find_all("option"): value = option["value"] if "All" not in value: data.append(value) return data def test(): data = extract_airports(html_page) assert len(data) == 15 assert "ATL" in data assert "ABR" in data test()
sclorg/s2i-python-container
refs/heads/master
3.7/test/micropipenv-test-app/setup.py
125
from setuptools import setup, find_packages setup ( name = "testapp", version = "0.1", description = "Example application to be deployed.", packages = find_packages(), install_requires = ["gunicorn"], )
datsfosure/ansible
refs/heads/devel
test/units/plugins/__init__.py
7690
# (c) 2012-2014, Michael DeHaan <michael.dehaan@gmail.com> # # This file is part of Ansible # # Ansible is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # Ansible is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with Ansible. If not, see <http://www.gnu.org/licenses/>. # Make coding more python3-ish from __future__ import (absolute_import, division, print_function) __metaclass__ = type
ParadigmaXis/dadosabertos_cmporto_pt
refs/heads/master
ckanext/dados_cmporto_pt/controller.py
1
# -*- coding: utf-8 -*- import ckan.lib.base as base from ckan.controllers import admin import json from ckan.model import Package class AdminController(admin.AdminController): def _get_config_form_items(self): items = super(AdminController, self)._get_config_form_items() items[7]['options'].append({'value': '4', 'text': 'dados.cm-porto.pt'}) return items class StaticPagesController(base.BaseController): def linked_data(self): return base.render('home/cmp/linked-data.html') def terms_of_use(self): return base.render('home/cmp/terms-of-use.html') def privacy_policy(self): return base.render('home/cmp/privacy-policy.html') def moderation_policy(self): return base.render('home/cmp/moderation-policy.html') def list_license(self): return base.render('home/cmp/license-list.html', extra_vars = {'licenses': Package.get_license_register().items()}) def data_cubes(self): return base.render('home/cmp/data-cubes.html')
ptemplier/ansible
refs/heads/devel
lib/ansible/plugins/action/asa_config.py
45
# # (c) 2017, Red Hat, Inc. # # This file is part of Ansible # # Ansible is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # Ansible is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with Ansible. If not, see <http://www.gnu.org/licenses/>. # from __future__ import (absolute_import, division, print_function) __metaclass__ = type import os import re import time import glob from ansible.plugins.action.asa import ActionModule as _ActionModule from ansible.module_utils._text import to_text from ansible.module_utils.six.moves.urllib.parse import urlsplit from ansible.utils.vars import merge_hash PRIVATE_KEYS_RE = re.compile('__.+__') class ActionModule(_ActionModule): def run(self, tmp=None, task_vars=None): if self._task.args.get('src'): try: self._handle_template() except ValueError as exc: return dict(failed=True, msg=exc.message) result = super(ActionModule, self).run(tmp, task_vars) if self._task.args.get('backup') and result.get('__backup__'): # User requested backup and no error occurred in module. # NOTE: If there is a parameter error, _backup key may not be in results. filepath = self._write_backup(task_vars['inventory_hostname'], result['__backup__']) result['backup_path'] = filepath # strip out any keys that have two leading and two trailing # underscore characters for key in result.keys(): if PRIVATE_KEYS_RE.match(key): del result[key] return result def _get_working_path(self): cwd = self._loader.get_basedir() if self._task._role is not None: cwd = self._task._role._role_path return cwd def _write_backup(self, host, contents): backup_path = self._get_working_path() + '/backup' if not os.path.exists(backup_path): os.mkdir(backup_path) for fn in glob.glob('%s/%s*' % (backup_path, host)): os.remove(fn) tstamp = time.strftime("%Y-%m-%d@%H:%M:%S", time.localtime(time.time())) filename = '%s/%s_config.%s' % (backup_path, host, tstamp) open(filename, 'w').write(contents) return filename def _handle_template(self): src = self._task.args.get('src') working_path = self._get_working_path() if os.path.isabs(src) or urlsplit('src').scheme: source = src else: source = self._loader.path_dwim_relative(working_path, 'templates', src) if not source: source = self._loader.path_dwim_relative(working_path, src) if not os.path.exists(source): raise ValueError('path specified in src not found') try: with open(source, 'r') as f: template_data = to_text(f.read()) except IOError: return dict(failed=True, msg='unable to load src file') # Create a template search path in the following order: # [working_path, self_role_path, dependent_role_paths, dirname(source)] searchpath = [working_path] if self._task._role is not None: searchpath.append(self._task._role._role_path) if hasattr(self._task, "_block:"): dep_chain = self._task._block.get_dep_chain() if dep_chain is not None: for role in dep_chain: searchpath.append(role._role_path) searchpath.append(os.path.dirname(source)) self._templar.environment.loader.searchpath = searchpath self._task.args['src'] = self._templar.template(template_data)
Stackvana/stack
refs/heads/master
bin/binaries/lib/python/microcule/__init__.py
1
import sys import json import logging import traceback import pkg_resources import wsgiref.handlers import os # open incoming connection from fd3 if sys.version_info[0] < 3: fd3 = os.fdopen(3, 'w+') else: fd3 = os.fdopen(3, 'wb+', buffering=0) class FullMicroculeJSONFormatter(logging.Formatter): def format(self, record): record.message = record.getMessage() if record.exc_info and not record.exc_text: record.exc_text = self.formatException(record.exc_info) record = record.__dict__.copy() record['exc_info'] = None try: json.dumps(record['args']) except Exception: del record['args'] record['msg'] = record['message'] res = {'type': 'log', 'payload': {'entry': record}} return json.dumps(res) class SimpleMicroculeJSONFormatter(logging.Formatter): def format(self, record): msg = logging.Formatter.format(self, record) res = {'type': 'log', 'payload': {'entry': msg}} return json.dumps(res) class MicroculeExceptHook: def __init__(self, display=1, verbose=1): self.display = display self.verbose = verbose def __call__(self, etype, evalue, etb): self.handle((etype, evalue, etb)) def handle(self, info=None): self.send_exception(info) sys.exit(1) def send_exception(self, info=None): info = info or sys.exc_info() code = info[0].__name__ if getattr(info[0], '__module__ ', None): code = info[0].__module__ + '.' + code payload = {'code': code} if hasattr(info[1], 'args'): payload['args'] = repr(info[1].args) if self.verbose: payload['error'] = ''.join(traceback.format_exception(*info)) else: payload['error'] = str(info[1]) res = {'type': 'error', 'payload': payload} if isinstance(info[1], ImportError) and info[1].message.startswith('No module named '): payload['code'] = 'MODULE_NOT_FOUND' payload['module'] = info[1].message.replace('No module named ', '') if isinstance(info[1], (pkg_resources.VersionConflict, pkg_resources.DistributionNotFound)): req = None try: if hasattr(info[1], 'req'): req = info[1].req # This is check for compatibility with old version of setuptools if req is None: for arg in info[1].args: if isinstance(arg, pkg_resources.Requirement): req = arg except BaseException: # unable to parse exception to requirement - it's ok pass if req is not None: payload['code'] = 'MODULE_NOT_FOUND' payload['module'] = str(req) error = '%s(%s): %s' % (payload['code'], code, str(info[1])) if self.verbose: payload['error'] += error else: payload['error'] = error fd3.write(json.dumps(res)+'\n') fd3.flush() fd3.write(json.dumps({'type': 'statusCode', 'payload': {'value': 500}})+'\n') fd3.flush() if self.display: sys.stdout.write(payload['error'].rstrip('\n')+'\n') sys.stdout.flush() fd3.write(json.dumps({'type': 'end'})+'\n') fd3.flush() class wsgi(wsgiref.handlers.CGIHandler): def __init__(self, Hook=None): self.Hook = Hook or getattr(sys.modules.get('__main__',sys), 'Hook', None) wsgiref.handlers.BaseCGIHandler.__init__( self, sys.stdin, sys.stdout, sys.stderr, {}, multithread=False, multiprocess=True ) def send_headers(self): self.cleanup_headers() self.headers_sent = True # remark: the status code needs to be sent to the parent process as an 3 digit integer value, not a string value with label # todo: make parse int code for status more robust. head = {'type': 'writeHead', 'payload': {'code': int(self.status[:3]), 'headers': dict(self.headers)}} fd3.write(json.dumps(head)+'\n') fd3.flush() def add_cgi_vars(self): #assert not Hook['isStreaming'], 'Streaming hooks not yet supported by WSGI gateway' self.environ.update(self.Hook['env']) if 'hookAccessKey' in self.Hook: self.environ['hook_private_key'] = self.Hook['hookAccessKey'] self.environ['SERVER_NAME'] = self.Hook['req']['host'] self.environ['GATEWAY_INTERFACE'] = 'CGI/1.1' self.environ['SERVER_PORT'] = '443' self.environ['REMOTE_ADDR'] = self.environ['REMOTE_HOST'] = self.Hook['req']['connection']['remoteAddress'] self.environ['CONTENT_LENGTH'] = '' self.environ['SCRIPT_NAME'] = '' self.environ['SERVER_PROTOCOL'] = 'HTTP/1.0' self.environ['REQUEST_METHOD'] = self.Hook['req']['method'] self.environ['PATH_INFO'] = self.Hook['req']['path'] self.environ['QUERY_STRING'] = self.Hook['req']['url'][len(self.Hook['req']['path'])+1:] self.environ['CONTENT_TYPE'] = self.Hook['req']['headers'].get('content-type', '') if 'content-length' in self.Hook['req']['headers']: self.environ['CONTENT_LENGTH'] = self.Hook['req']['headers']['content-length'] for k,v in self.Hook['req']['headers'].items(): k = k.replace('-', '_').upper() v = v.strip() if k in self.environ: continue self.environ['HTTP_'+k] = v if self.Hook.get('isHookio', False): hookname = '/%(owner)s/%(hook)s/' % self.Hook['req']['params'] assert (self.Hook['req']['path']+'/').startswith(hookname) wsgiref.util.shift_path_info(self.environ) # user wsgiref.util.shift_path_info(self.environ) # hook def hookioLoggingConfig(level=None, format=None, datefmt=None): logging._acquireLock() try: if level is not None: logging.root.setLevel(level) if len(logging.root.handlers) > 0: return hdlr = logging.StreamHandler(sys.stderr) if format is None: fmt = FullMicroculeJSONFormatter() else: fmt = SimpleMicroculeJSONFormatter(format, datefmt) hdlr.setFormatter(fmt) logging.root.addHandler(hdlr) finally: logging._releaseLock() def parse_argv(argv=None): argv = argv or sys.argv # print(argv[0]) # binary # print(argv[1]) # -c # print(argv[2]) # the code code = argv[2] # print(argv[3]) # -e # print(argv[4]) # the env envJSON = argv[4] Hook = json.loads(envJSON) #print(Hook.keys()) #print(Hook['resource'].keys()) Hook['req'] = Hook['input'] # TODO: full mapping of http including streaming / multipart streaming # print(argv[5]) # -s # print(argv[6]) # the service # TODO: do something with service variable service = argv[6] debug_output = 'gateway' in Hook['resource'].get('name', 'gateway') prod_mode = Hook['resource'].get('mode', 'Debug') == 'Production' #if debug_output: # sys.stderr = sys.stdout level = [logging.DEBUG, logging.INFO][prod_mode] hookioLoggingConfig(level) sys.excepthook = MicroculeExceptHook(debug_output, not prod_mode) return code, service, Hook
sustainableis/python-sis
refs/heads/master
pysis/services/organizations/__init__.py
1
# -*- encoding: utf-8 -*- from pysis.services.base import Service class Organizations(Service): """Organizations Service Consumes Organizations API: <{url}/organizations> """ def __init__(self, client): """Creates Organizations object with a client""" super(Organizations, self).__init__(client) def get(self, id=None): """Gets Organizations from the API Args: id (int): id of the organization. if None, returns all Organizations Returns: Organizations resources """ if id is None: request = self.request_builder('organizations.get') else: assert isinstance(id, int) request = self.request_builder('organizations.get', id=id) return self._get(request) def create(self, data): """Create an Organization Args: data (dict): {'name': 'Sample Organization'} Returns: Response """ #TODO: Make sure name is not '' if the node server doesn't request = self.request_builder('organizations.create', body=data) return self._post(request) def delete(self, id): """Delete an Organization Args: id (int): id of the organization. Returns: Response """ assert isinstance(id, int) request = self.request_builder('organizations.delete', id=id) return self._delete(request) def update(self, id, data): """Create an Organization Args: id (int): id of the organization. data (dict): {'name': 'Sample Organization'} Returns: Resource """ assert isinstance(id, int) request = self.request_builder('organizations.update', id=id, body=data) return self._put(request) def getFacilities(self, id): """Get the facilities of an organization Args: id (int): id of the organization. Returns: Facilities resources """ assert isinstance(id, int) request = self.request_builder('organizations.getFacilities', id=id) return self._get(request) def getBuildings(self, id): """Get the buildings of an organization Args: id (int): id of the organization. Returns: Buildings resources """ assert isinstance(id, int) request = self.request_builder('organizations.getBuildings', id=id) return self._get(request) def getUsers(self, id): """Get the users of an organization Args: id (int): id of the organization. Returns: Users resources """ assert isinstance(id, int) request = self.request_builder('organizations.getUsers', id=id) return self._get(request) def getFeeds(self, id): """Get the feeds of an organization Args: id (int): id of the organization. Returns: Feeds resources """ assert isinstance(id, int) request = self.request_builder('organizations.getFeeds', id=id) return self._get(request) def getOutputs(self, id): """Get the outputs of an organization Args: id (int): id of the organization. Returns: Outputs resources """ assert isinstance(id, int) request = self.request_builder('organizations.getOutputs', id=id) return self._get(request) def getBlastcells(self, id): """Get the blastcells of an organization Args: id (int): id of the organization. Returns: Blastcells resources """ assert isinstance(id, int) request = self.request_builder('organizations.getBlastcells', id=id) return self._get(request)
adjspecies/furrypoll-munger
refs/heads/master
bin/other/weasyl-dates.py
2
data = ( ( 2013,1,6), ( 2013,1,11), ( 2013,1,11), ( 2013,1,11), ( 2013,1,12), ( 2013,1,12), ( 2013,1,12), ( 2013,1,12), ( 2013,1,12), ( 2013,1,12), ( 2013,1,12), ( 2013,1,12), ( 2013,1,12), ( 2013,1,12), ( 2013,1,12), ( 2013,1,12), ( 2013,1,13), ( 2013,1,14), ( 2013,1,14), ( 2013,1,14), ( 2013,1,14), ( 2013,1,14), ( 2013,1,14), ( 2013,1,15), ( 2013,1,18), ( 2013,1,18), ( 2013,1,18), ( 2013,1,19), ( 2013,1,20), ( 2013,1,20), ( 2013,1,20), ( 2013,1,21), ( 2013,1,22), ( 2013,1,22), ( 2013,1,22), ( 2013,1,22), ( 2013,1,22), ( 2013,1,23), ( 2013,1,23), ( 2013,1,23), ( 2013,1,23), ( 2013,1,23), ( 2013,1,24), ( 2013,1,24), ( 2013,1,24), ( 2013,1,25), ( 2013,1,25), ( 2013,1,25), ( 2013,1,26), ( 2013,1,26), ( 2013,1,26), ( 2013,1,27), ( 2013,1,27), ( 2013,1,27), ( 2013,1,27), ( 2013,1,27), ( 2013,1,29), ( 2013,1,30), ( 2013,1,30), ( 2013,1,30), ( 2013,2,1), ( 2013,2,1), ( 2013,2,1), ( 2013,2,2), ( 2013,2,2), ( 2013,2,2), ( 2013,2,2), ( 2013,2,2), ( 2013,2,2), ( 2013,2,2), ( 2013,2,3), ( 2013,2,3), ( 2013,2,4), ( 2013,2,4), ( 2013,2,4), ( 2013,2,4), ( 2013,2,4), ( 2013,2,4), ( 2013,2,4), ( 2013,2,4), ( 2013,2,4), ( 2013,2,5), ( 2013,2,5), ( 2013,2,5), ( 2013,2,7), ( 2013,2,8), ( 2013,2,8), ( 2013,2,8), ( 2013,2,8), ( 2013,2,9), ( 2013,2,9), ( 2013,2,9), ( 2013,2,10), ( 2013,2,13), ( 2013,2,13), ( 2013,2,14), ( 2013,2,14), ( 2013,2,14), ( 2013,2,15), ( 2013,2,16), ( 2013,2,16), ( 2013,2,17), ( 2013,2,17), ( 2013,2,18), ( 2013,2,18), ( 2013,2,18), ( 2013,2,18), ( 2013,2,18), ( 2013,2,18), ( 2013,2,18), ( 2013,2,18), ( 2013,2,18), ( 2013,2,18), ( 2013,2,19), ( 2013,2,19), ( 2013,2,19), ( 2013,2,19), ( 2013,2,19), ( 2013,2,19), ( 2013,2,19), ( 2013,2,19), ( 2013,2,19), ( 2013,2,19), ( 2013,2,19), ( 2013,2,20), ( 2013,2,20), ( 2013,2,20), ( 2013,2,20), ( 2013,2,21), ( 2013,2,21), ( 2013,2,22), ( 2013,2,24), ( 2013,2,24), ( 2013,2,25), ( 2013,2,25), ( 2013,2,26), ( 2013,2,26), ( 2013,2,26), ( 2013,2,27), ( 2013,2,27), ( 2013,2,27), ( 2013,2,28), ( 2013,3,1), ( 2013,3,3), ( 2013,3,3), ( 2013,3,3), ( 2013,3,3), ( 2013,3,4), ( 2013,3,4), ( 2013,3,4), ( 2013,3,5), ( 2013,3,5), ( 2013,3,7), ( 2013,3,12), ( 2013,3,13), ( 2013,3,13), ( 2013,3,15), ( 2013,3,16), ( 2013,3,16), ( 2013,3,18), ( 2013,3,18), ( 2013,3,19), ( 2013,3,19), ( 2013,3,19), ( 2013,3,20), ( 2013,3,21), ( 2013,3,21), ( 2013,3,21), ( 2013,3,26), ( 2013,3,26), ( 2013,3,30), ( 2013,3,31), ( 2013,3,31), ( 2013,4,3), ( 2013,4,3), ( 2013,4,4), ( 2013,4,4), ( 2013,4,6), ( 2013,4,6), ( 2013,4,6), ( 2013,4,9), ( 2013,4,9), ( 2013,4,10), ( 2013,4,12), ( 2013,4,13), ( 2013,4,18), ( 2013,4,20), ( 2013,4,21), ( 2013,4,21), ( 2013,4,25), ( 2013,4,25), ( 2013,4,26), ( 2013,4,27), ( 2013,4,30), ( 2013,4,30), ( 2013,5,1), ( 2013,5,2), ( 2013,5,2), ( 2013,5,2), ( 2013,5,2), ( 2013,5,2), ( 2013,5,2), ( 2013,5,3), ( 2013,5,4), ( 2013,5,4), ( 2013,5,5), ( 2013,5,8), ( 2013,5,8), ( 2013,5,9), ( 2013,5,9), ( 2013,5,12), ( 2013,5,12), ( 2013,5,15), ( 2013,5,15), ( 2013,5,16), ( 2013,5,16), ( 2013,5,20), ( 2013,5,21), ( 2013,5,22), ( 2013,5,25), ( 2013,5,26), ( 2013,5,28), ( 2013,5,28), ( 2013,5,29), ( 2013,5,30), ( 2013,6,3), ( 2013,6,5), ( 2013,6,5), ( 2013,6,5), ( 2013,6,7), ( 2013,6,7), ( 2013,6,7), ( 2013,6,8), ( 2013,6,9), ( 2013,6,9), ( 2013,6,11), ( 2013,6,11), ( 2013,6,13), ( 2013,6,15), ( 2013,6,18), ( 2013,6,20), ( 2013,6,20), ( 2013,6,22), ( 2013,6,23), ( 2013,6,23), ( 2013,6,29), ( 2013,7,2), ( 2013,7,4), ( 2013,7,6), ( 2013,7,8), ( 2013,7,11), ( 2013,7,11), ( 2013,7,12), ( 2013,7,12), ( 2013,7,12), ( 2013,7,12), ( 2013,7,12), ( 2013,7,12), ( 2013,7,13), ( 2013,7,14), ( 2013,7,14), ( 2013,7,16), ( 2013,7,17), ( 2013,7,18), ( 2013,7,18), ( 2013,7,22), ( 2013,7,22), ( 2013,7,22), ( 2013,7,22), ( 2013,7,24), ( 2013,7,24), ( 2013,7,24), ( 2013,7,25), ( 2013,7,25), ( 2013,7,25), ( 2013,7,25), ( 2013,7,25), ( 2013,7,25), ( 2013,7,25), ( 2013,7,25), ( 2013,7,25), ( 2013,7,25), ( 2013,7,25), ( 2013,7,25), ( 2013,7,25), ( 2013,7,25), ( 2013,7,25), ( 2013,7,25), ( 2013,7,25), ( 2013,7,25), ( 2013,7,25), ( 2013,7,25), ( 2013,7,25), ( 2013,7,25), ( 2013,7,25), ( 2013,7,25), ( 2013,7,25), ( 2013,7,25), ( 2013,7,25), ( 2013,7,26), ( 2013,7,26), ( 2013,7,26), ( 2013,7,26), ( 2013,7,26), ( 2013,7,26), ( 2013,7,26), ( 2013,7,26), ( 2013,7,26), ( 2013,7,26), ( 2013,7,26), ( 2013,7,27), ( 2013,7,27), ( 2013,7,27), ( 2013,7,27), ( 2013,7,27), ( 2013,7,27), ( 2013,7,28), ( 2013,7,28), ( 2013,7,29), ( 2013,7,30), ( 2013,7,31), ( 2013,7,31), ( 2013,7,31), ( 2013,7,31), ( 2013,7,31), ( 2013,7,31), ( 2013,7,31), ( 2013,7,31), ( 2013,7,31), ( 2013,7,31), ( 2013,7,31), ( 2013,7,31), ( 2013,7,31), ( 2013,7,31), ( 2013,7,31), ( 2013,7,31), ( 2013,7,31), ( 2013,7,31), ( 2013,7,31), ( 2013,7,31), ( 2013,7,31), ( 2013,7,31), ( 2013,8,1), ( 2013,8,1), ( 2013,8,1), ( 2013,8,1), ( 2013,8,1), ( 2013,8,1), ( 2013,8,1), ( 2013,8,1), ( 2013,8,1), ( 2013,8,1), ( 2013,8,1), ( 2013,8,1), ( 2013,8,1), ( 2013,8,2), ( 2013,8,2), ( 2013,8,2), ( 2013,8,2), ( 2013,8,2), ( 2013,8,4), ( 2013,8,4), ( 2013,8,5), ( 2013,8,5), ( 2013,8,5), ( 2013,8,5), ( 2013,8,5), ( 2013,8,5), ( 2013,8,5), ( 2013,8,6), ( 2013,8,6), ( 2013,8,7), ( 2013,8,8), ( 2013,8,8), ( 2013,8,8), ( 2013,8,8), ( 2013,8,8), ( 2013,8,8), ( 2013,8,8), ( 2013,8,8), ( 2013,8,8), ( 2013,8,9), ( 2013,8,9), ( 2013,8,9), ( 2013,8,9), ( 2013,8,11), ( 2013,8,12), ( 2013,8,13), ( 2013,8,13), ( 2013,8,13), ( 2013,8,13), ( 2013,8,14), ( 2013,8,14), ( 2013,8,15), ( 2013,8,15), ( 2013,8,15), ( 2013,8,15), ( 2013,8,15), ( 2013,8,15), ( 2013,8,15), ( 2013,8,16), ( 2013,8,16), ( 2013,8,17), ( 2013,8,18), ( 2013,8,19), ( 2013,8,21), ( 2013,8,22), ( 2013,8,22), ( 2013,8,22), ( 2013,8,22), ( 2013,8,23), ( 2013,8,23), ( 2013,8,24), ( 2013,8,24), ( 2013,8,24), ( 2013,8,27), ( 2013,8,29), ( 2013,8,30), ( 2013,8,30), ( 2013,8,31), ( 2013,9,2), ( 2013,9,2), ( 2013,9,3), ( 2013,9,3), ( 2013,9,6), ( 2013,9,6), ( 2013,9,7), ( 2013,9,7), ( 2013,9,8), ( 2013,9,9), ( 2013,9,10), ( 2013,9,12), ( 2013,9,14), ( 2013,9,15), ( 2013,9,15), ( 2013,9,15), ( 2013,9,15), ( 2013,9,15), ( 2013,9,16), ( 2013,9,17), ( 2013,9,19), ( 2013,9,19), ( 2013,9,21), ( 2013,9,22), ( 2013,9,23), ( 2013,9,24), ( 2013,9,24), ( 2013,9,30), ( 2013,9,30), ( 2013,9,30), ( 2013,9,30), ( 2013,9,30), ( 2013,9,30), ( 2013,9,30), ( 2013,9,30), ( 2013,9,30), ( 2013,9,30), ( 2013,9,30), ( 2013,9,30), ( 2013,9,30), ( 2013,9,30), ( 2013,9,30), ( 2013,9,30), ( 2013,9,30), ( 2013,10,1), ( 2013,10,1), ( 2013,10,2), ( 2013,10,3), ( 2013,10,3), ( 2013,10,4), ( 2013,10,4), ( 2013,10,4), ( 2013,10,4), ( 2013,10,5), ( 2013,10,6), ( 2013,10,7), ( 2013,10,7), ( 2013,10,7), ( 2013,10,7), ( 2013,10,7), ( 2013,10,7), ( 2013,10,7), ( 2013,10,7), ( 2013,10,7), ( 2013,10,7), ( 2013,10,7), ( 2013,10,7), ( 2013,10,7), ( 2013,10,7), ( 2013,10,7), ( 2013,10,7), ( 2013,10,7), ( 2013,10,8), ( 2013,10,8), ( 2013,10,8), ( 2013,10,8), ( 2013,10,8), ( 2013,10,8), ( 2013,10,8), ( 2013,10,8), ( 2013,10,8), ( 2013,10,8), ( 2013,10,8), ( 2013,10,8), ( 2013,10,8), ( 2013,10,8), ( 2013,10,8), ( 2013,10,8), ( 2013,10,8), ( 2013,10,8), ( 2013,10,9), ( 2013,10,9), ( 2013,10,9), ( 2013,10,9), ( 2013,10,9), ( 2013,10,9), ( 2013,10,9), ( 2013,10,9), ( 2013,10,9), ( 2013,10,9), ( 2013,10,9), ( 2013,10,10), ( 2013,10,10), ( 2013,10,10), ( 2013,10,10), ( 2013,10,10), ( 2013,10,10), ( 2013,10,10), ( 2013,10,10), ( 2013,10,10), ( 2013,10,10), ( 2013,10,10), ( 2013,10,10), ( 2013,10,10), ( 2013,10,10), ( 2013,10,10), ( 2013,10,10), ( 2013,10,11), ( 2013,10,11), ( 2013,10,11), ( 2013,10,11), ( 2013,10,11), ( 2013,10,11), ( 2013,10,11), ( 2013,10,11), ( 2013,10,11), ( 2013,10,11), ( 2013,10,11), ( 2013,10,12), ( 2013,10,12), ( 2013,10,12), ( 2013,10,12), ( 2013,10,12), ( 2013,10,12), ( 2013,10,12), ( 2013,10,12), ( 2013,10,12), ( 2013,10,12), ( 2013,10,12), ( 2013,10,12), ( 2013,10,12), ( 2013,10,12), ( 2013,10,12), ( 2013,10,12), ( 2013,10,12), ( 2013,10,12), ( 2013,10,12), ( 2013,10,13), ( 2013,10,13), ( 2013,10,13), ( 2013,10,13), ( 2013,10,13), ( 2013,10,13), ( 2013,10,13), ( 2013,10,13), ( 2013,10,13), ( 2013,10,13), ( 2013,10,13), ( 2013,10,13), ( 2013,10,13), ( 2013,10,13), ( 2013,10,13), ( 2013,10,13), ( 2013,10,14), ( 2013,10,14), ( 2013,10,14), ( 2013,10,14), ( 2013,10,14), ( 2013,10,14), ( 2013,10,14), ( 2013,10,14), ( 2013,10,14), ( 2013,10,14), ( 2013,10,14), ( 2013,10,14), ( 2013,10,15), ( 2013,10,15), ( 2013,10,15), ( 2013,10,15), ( 2013,10,15), ( 2013,10,15), ( 2013,10,15), ( 2013,10,15), ( 2013,10,15), ( 2013,10,15), ( 2013,10,15), ( 2013,10,15), ( 2013,10,15), ( 2013,10,15), ( 2013,10,15), ( 2013,10,16), ( 2013,10,16), ( 2013,10,16), ( 2013,10,16), ( 2013,10,16), ( 2013,10,16), ( 2013,10,16), ( 2013,10,16), ( 2013,10,16), ( 2013,10,16), ( 2013,10,16), ( 2013,10,17), ( 2013,10,17), ( 2013,10,17), ( 2013,10,17), ( 2013,10,17), ( 2013,10,17), ( 2013,10,17), ( 2013,10,17), ( 2013,10,17), ( 2013,10,17), ( 2013,10,17), ( 2013,10,17), ( 2013,10,17), ( 2013,10,18), ( 2013,10,18), ( 2013,10,18), ( 2013,10,18), ( 2013,10,18), ( 2013,10,18), ( 2013,10,18), ( 2013,10,19), ( 2013,10,19), ( 2013,10,19), ( 2013,10,19), ( 2013,10,20), ( 2013,10,20), ( 2013,10,20), ( 2013,10,20), ( 2013,10,20), ( 2013,10,20), ( 2013,10,20), ( 2013,10,21), ( 2013,10,21), ( 2013,10,21), ( 2013,10,21), ( 2013,10,21), ( 2013,10,21), ( 2013,10,21), ( 2013,10,21), ( 2013,10,21), ( 2013,10,21), ( 2013,10,21), ( 2013,10,22), ( 2013,10,22), ( 2013,10,22), ( 2013,10,22), ( 2013,10,22), ( 2013,10,22), ( 2013,10,22), ( 2013,10,23), ( 2013,10,23), ( 2013,10,23), ( 2013,10,23), ( 2013,10,23), ( 2013,10,23), ( 2013,10,23), ( 2013,10,23), ( 2013,10,23), ( 2013,10,23), ( 2013,10,23), ( 2013,10,23), ( 2013,10,24), ( 2013,10,24), ( 2013,10,24), ( 2013,10,24), ( 2013,10,24), ( 2013,10,24), ( 2013,10,24), ( 2013,10,24), ( 2013,10,25), ( 2013,10,25), ( 2013,10,25), ( 2013,10,25), ( 2013,10,25), ( 2013,10,26), ( 2013,10,26), ( 2013,10,26), ( 2013,10,26), ( 2013,10,26), ( 2013,10,26), ( 2013,10,27), ( 2013,10,27), ( 2013,10,27), ( 2013,10,27), ( 2013,10,27), ( 2013,10,27), ( 2013,10,27), ( 2013,10,28), ( 2013,10,28), ( 2013,10,28), ( 2013,10,28), ( 2013,10,28), ( 2013,10,28), ( 2013,10,28), ( 2013,10,28), ( 2013,10,28), ( 2013,10,28), ( 2013,10,28), ( 2013,10,28), ( 2013,10,29), ( 2013,10,29), ( 2013,10,29), ( 2013,10,29), ( 2013,10,29), ( 2013,10,29), ( 2013,10,29), ( 2013,10,29), ( 2013,10,29), ( 2013,10,29), ( 2013,10,29), ( 2013,10,29), ( 2013,10,29), ( 2013,10,29), ( 2013,10,29), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,30), ( 2013,10,31), ( 2013,10,31), ( 2013,10,31), ( 2013,10,31), ( 2013,10,31), ( 2013,10,31), ( 2013,11,1), ( 2013,11,1), ( 2013,11,1), ( 2013,11,1), ( 2013,11,1), ( 2013,11,1), ( 2013,11,1), ( 2013,11,2), ( 2013,11,2), ( 2013,11,2), ( 2013,11,2), ( 2013,11,2), ( 2013,11,3), ( 2013,11,3), ( 2013,11,3), ( 2013,11,3), ( 2013,11,3), ( 2013,11,3), ( 2013,11,3), ( 2013,11,3), ( 2013,11,3), ( 2013,11,4), ( 2013,11,4), ( 2013,11,4), ( 2013,11,4), ( 2013,11,4), ( 2013,11,4), ( 2013,11,4), ( 2013,11,5), ( 2013,11,5), ( 2013,11,5), ( 2013,11,5), ( 2013,11,5), ( 2013,11,5), ( 2013,11,5), ( 2013,11,5), ( 2013,11,5), ( 2013,11,5), ( 2013,11,6), ( 2013,11,6), ( 2013,11,6), ( 2013,11,6), ( 2013,11,6), ( 2013,11,6), ( 2013,11,7), ( 2013,11,7), ( 2013,11,7), ( 2013,11,7), ( 2013,11,7), ( 2013,11,7), ( 2013,11,7), ( 2013,11,7), ( 2013,11,7), ( 2013,11,7), ( 2013,11,7), ( 2013,11,8), ( 2013,11,8), ( 2013,11,8), ( 2013,11,8), ( 2013,11,8), ( 2013,11,8), ( 2013,11,8), ( 2013,11,8), ( 2013,11,9), ( 2013,11,9), ( 2013,11,9), ( 2013,11,9), ( 2013,11,9), ( 2013,11,9), ( 2013,11,9), ( 2013,11,9), ( 2013,11,9), ( 2013,11,10), ( 2013,11,10), ( 2013,11,10), ( 2013,11,12), ( 2013,11,16), ( 2013,11,16), ( 2013,11,16), ( 2013,11,17), ( 2013,11,18), ( 2013,11,20), ( 2013,11,22), ( 2013,11,23), ( 2013,11,23), ( 2013,11,23), ( 2013,11,24), ( 2013,11,25), ( 2013,11,26), ( 2013,11,27), ( 2013,12,2), ( 2013,12,5), ( 2013,12,7), ( 2013,12,9), ( 2013,12,10), ( 2013,12,10), ( 2013,12,11), ( 2013,12,11), ( 2013,12,11), ( 2013,12,13), ( 2013,12,14), ( 2013,12,15), ( 2013,12,16), ( 2013,12,16), ( 2013,12,17), ( 2013,12,17), ( 2013,12,17), ( 2013,12,17), ( 2013,12,17), ( 2013,12,17), ( 2013,12,18), ( 2013,12,21), ( 2013,12,21), ( 2013,12,25), ( 2013,12,26), ( 2013,12,26), ( 2013,12,26), ( 2013,12,29), ( 2013,12,29), ( 2013,12,29), ( 2013,12,30), ( 2013,12,30), ( 2013,12,31), ( 2013,12,31), ( 2013,12,31), ( 2013,12,31), ( 2013,12,31), ) results = [[]] for entry in data: if len(results) < entry[1]: results.append([]) if len(results[entry[1] - 1]) < entry[2]: for _ in range(len(results[entry[1] - 1]), entry[2]): results[entry[1] - 1].append(0) results[entry[1] - 1][entry[2] - 1] += 1 print results print [sum(i) for i in results]
google/makani
refs/heads/master
gs/monitor2/apps/plugins/layouts/estimator_layout.py
1
# Copyright 2020 Makani Technologies LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Layout to monitor motors.""" from makani.avionics.common import gps_receiver from makani.gs.monitor2.apps.layout import base from makani.gs.monitor2.apps.plugins import common from makani.gs.monitor2.apps.plugins.indicators import control from makani.gs.monitor2.apps.plugins.indicators import estimator from makani.gs.monitor2.apps.plugins.indicators import gps from makani.gs.monitor2.apps.plugins.indicators import ground_station from makani.gs.monitor2.apps.plugins.layouts import gps_util class EstimatorLayout(base.BaseLayout): """Layout to monitor estimators.""" _NAME = 'Estimator' _DESIRED_VIEW_COLS = 12 _MODE = common.FULL_COMMS_MODE _ORDER_HORIZONTALLY = False def Initialize(self): self._AddIndicators('Network', [ control.FlightModeIndicator(self._MODE), control.ControlTimeIndicator(self._MODE), # {col_0, INDICATOR_TYPE, (UpdateFuncType)&UpdateThrottle}, control.HoverAnglesChart(self._MODE), ], {'cols': 3}) self._AddBreak() self._AddIndicators('Imus', [ estimator.EstimatorGyroBiasChart(self._MODE), estimator.EstimatorGyroBiasIndicator(self._MODE), # {col_0, INDICATOR_TYPE, (UpdateFuncType)&UpdateEstimatorAccBDiff}, estimator.EstimatorGyroDiffIndicator(self._MODE), estimator.EstimatorMagnetometerDiffIndicator(self._MODE), estimator.EstimatorAttitudeDiffChart(self._MODE), control.FdDisabledIndicator(), control.FdAllActiveIndicator(), # {col_1, INDICATOR_CHART_TYPE, # (UpdateFuncType)&UpdateEstimatorTetherForce}, control.WingPosChart(self._MODE), # {col_1, INDICATOR_TYPE, # (UpdateFuncType)&UpdateEstimatorCurrentGpsReceiver}, ], {'cols': 3}) self._AddBreak() self._AddIndicators('Gs GPS (NovAtel)', [ gps.NovAtelSigmasIndicator('GpsBaseStation'), gps.GsNovAtelPosVelTypeIndicator('GpsBaseStation'), gps.NovAtelCn0Indicator('GpsBaseStation'), ], properties={'cols': 3}) wing_gps_indicators = [] for gps_type, fc_name in gps_util.GpsSelector(): if gps_type == gps_receiver.GpsReceiverType.NOV_ATEL.value: wing_gps_indicators += [ gps.NovAtelSigmasIndicator(fc_name, name='Sigmas (%s)' % fc_name), gps.WingNovAtelPosVelTypeIndicator( fc_name, name='Pos./Vel. Type (%s)' % fc_name), gps.NovAtelCn0Indicator(fc_name, name='Cn0 (%s)' % fc_name), ] elif gps_type == gps_receiver.GpsReceiverType.SEPTENTRIO.value: wing_gps_indicators += [ gps.SeptentrioSigmasIndicator(fc_name, name='Sigmas (%s)' % fc_name), gps.SeptentrioModeIndicator( fc_name, name='Mode (%s)' % fc_name), gps.SeptentrioCn0Indicator(fc_name, name='Cn0 (%s)' % fc_name), ] else: raise ValueError('Invalid GPS type: %d.' % gps_type) self._AddIndicators('Wing GPS', wing_gps_indicators + [ # {col_1, INDICATOR_TYPE, (UpdateFuncType)&UpdateEstimatorGpsDiff}, # {col_1, INDICATOR_TYPE, (UpdateFuncType)&UpdateEstimatorGsPosEcef}, # {col_1, INDICATOR_CHART_TYPE, (UpdateFuncType)&UpdateGsGpsToWingDist}, ], properties={'cols': 3}) self._AddIndicators('Platform', [ # {col_2, INDICATOR_TYPE, (UpdateFuncType)&UpdateEstimatorGsgBias}, # {col_2, XY_PLOT_TYPE, (UpdateFuncType)&UpdatePerchPosition}, ground_station.PerchAzimuthIndicator(), ], {'cols': 3})
ZombieNinjaPirate/pypkg
refs/heads/master
HonSSH/DailyLogs.py
1
""" Copyright (c) 2014, Are Hansen - Honeypot Development. All rights reserved. Redistribution and use in source and binary forms, with or without modification, are permitted provided that the following conditions are met: 1. Redistributions of source code must retain the above copyright notice, this list of conditions and the following disclaimer. 2. Redistributions in binary form must reproduce the above copyright notice, this list of conditions and the following disclaimer in the documentation and/or other materials provided with the distribution. THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS "AS IS" AND AN EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT HOLDER OR CONTRIBUTORS BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF SUCH DAMAGE. """ __author__ = 'Are Hansen' __date__ = '2014, July 25' __version__ = '0.0.1' def sourceIPv4(loglines): """Splits the loglines on ',' and assumes that there will be an IP address at index[1], this IP address is appended to the output list and returned from the function. """ output = [] for line in loglines: line = line.split(',') output.append(line[1]) return output def passwords(loglines): """Splits the loglines on ',' and assumes that there will be a usename at index[3], which will be appended to the output list and returned from the function. """ output = [] for line in loglines: line = line.split(',') output.append('{0}'.format(line[3])) return output def passwords(loglines): """Splits the loglines on ',' and assumes that there will be a password at index[3], which will be appended to the output list and returned from the function. """ output = [] for line in loglines: line = line.split(',') output.append('{0}'.format(line[3])) return output def users(loglines): """Splits the loglines on ',' and assumes that there will be a username at index[2], which will be appended to the output list and returned from the function. """ output = [] for line in loglines: line = line.split(',') output.append('{0}'.format(line[2])) return output def combos(loglines): """Splits the loglines on ',' and assumes that there will be a username at index[2] and a password at index[4], which will be appended to the output list and returned from the function. """ output = [] for line in loglines: line = line.split(',') output.append('{0}/{1}'.format(line[2], line[3])) return output def access(loglines): """Checks for successful logins in the HonSSH loglines. This is done by itterating trough the loglines while splitting on ','. The last index of the list item thats created by doing so will either be a 0 (failed login) or a 1 (successful login). Any line with a 1 in index[4] will be appended to the output list and returned from the function. """ output = [] for line in loglines: line = line.split(',') if '1' in line[4]: output.append('{0} {1} {2} {3}'.format(line[0], line[1], line[2], line[3])) return output
tmaeda11235/schroedinger_solver
refs/heads/master
tests/nelsonrun.py
1
import schrpy as sch import scipy as sci cache = sci.zeros((100, 100000)) print("cache") psi = sci.load('Huskp_v[6.5]a[1.4]b[0.2]k[2.0].npy')[:-15001] + 2 ** -50 print("psi") x = sci.arange(-60, 60, 0.005) print("x") t = sci.arange(0, 15, 0.001) print("t") init = sci.random.normal(loc=-15, size=100000, scale=3.0) print("init") nel = sch.Nelson(x, t, psi, init) print("nel") k = 0 s = 0 while nel.t <= 15.: cache[k] = nel.run(0.001) k += 1 print('\r{}.{} doing! '.format(s, k), end=' ', flush=True) if k == 100: sci.save("RW_{}.npy".format(s), cache) s += 1 k = 0
bongtrop/python-btlogic
refs/heads/master
btlogic.py
1
#!/usr/bin/python import math #Minterm Maxterm class #Author: Pongsakorn Sommalai #Detail: Use for management logic term class term: def __init__(self, term=[], type="", bit=0): if type=="maxterm": self.term = term self.type = type else: self.term = term self.type = "minterm" self.bit=bit self.term.sort() def __m2m(self): out = [] if self.bit==0: return [] for i in range(0,int(math.pow(2,self.bit))): if i not in self.term: out.append(i) return out def __dec2bin(self, dec,bit): return ('0'*(bit-len(bin(dec)[2:])))+bin(dec)[2:] #Set variable def setTerm(self, term=[], type="", bit=0): if type=="maxterm": self.term = term.sort() self.type = type else: self.term = term.sort() self.type = "minterm" def getBit(self): return self.bit #Return minterm by list def getMinterm(self): if self.type == "minterm": return self.term else: return self.__m2m() #Return maxterm by list def getMaxterm(self): if self.type == "maxterm": return self.term else: return self.__m2m() #Return minterm by string def getMintermByAlphabet(self): alfa = "abcdefghijklmnopqrstuvwxyz" alfa = list(alfa) if self.type == "maxterm": term = self.__m2m() else: term = self.term out = "" for i in term: binterm=self.__dec2bin(i, self.bit) c=0 for j in list(binterm): if j == "0": out = out+alfa[c]+"\'" else: out = out+alfa[c] c = c+1 out = out+'+' return out[:-1] #Return maxterm by string def getMaxtermByAlphabet(self): alfa = "abcdefghijklmnopqrstuvwxyz" alfa = list(alfa) if self.type == "minterm": term = self.__m2m() else: term = self.term out = "" for i in term: out = out+'(' binterm=self.__dec2bin(i, self.bit) c=0 for j in list(binterm): if j == "1": out = out+alfa[c]+"\'" else: out = out+alfa[c] out = out+'+' c = c+1 out = out[:-1]+')' return out #Return minterm in binary def getMintermByBinary(self): alfa = "abcdefghijklmnopqrstuvwxyz" alfa = list(alfa) if self.type == "maxterm": term = self.__m2m() else: term = self.term out = [] for i in term: out.append(self.__dec2bin(i, self.bit)) return out #Return maxterm in binary def getMaxtermByBinary(self): alfa = "abcdefghijklmnopqrstuvwxyz" alfa = list(alfa) if self.type == "minterm": term = self.__m2m() else: term = self.term out = [] for i in term: out.append(self.__dec2bin(i, self.bit)) return out #Set to not term def opposite(self): self.term=self.__m2m() #Minterm Maxterm class #Author: Pongsakorn Sommalai #Detail: Use for abbreviated offseason a logic term class quine: #constucter def __init__(self, objterm=None, minterm=[], bit=0): if objterm!=None: self.term=objterm else: self.term=term(minterm, "minterm", bit) #check 1 bit not match def __check1bit(self, a, b): c = 0 for i in range(0, self.term.getBit()): if a[i]!=b[i]: c = c+1 if c==1: return True return False #match group def __matchGroup(self, group1, group2): newGroup=[] ngroup1=[] ngroup2=[] for term1 in group1: for term2 in group2: if self.__check1bit(term1, term2): newterm = "" for i in range(0, self.term.getBit()): if term1[i]!=term2[i]: newterm = newterm+"-" else: newterm = newterm+term1[i] newterm = newterm+"-" if term1[:-1]+'+' not in ngroup1: ngroup1.append(term1[:-1]+'+') if term2[:-1]+'+' not in ngroup2: ngroup2.append(term2[:-1]+'+') if newterm not in newGroup: newGroup.append(newterm) for term1 in group1: if term1[:-1]+'+' not in ngroup1: ngroup1.append(term1) for term2 in group2: if term2[:-1]+'+' not in ngroup2: ngroup2.append(term2) return [ngroup1,ngroup2,newGroup] #Set term def setTerm(self, objterm=None, minterm=[], bit=0): if objterm==None: self.term=objterm else: self.term=term(minterm, "minterm", bit) #Sort group def sortGroup(self, terms): out = [] for i in range(0, terms.getBit()+1): out.append([]) for term in terms.getMintermByBinary(): out[term.count('1')].append(term+"-") c=0 for o in out: if o==[]: del out[c] c = c+1 return out #Start quine return list of logic term def start(self): groups = self.sortGroup(self.term) c = 99 newgroups = [] ng = [] end = [] while c==1 or c==99: ng = [] if c==99: newgroups = groups c=0 for i in range (0,len(newgroups)-1): result = self.__matchGroup(newgroups[i], newgroups[i+1]) newgroups[i+1] = result[1] end.append(result[0]) if result[2]!=[]: c=1 ng.append(result[2]) if len(newgroups)-1==0: end.append(newgroups[i]) elif c==0: end.append(result[1]) newgroups = ng merge = [] for e in end: merge = merge+e out = [] for m in merge: if m[-1]=='-': out.append(m[:-1]) return out #Change list of term to string term def getTermByAlphabet(self, term): alfa = "abcdefghijklmnopqrstuvwxyz" alfa = list(alfa) out = "" for i in term: c=0 for j in list(i): if j == "0": out = out+alfa[c]+"\'" elif j == "-": c = c+1 continue else: out = out+alfa[c] c = c+1 out = out+'+' return out[:-1]
shepdelacreme/ansible
refs/heads/devel
lib/ansible/utils/__init__.py
2520
# (c) 2012-2014, Michael DeHaan <michael.dehaan@gmail.com> # # This file is part of Ansible # # Ansible is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # Ansible is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with Ansible. If not, see <http://www.gnu.org/licenses/>. # Make coding more python3-ish from __future__ import (absolute_import, division, print_function) __metaclass__ = type
Kimball12/DOUAudioStreamer-master
refs/heads/master
python/setup.py
11
# vim: set ft=python fenc=utf-8 sw=4 ts=4 et: # # DOUAudioStreamer - A Core Audio based streaming audio player for iOS/Mac: # # https://github.com/douban/DOUAudioStreamer # # Copyright 2013-2014 Douban Inc. All rights reserved. # # Use and distribution licensed under the BSD license. See # the LICENSE file for full text. # # Authors: # Chongyu Zhu <i@lembacon.com> # # from distutils.core import setup, Extension import glob FRAMEWORKS = ['Accelerate', 'CFNetwork', 'CoreAudio', 'AudioToolbox', 'AudioUnit', 'CoreServices'] DOUAS_MODULE = Extension('douas', sources=glob.glob('../src/*.m') + ['douas.m'], include_dirs=['../src'], extra_compile_args=['-fobjc-arc'], extra_link_args=[item for f in FRAMEWORKS for item in ['-framework', f]]) setup( name='douas', version='0.2.10', description='A Core Audio based streaming audio player for iOS/Mac', url='https://github.com/douban/DOUAudioStreamer', ext_modules=[DOUAS_MODULE], keywords=['douas', 'audio', 'streamer'], license='BSD', author='Chongyu Zhu', author_email='i@lembacon.com' )
fyffyt/scikit-learn
refs/heads/master
examples/linear_model/plot_lasso_model_selection.py
311
""" =================================================== Lasso model selection: Cross-Validation / AIC / BIC =================================================== Use the Akaike information criterion (AIC), the Bayes Information criterion (BIC) and cross-validation to select an optimal value of the regularization parameter alpha of the :ref:`lasso` estimator. Results obtained with LassoLarsIC are based on AIC/BIC criteria. Information-criterion based model selection is very fast, but it relies on a proper estimation of degrees of freedom, are derived for large samples (asymptotic results) and assume the model is correct, i.e. that the data are actually generated by this model. They also tend to break when the problem is badly conditioned (more features than samples). For cross-validation, we use 20-fold with 2 algorithms to compute the Lasso path: coordinate descent, as implemented by the LassoCV class, and Lars (least angle regression) as implemented by the LassoLarsCV class. Both algorithms give roughly the same results. They differ with regards to their execution speed and sources of numerical errors. Lars computes a path solution only for each kink in the path. As a result, it is very efficient when there are only of few kinks, which is the case if there are few features or samples. Also, it is able to compute the full path without setting any meta parameter. On the opposite, coordinate descent compute the path points on a pre-specified grid (here we use the default). Thus it is more efficient if the number of grid points is smaller than the number of kinks in the path. Such a strategy can be interesting if the number of features is really large and there are enough samples to select a large amount. In terms of numerical errors, for heavily correlated variables, Lars will accumulate more errors, while the coordinate descent algorithm will only sample the path on a grid. Note how the optimal value of alpha varies for each fold. This illustrates why nested-cross validation is necessary when trying to evaluate the performance of a method for which a parameter is chosen by cross-validation: this choice of parameter may not be optimal for unseen data. """ print(__doc__) # Author: Olivier Grisel, Gael Varoquaux, Alexandre Gramfort # License: BSD 3 clause import time import numpy as np import matplotlib.pyplot as plt from sklearn.linear_model import LassoCV, LassoLarsCV, LassoLarsIC from sklearn import datasets diabetes = datasets.load_diabetes() X = diabetes.data y = diabetes.target rng = np.random.RandomState(42) X = np.c_[X, rng.randn(X.shape[0], 14)] # add some bad features # normalize data as done by Lars to allow for comparison X /= np.sqrt(np.sum(X ** 2, axis=0)) ############################################################################## # LassoLarsIC: least angle regression with BIC/AIC criterion model_bic = LassoLarsIC(criterion='bic') t1 = time.time() model_bic.fit(X, y) t_bic = time.time() - t1 alpha_bic_ = model_bic.alpha_ model_aic = LassoLarsIC(criterion='aic') model_aic.fit(X, y) alpha_aic_ = model_aic.alpha_ def plot_ic_criterion(model, name, color): alpha_ = model.alpha_ alphas_ = model.alphas_ criterion_ = model.criterion_ plt.plot(-np.log10(alphas_), criterion_, '--', color=color, linewidth=3, label='%s criterion' % name) plt.axvline(-np.log10(alpha_), color=color, linewidth=3, label='alpha: %s estimate' % name) plt.xlabel('-log(alpha)') plt.ylabel('criterion') plt.figure() plot_ic_criterion(model_aic, 'AIC', 'b') plot_ic_criterion(model_bic, 'BIC', 'r') plt.legend() plt.title('Information-criterion for model selection (training time %.3fs)' % t_bic) ############################################################################## # LassoCV: coordinate descent # Compute paths print("Computing regularization path using the coordinate descent lasso...") t1 = time.time() model = LassoCV(cv=20).fit(X, y) t_lasso_cv = time.time() - t1 # Display results m_log_alphas = -np.log10(model.alphas_) plt.figure() ymin, ymax = 2300, 3800 plt.plot(m_log_alphas, model.mse_path_, ':') plt.plot(m_log_alphas, model.mse_path_.mean(axis=-1), 'k', label='Average across the folds', linewidth=2) plt.axvline(-np.log10(model.alpha_), linestyle='--', color='k', label='alpha: CV estimate') plt.legend() plt.xlabel('-log(alpha)') plt.ylabel('Mean square error') plt.title('Mean square error on each fold: coordinate descent ' '(train time: %.2fs)' % t_lasso_cv) plt.axis('tight') plt.ylim(ymin, ymax) ############################################################################## # LassoLarsCV: least angle regression # Compute paths print("Computing regularization path using the Lars lasso...") t1 = time.time() model = LassoLarsCV(cv=20).fit(X, y) t_lasso_lars_cv = time.time() - t1 # Display results m_log_alphas = -np.log10(model.cv_alphas_) plt.figure() plt.plot(m_log_alphas, model.cv_mse_path_, ':') plt.plot(m_log_alphas, model.cv_mse_path_.mean(axis=-1), 'k', label='Average across the folds', linewidth=2) plt.axvline(-np.log10(model.alpha_), linestyle='--', color='k', label='alpha CV') plt.legend() plt.xlabel('-log(alpha)') plt.ylabel('Mean square error') plt.title('Mean square error on each fold: Lars (train time: %.2fs)' % t_lasso_lars_cv) plt.axis('tight') plt.ylim(ymin, ymax) plt.show()
qmagico/sampleappqm
refs/heads/master
src/django/contrib/localflavor/si/si_postalcodes.py
89
# *-* coding: utf-8 *-* SI_POSTALCODES = [ (1000, u'Ljubljana'), (1215, u'Medvode'), (1216, u'Smlednik'), (1217, u'Vodice'), (1218, u'Komenda'), (1219, u'Laze v Tuhinju'), (1221, u'Motnik'), (1222, u'Trojane'), (1223, u'Blagovica'), (1225, u'Lukovica'), (1230, u'Dom\u017eale'), (1233, u'Dob'), (1234, u'Menge\u0161'), (1235, u'Radomlje'), (1236, u'Trzin'), (1241, u'Kamnik'), (1242, u'Stahovica'), (1251, u'Morav\u010de'), (1252, u'Va\u010de'), (1262, u'Dol pri Ljubljani'), (1270, u'Litija'), (1272, u'Pol\u0161nik'), (1273, u'Dole pri Litiji'), (1274, u'Gabrovka'), (1275, u'\u0160martno pri Litiji'), (1276, u'Primskovo'), (1281, u'Kresnice'), (1282, u'Sava'), (1290, u'Grosuplje'), (1291, u'\u0160kofljica'), (1292, u'Ig'), (1293, u'\u0160marje - Sap'), (1294, u'Vi\u0161nja Gora'), (1295, u'Ivan\u010dna Gorica'), (1296, u'\u0160entvid pri Sti\u010dni'), (1301, u'Krka'), (1303, u'Zagradec'), (1310, u'Ribnica'), (1311, u'Turjak'), (1312, u'Videm - Dobrepolje'), (1313, u'Struge'), (1314, u'Rob'), (1315, u'Velike La\u0161\u010de'), (1316, u'Ortnek'), (1317, u'Sodra\u017eica'), (1318, u'Lo\u0161ki Potok'), (1319, u'Draga'), (1330, u'Ko\u010devje'), (1331, u'Dolenja vas'), (1332, u'Stara Cerkev'), (1336, u'Kostel'), (1337, u'Osilnica'), (1338, u'Ko\u010devska Reka'), (1351, u'Brezovica pri Ljubljani'), (1352, u'Preserje'), (1353, u'Borovnica'), (1354, u'Horjul'), (1355, u'Polhov Gradec'), (1356, u'Dobrova'), (1357, u'Notranje Gorice'), (1358, u'Log pri Brezovici'), (1360, u'Vrhnika'), (1370, u'Logatec'), (1372, u'Hotedr\u0161ica'), (1373, u'Rovte'), (1380, u'Cerknica'), (1381, u'Rakek'), (1382, u'Begunje pri Cerknici'), (1384, u'Grahovo'), (1385, u'Nova vas'), (1386, u'Stari trg pri Lo\u017eu'), (1410, u'Zagorje ob Savi'), (1411, u'Izlake'), (1412, u'Kisovec'), (1413, u'\u010cem\u0161enik'), (1414, u'Podkum'), (1420, u'Trbovlje'), (1423, u'Dobovec'), (1430, u'Hrastnik'), (1431, u'Dol pri Hrastniku'), (1432, u'Zidani Most'), (1433, u'Rade\u010de'), (1434, u'Loka pri Zidanem Mostu'), (2000, u'Maribor'), (2201, u'Zgornja Kungota'), (2204, u'Miklav\u017e na Dravskem polju'), (2205, u'Star\u0161e'), (2206, u'Marjeta na Dravskem polju'), (2208, u'Pohorje'), (2211, u'Pesnica pri Mariboru'), (2212, u'\u0160entilj v Slovenskih goricah'), (2213, u'Zgornja Velka'), (2214, u'Sladki vrh'), (2215, u'Cer\u0161ak'), (2221, u'Jarenina'), (2222, u'Jakobski Dol'), (2223, u'Jurovski Dol'), (2229, u'Male\u010dnik'), (2230, u'Lenart v Slovenskih goricah'), (2231, u'Pernica'), (2232, u'Voli\u010dina'), (2233, u'Sveta Ana v Slovenskih goricah'), (2234, u'Benedikt'), (2235, u'Sveta Trojica v Slovenskih goricah'), (2236, u'Cerkvenjak'), (2241, u'Spodnji Duplek'), (2242, u'Zgornja Korena'), (2250, u'Ptuj'), (2252, u'Dornava'), (2253, u'Destrnik'), (2254, u'Trnovska vas'), (2255, u'Vitomarci'), (2256, u'Jur\u0161inci'), (2257, u'Polen\u0161ak'), (2258, u'Sveti Toma\u017e'), (2259, u'Ivanjkovci'), (2270, u'Ormo\u017e'), (2272, u'Gori\u0161nica'), (2273, u'Podgorci'), (2274, u'Velika Nedelja'), (2275, u'Miklav\u017e pri Ormo\u017eu'), (2276, u'Kog'), (2277, u'Sredi\u0161\u010de ob Dravi'), (2281, u'Markovci'), (2282, u'Cirkulane'), (2283, u'Zavr\u010d'), (2284, u'Videm pri Ptuju'), (2285, u'Zgornji Leskovec'), (2286, u'Podlehnik'), (2287, u'\u017detale'), (2288, u'Hajdina'), (2289, u'Stoperce'), (2310, u'Slovenska Bistrica'), (2311, u'Ho\u010de'), (2312, u'Orehova vas'), (2313, u'Fram'), (2314, u'Zgornja Polskava'), (2315, u'\u0160martno na Pohorju'), (2316, u'Zgornja Lo\u017enica'), (2317, u'Oplotnica'), (2318, u'Laporje'), (2319, u'Polj\u010dane'), (2321, u'Makole'), (2322, u'Maj\u0161perk'), (2323, u'Ptujska Gora'), (2324, u'Lovrenc na Dravskem polju'), (2325, u'Kidri\u010devo'), (2326, u'Cirkovce'), (2327, u'Ra\u010de'), (2331, u'Pragersko'), (2341, u'Limbu\u0161'), (2342, u'Ru\u0161e'), (2343, u'Fala'), (2344, u'Lovrenc na Pohorju'), (2345, u'Bistrica ob Dravi'), (2351, u'Kamnica'), (2352, u'Selnica ob Dravi'), (2353, u'Sv. Duh na Ostrem Vrhu'), (2354, u'Bresternica'), (2360, u'Radlje ob Dravi'), (2361, u'O\u017ebalt'), (2362, u'Kapla'), (2363, u'Podvelka'), (2364, u'Ribnica na Pohorju'), (2365, u'Vuhred'), (2366, u'Muta'), (2367, u'Vuzenica'), (2370, u'Dravograd'), (2371, u'Trbonje'), (2372, u'Libeli\u010de'), (2373, u'\u0160entjan\u017e pri Dravogradu'), (2380, u'Slovenj Gradec'), (2381, u'Podgorje pri Slovenj Gradcu'), (2382, u'Mislinja'), (2383, u'\u0160martno pri Slovenj Gradcu'), (2390, u'Ravne na Koro\u0161kem'), (2391, u'Prevalje'), (2392, u'Me\u017eica'), (2393, u'\u010crna na Koro\u0161kem'), (2394, u'Kotlje'), (3000, u'Celje'), (3201, u'\u0160martno v Ro\u017eni dolini'), (3202, u'Ljube\u010dna'), (3203, u'Nova Cerkev'), (3204, u'Dobrna'), (3205, u'Vitanje'), (3206, u'Stranice'), (3210, u'Slovenske Konjice'), (3211, u'\u0160kofja vas'), (3212, u'Vojnik'), (3213, u'Frankolovo'), (3214, u'Zre\u010de'), (3215, u'Lo\u010de'), (3220, u'\u0160tore'), (3221, u'Teharje'), (3222, u'Dramlje'), (3223, u'Loka pri \u017dusmu'), (3224, u'Dobje pri Planini'), (3225, u'Planina pri Sevnici'), (3230, u'\u0160entjur'), (3231, u'Grobelno'), (3232, u'Ponikva'), (3233, u'Kalobje'), (3240, u'\u0160marje pri Jel\u0161ah'), (3241, u'Podplat'), (3250, u'Roga\u0161ka Slatina'), (3252, u'Rogatec'), (3253, u'Pristava pri Mestinju'), (3254, u'Pod\u010detrtek'), (3255, u'Bu\u010de'), (3256, u'Bistrica ob Sotli'), (3257, u'Podsreda'), (3260, u'Kozje'), (3261, u'Lesi\u010dno'), (3262, u'Prevorje'), (3263, u'Gorica pri Slivnici'), (3264, u'Sveti \u0160tefan'), (3270, u'La\u0161ko'), (3271, u'\u0160entrupert'), (3272, u'Rimske Toplice'), (3273, u'Jurklo\u0161ter'), (3301, u'Petrov\u010de'), (3302, u'Gri\u017ee'), (3303, u'Gomilsko'), (3304, u'Tabor'), (3305, u'Vransko'), (3310, u'\u017dalec'), (3311, u'\u0160empeter v Savinjski dolini'), (3312, u'Prebold'), (3313, u'Polzela'), (3314, u'Braslov\u010de'), (3320, u'Velenje - dostava'), (3322, u'Velenje - po\u0161tni predali'), (3325, u'\u0160o\u0161tanj'), (3326, u'Topol\u0161ica'), (3327, u'\u0160martno ob Paki'), (3330, u'Mozirje'), (3331, u'Nazarje'), (3332, u'Re\u010dica ob Savinji'), (3333, u'Ljubno ob Savinji'), (3334, u'Lu\u010de'), (3335, u'Sol\u010dava'), (3341, u'\u0160martno ob Dreti'), (3342, u'Gornji Grad'), (4000, u'Kranj'), (4201, u'Zgornja Besnica'), (4202, u'Naklo'), (4203, u'Duplje'), (4204, u'Golnik'), (4205, u'Preddvor'), (4206, u'Zgornje Jezersko'), (4207, u'Cerklje na Gorenjskem'), (4208, u'\u0160en\u010dur'), (4209, u'\u017dabnica'), (4210, u'Brnik - aerodrom'), (4211, u'Mav\u010di\u010de'), (4212, u'Visoko'), (4220, u'\u0160kofja Loka'), (4223, u'Poljane nad \u0160kofjo Loko'), (4224, u'Gorenja vas'), (4225, u'Sovodenj'), (4226, u'\u017diri'), (4227, u'Selca'), (4228, u'\u017delezniki'), (4229, u'Sorica'), (4240, u'Radovljica'), (4243, u'Brezje'), (4244, u'Podnart'), (4245, u'Kropa'), (4246, u'Kamna Gorica'), (4247, u'Zgornje Gorje'), (4248, u'Lesce'), (4260, u'Bled'), (4263, u'Bohinjska Bela'), (4264, u'Bohinjska Bistrica'), (4265, u'Bohinjsko jezero'), (4267, u'Srednja vas v Bohinju'), (4270, u'Jesenice'), (4273, u'Blejska Dobrava'), (4274, u'\u017dirovnica'), (4275, u'Begunje na Gorenjskem'), (4276, u'Hru\u0161ica'), (4280, u'Kranjska Gora'), (4281, u'Mojstrana'), (4282, u'Gozd Martuljek'), (4283, u'Rate\u010de - Planica'), (4290, u'Tr\u017ei\u010d'), (4294, u'Kri\u017ee'), (5000, u'Nova Gorica'), (5210, u'Deskle'), (5211, u'Kojsko'), (5212, u'Dobrovo v Brdih'), (5213, u'Kanal'), (5214, u'Kal nad Kanalom'), (5215, u'Ro\u010dinj'), (5216, u'Most na So\u010di'), (5220, u'Tolmin'), (5222, u'Kobarid'), (5223, u'Breginj'), (5224, u'Srpenica'), (5230, u'Bovec'), (5231, u'Log pod Mangartom'), (5232, u'So\u010da'), (5242, u'Grahovo ob Ba\u010di'), (5243, u'Podbrdo'), (5250, u'Solkan'), (5251, u'Grgar'), (5252, u'Trnovo pri Gorici'), (5253, u'\u010cepovan'), (5261, u'\u0160empas'), (5262, u'\u010crni\u010de'), (5263, u'Dobravlje'), (5270, u'Ajdov\u0161\u010dina'), (5271, u'Vipava'), (5272, u'Podnanos'), (5273, u'Col'), (5274, u'\u010crni Vrh nad Idrijo'), (5275, u'Godovi\u010d'), (5280, u'Idrija'), (5281, u'Spodnja Idrija'), (5282, u'Cerkno'), (5283, u'Slap ob Idrijci'), (5290, u'\u0160empeter pri Gorici'), (5291, u'Miren'), (5292, u'Ren\u010de'), (5293, u'Vol\u010dja Draga'), (5294, u'Dornberk'), (5295, u'Branik'), (5296, u'Kostanjevica na Krasu'), (5297, u'Prva\u010dina'), (6000, u'Koper'), (6210, u'Se\u017eana'), (6215, u'Diva\u010da'), (6216, u'Podgorje'), (6217, u'Vremski Britof'), (6219, u'Lokev'), (6221, u'Dutovlje'), (6222, u'\u0160tanjel'), (6223, u'Komen'), (6224, u'Seno\u017ee\u010de'), (6225, u'Hru\u0161evje'), (6230, u'Postojna'), (6232, u'Planina'), (6240, u'Kozina'), (6242, u'Materija'), (6243, u'Obrov'), (6244, u'Podgrad'), (6250, u'Ilirska Bistrica'), (6251, u'Ilirska Bistrica - Trnovo'), (6253, u'Kne\u017eak'), (6254, u'Jel\u0161ane'), (6255, u'Prem'), (6256, u'Ko\u0161ana'), (6257, u'Pivka'), (6258, u'Prestranek'), (6271, u'Dekani'), (6272, u'Gra\u010di\u0161\u010de'), (6273, u'Marezige'), (6274, u'\u0160marje'), (6275, u'\u010crni Kal'), (6276, u'Pobegi'), (6280, u'Ankaran - Ancarano'), (6281, u'\u0160kofije'), (6310, u'Izola - Isola'), (6320, u'Portoro\u017e - Portorose'), (6330, u'Piran - Pirano'), (6333, u'Se\u010dovlje - Sicciole'), (8000, u'Novo mesto'), (8210, u'Trebnje'), (8211, u'Dobrni\u010d'), (8212, u'Velika Loka'), (8213, u'Veliki Gaber'), (8216, u'Mirna Pe\u010d'), (8220, u'\u0160marje\u0161ke Toplice'), (8222, u'Oto\u010dec'), (8230, u'Mokronog'), (8231, u'Trebelno'), (8232, u'\u0160entrupert'), (8233, u'Mirna'), (8250, u'Bre\u017eice'), (8251, u'\u010cate\u017e ob Savi'), (8253, u'Arti\u010de'), (8254, u'Globoko'), (8255, u'Pi\u0161ece'), (8256, u'Sromlje'), (8257, u'Dobova'), (8258, u'Kapele'), (8259, u'Bizeljsko'), (8261, u'Jesenice na Dolenjskem'), (8262, u'Kr\u0161ka vas'), (8263, u'Cerklje ob Krki'), (8270, u'Kr\u0161ko'), (8272, u'Zdole'), (8273, u'Leskovec pri Kr\u0161kem'), (8274, u'Raka'), (8275, u'\u0160kocjan'), (8276, u'Bu\u010dka'), (8280, u'Brestanica'), (8281, u'Senovo'), (8282, u'Koprivnica'), (8283, u'Blanca'), (8290, u'Sevnica'), (8292, u'Zabukovje'), (8293, u'Studenec'), (8294, u'Bo\u0161tanj'), (8295, u'Tr\u017ei\u0161\u010de'), (8296, u'Krmelj'), (8297, u'\u0160entjan\u017e'), (8310, u'\u0160entjernej'), (8311, u'Kostanjevica na Krki'), (8312, u'Podbo\u010dje'), (8321, u'Brusnice'), (8322, u'Stopi\u010de'), (8323, u'Ur\u0161na sela'), (8330, u'Metlika'), (8331, u'Suhor'), (8332, u'Gradac'), (8333, u'Semi\u010d'), (8340, u'\u010crnomelj'), (8341, u'Adle\u0161i\u010di'), (8342, u'Stari trg ob Kolpi'), (8343, u'Dragatu\u0161'), (8344, u'Vinica pri \u010crnomlju'), (8350, u'Dolenjske Toplice'), (8351, u'Stra\u017ea'), (8360, u'\u017du\u017eemberk'), (8361, u'Dvor'), (8362, u'Hinje'), (9000, u'Murska Sobota'), (9201, u'Puconci'), (9202, u'Ma\u010dkovci'), (9203, u'Petrovci'), (9204, u'\u0160alovci'), (9205, u'Hodo\u0161 - Hodos'), (9206, u'Kri\u017eevci'), (9207, u'Prosenjakovci - Partosfalva'), (9208, u'Fokovci'), (9220, u'Lendava - Lendva'), (9221, u'Martjanci'), (9222, u'Bogojina'), (9223, u'Dobrovnik - Dobronak'), (9224, u'Turni\u0161\u010de'), (9225, u'Velika Polana'), (9226, u'Moravske Toplice'), (9227, u'Kobilje'), (9231, u'Beltinci'), (9232, u'\u010cren\u0161ovci'), (9233, u'Odranci'), (9240, u'Ljutomer'), (9241, u'Ver\u017eej'), (9242, u'Kri\u017eevci pri Ljutomeru'), (9243, u'Mala Nedelja'), (9244, u'Sveti Jurij ob \u0160\u010davnici'), (9245, u'Spodnji Ivanjci'), (9250, u'Gornja Radgona'), (9251, u'Ti\u0161ina'), (9252, u'Radenci'), (9253, u'Apa\u010de'), (9261, u'Cankova'), (9262, u'Roga\u0161ovci'), (9263, u'Kuzma'), (9264, u'Grad'), (9265, u'Bodonci'), ] SI_POSTALCODES_CHOICES = sorted(SI_POSTALCODES, key=lambda k: k[1])
byterom/android_external_chromium_org
refs/heads/12.1
chrome/common/extensions/docs/server2/manifest_data_source_test.py
87
#!/usr/bin/env python # Copyright 2013 The Chromium Authors. All rights reserved. # Use of this source code is governed by a BSD-style license that can be # found in the LICENSE file. from copy import deepcopy import json import unittest from future import Future import manifest_data_source from object_store_creator import ObjectStoreCreator convert_and_annotate_docs = { 'name': { 'example': "My {{platform}}", 'name': 'name', 'level': 'required' }, 'doc2': { 'level': 'required', 'name': 'doc2' }, 'doc1': { 'level': 'required', 'name': 'doc1', 'children': { 'sub1': { 'annotations': ['not so important'], 'level': 'optional', 'name': 'sub1' }, 'sub2': { 'level': 'required', 'name': 'sub2' } } }, 'doc3': { 'level': 'only_one', 'name': 'doc3' }, 'doc4': { 'level': 'recommended', 'name': 'doc4' }, 'doc5': { 'level': 'only_one', 'name': 'doc5' }, 'doc6': { 'level': 'optional', 'name': 'doc6' } } class ManifestDataSourceTest(unittest.TestCase): def testListifyAndSortDocs(self): expected_docs = [ { 'level': 'required', 'name': 'doc1', 'children': [ { 'level': 'required', 'name': 'sub2' }, { 'annotations': ['not so important'], 'level': 'optional', 'name': 'sub1' } ] }, { 'level': 'required', 'name': 'doc2' }, { 'level': 'required', 'example': '"My App"', 'has_example': True, 'name': 'name' }, { 'level': 'recommended', 'name': 'doc4' }, { 'level': 'only_one', 'name': 'doc3' }, { 'level': 'only_one', 'name': 'doc5' }, { 'level': 'optional', 'name': 'doc6' } ] self.assertEqual(expected_docs, manifest_data_source._ListifyAndSortDocs( deepcopy(convert_and_annotate_docs), 'App')) def testAnnotate(self): expected_docs = [ { 'level': 'required', 'name': 'doc1', 'children': [ { 'level': 'required', 'name': 'sub2' }, { 'annotations': ['Optional', 'not so important'], 'level': 'optional', 'name': 'sub1', 'is_last': True } ] }, { 'level': 'required', 'name': 'doc2' }, { 'name': 'name', 'level': 'required', 'example': '"My App"', 'has_example': True }, { 'annotations': ['Recommended'], 'level': 'recommended', 'name': 'doc4' }, { 'annotations': ['Pick one (or none)'], 'level': 'only_one', 'name': 'doc3' }, { 'level': 'only_one', 'name': 'doc5' }, { 'annotations': ['Optional'], 'level': 'optional', 'name': 'doc6', 'is_last': True } ] annotated = manifest_data_source._ListifyAndSortDocs( deepcopy(convert_and_annotate_docs), 'App') manifest_data_source._AddLevelAnnotations(annotated) self.assertEqual(expected_docs, annotated) def testExpandedExamples(self): docs = { 'doc1': { 'name': 'doc1', 'example': { 'big': { 'nested': { 'json_example': ['with', 'more', 'json'] } } } } } expected_docs = [ { 'name': 'doc1', 'children': [ { 'name': 'big', 'children': [ { 'name': 'nested', 'children': [ { 'name': 'json_example', 'example': json.dumps(['with', 'more', 'json']), 'has_example': True } ] } ] } ] } ] self.assertEqual( expected_docs, manifest_data_source._ListifyAndSortDocs(docs, 'apps')) def testNonExpandedExamples(self): docs = { 'doc1': { 'name': 'doc1', 'example': {} }, 'doc2': { 'name': 'doc2', 'example': [] }, 'doc3': { 'name': 'doc3', 'example': [{}] } } expected_docs = [ { 'name': 'doc1', 'has_example': True, 'example': '{...}' }, { 'name': 'doc2', 'has_example': True, 'example': '[...]' }, { 'name': 'doc3', 'has_example': True, 'example': '[{...}]' } ] self.assertEqual( expected_docs, manifest_data_source._ListifyAndSortDocs(docs, 'apps')) def testManifestDataSource(self): manifest_features = { 'doc1': { 'name': 'doc1', 'platforms': ['apps', 'extensions'], 'example': {}, 'level': 'required' }, 'doc1.sub1': { 'name': 'doc1.sub1', 'platforms': ['apps'], 'annotations': ['important!'], 'level': 'recommended' } } expected_app = [ { 'example': '{...}', 'has_example': True, 'level': 'required', 'name': 'doc1', 'platforms': ['apps', 'extensions'], 'children': [ { 'annotations': [ 'Recommended', 'important!' ], 'level': 'recommended', 'name': 'sub1', 'platforms': ['apps'], 'is_last': True } ], 'is_last': True } ] class FakePlatformBundle(object): def GetFeaturesBundle(self, platform): return FakeFeaturesBundle() class FakeFeaturesBundle(object): def GetManifestFeatures(self): return Future(value=manifest_features) class FakeServerInstance(object): def __init__(self): self.platform_bundle = FakePlatformBundle() self.object_store_creator = ObjectStoreCreator.ForTest() mds = manifest_data_source.ManifestDataSource(FakeServerInstance(), None) self.assertEqual(expected_app, mds.get('apps')) if __name__ == '__main__': unittest.main()
RTHMaK/RPGOne
refs/heads/master
Documents/scikit-image-0.12.3/skimage/scripts/skivi.py
49
"""skimage viewer""" def main(): import skimage.io as io import sys if len(sys.argv) != 2: print("Usage: skivi <image-file>") sys.exit(-1) io.use_plugin('qt') io.imshow(io.imread(sys.argv[1]), fancy=True) io.show()
laosiaudi/tensorflow
refs/heads/master
tensorflow/contrib/slim/python/slim/queues.py
35
# Copyright 2015 The TensorFlow Authors. All Rights Reserved. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # ============================================================================== """Contains a helper context for running queue runners. @@NestedQueueRunnerError @@QueueRunners """ from __future__ import absolute_import from __future__ import division from __future__ import print_function from contextlib import contextmanager import threading from tensorflow.python.framework import ops from tensorflow.python.training import coordinator __all__ = [ 'NestedQueueRunnerError', 'QueueRunners', ] _queue_runner_lock = threading.Lock() class NestedQueueRunnerError(Exception): pass @contextmanager def QueueRunners(session): """Creates a context manager that handles starting and stopping queue runners. Args: session: the currently running session. Yields: a context in which queues are run. Raises: NestedQueueRunnerError: if a QueueRunners context is nested within another. """ if not _queue_runner_lock.acquire(False): raise NestedQueueRunnerError('QueueRunners cannot be nested') coord = coordinator.Coordinator() threads = [] for qr in ops.get_collection(ops.GraphKeys.QUEUE_RUNNERS): threads.extend(qr.create_threads(session, coord=coord, daemon=True, start=True)) try: yield finally: coord.request_stop() coord.join(threads, stop_grace_period_secs=120) _queue_runner_lock.release()
albertomurillo/ansible
refs/heads/devel
lib/ansible/modules/system/aix_devices.py
44
#!/usr/bin/python # -*- coding: utf-8 -*- # Copyright: (c) 2017, 2018 Kairo Araujo <kairo@kairo.eti.br> # GNU General Public License v3.0+ (see COPYING or https://www.gnu.org/licenses/gpl-3.0.txt) from __future__ import absolute_import, division, print_function __metaclass__ = type ANSIBLE_METADATA = { 'metadata_version': '1.1', 'status': ['preview'], 'supported_by': 'community' } DOCUMENTATION = r''' --- author: - Kairo Araujo (@kairoaraujo) module: aix_devices short_description: Manages AIX devices description: - This module discovers, defines, removes and modifies attributes of AIX devices. version_added: '2.8' options: attributes: description: - A list of device attributes. type: dict device: description: - The name of the device. - C(all) is valid to rescan C(available) all devices (AIX cfgmgr command). type: str required: true force: description: - Forces action. type: bool default: no recursive: description: - Removes or defines a device and children devices. type: bool default: no state: description: - Controls the device state. - C(available) (alias C(present)) rescan a specific device or all devices (when C(device) is not specified). - C(removed) (alias C(absent) removes a device. - C(defined) changes device to Defined state. type: str choices: [ available, defined, removed ] default: available ''' EXAMPLES = r''' - name: Scan new devices aix_devices: device: all state: available - name: Scan new virtual devices (vio0) aix_devices: device: vio0 state: available - name: Removing IP alias to en0 aix_devices: device: en0 attributes: delalias4: 10.0.0.100,255.255.255.0 - name: Removes ent2 aix_devices: device: ent2 state: removed - name: Put device en2 in Defined aix_devices: device: en2 state: defined - name: Removes ent4 (inexistent). aix_devices: device: ent4 state: removed - name: Put device en4 in Defined (inexistent) aix_devices: device: en4 state: defined - name: Put vscsi1 and children devices in Defined state. aix_devices: device: vscsi1 recursive: yes state: defined - name: Removes vscsi1 and children devices. aix_devices: device: vscsi1 recursive: yes state: removed - name: Changes en1 mtu to 9000 and disables arp. aix_devices: device: en1 attributes: mtu: 900 arp: off state: available - name: Configure IP, netmask and set en1 up. aix_devices: device: en1 attributes: netaddr: 192.168.0.100 netmask: 255.255.255.0 state: up state: available - name: Adding IP alias to en0 aix_devices: device: en0 attributes: alias4: 10.0.0.100,255.255.255.0 state: available ''' RETURN = r''' # ''' from ansible.module_utils.basic import AnsibleModule def _check_device(module, device): """ Check if device already exists and the state. Args: module: Ansible module. device: device to be checked. Returns: bool, device state """ lsdev_cmd = module.get_bin_path('lsdev', True) rc, lsdev_out, err = module.run_command(["%s" % lsdev_cmd, '-C', '-l', "%s" % device]) if rc != 0: module.fail_json(msg="Failed to run lsdev", rc=rc, err=err) if lsdev_out: device_state = lsdev_out.split()[1] return True, device_state device_state = None return False, device_state def _check_device_attr(module, device, attr): """ Args: module: Ansible module. device: device to check attributes. attr: attribute to be checked. Returns: """ lsattr_cmd = module.get_bin_path('lsattr', True) rc, lsattr_out, err = module.run_command(["%s" % lsattr_cmd, '-El', "%s" % device, '-a', "%s" % attr]) hidden_attrs = ['delalias4', 'delalias6'] if rc == 255: if attr in hidden_attrs: current_param = '' else: current_param = None return current_param elif rc != 0: module.fail_json(msg="Failed to run lsattr: %s" % err, rc=rc, err=err) current_param = lsattr_out.split()[1] return current_param def discover_device(module, device): """ Discover AIX devices.""" cfgmgr_cmd = module.get_bin_path('cfgmgr', True) if device is not None: device = "-l %s" % device else: device = '' changed = True msg = '' if not module.check_mode: rc, cfgmgr_out, err = module.run_command(["%s" % cfgmgr_cmd, "%s" % device]) changed = True msg = cfgmgr_out return changed, msg def change_device_attr(module, attributes, device, force): """ Change AIX device attribute. """ attr_changed = [] attr_not_changed = [] attr_invalid = [] chdev_cmd = module.get_bin_path('chdev', True) for attr in list(attributes.keys()): new_param = attributes[attr] current_param = _check_device_attr(module, device, attr) if current_param is None: attr_invalid.append(attr) elif current_param != new_param: if force: cmd = ["%s" % chdev_cmd, '-l', "%s" % device, '-a', "%s=%s" % (attr, attributes[attr]), "%s" % force] else: cmd = ["%s" % chdev_cmd, '-l', "%s" % device, '-a', "%s=%s" % (attr, attributes[attr])] if not module.check_mode: rc, chdev_out, err = module.run_command(cmd) if rc != 0: module.exit_json(msg="Failed to run chdev.", rc=rc, err=err) attr_changed.append(attributes[attr]) else: attr_not_changed.append(attributes[attr]) if len(attr_changed) > 0: changed = True attr_changed_msg = "Attributes changed: %s. " % ','.join(attr_changed) else: changed = False attr_changed_msg = '' if len(attr_not_changed) > 0: attr_not_changed_msg = "Attributes already set: %s. " % ','.join(attr_not_changed) else: attr_not_changed_msg = '' if len(attr_invalid) > 0: attr_invalid_msg = "Invalid attributes: %s " % ', '.join(attr_invalid) else: attr_invalid_msg = '' msg = "%s%s%s" % (attr_changed_msg, attr_not_changed_msg, attr_invalid_msg) return changed, msg def remove_device(module, device, force, recursive, state): """ Puts device in defined state or removes device. """ state_opt = { 'removed': '-d', 'absent': '-d', 'defined': '' } recursive_opt = { True: '-R', False: '' } recursive = recursive_opt[recursive] state = state_opt[state] changed = True msg = '' rmdev_cmd = module.get_bin_path('rmdev', True) if not module.check_mode: if state: rc, rmdev_out, err = module.run_command(["%s" % rmdev_cmd, "-l", "%s" % device, "%s" % recursive, "%s" % force]) else: rc, rmdev_out, err = module.run_command(["%s" % rmdev_cmd, "-l", "%s" % device, "%s" % recursive]) if rc != 0: module.fail_json(msg="Failed to run rmdev", rc=rc, err=err) msg = rmdev_out return changed, msg def main(): module = AnsibleModule( argument_spec=dict( attributes=dict(type='dict'), device=dict(type='str'), force=dict(type='bool', default=False), recursive=dict(type='bool', default=False), state=dict(type='str', default='available', choices=['available', 'defined', 'removed']), ), supports_check_mode=True, ) force_opt = { True: '-f', False: '', } attributes = module.params['attributes'] device = module.params['device'] force = force_opt[module.params['force']] recursive = module.params['recursive'] state = module.params['state'] result = dict( changed=False, msg='', ) if state == 'available' or state == 'present': if attributes: # change attributes on device device_status, device_state = _check_device(module, device) if device_status: result['changed'], result['msg'] = change_device_attr(module, attributes, device, force) else: result['msg'] = "Device %s does not exist." % device else: # discovery devices (cfgmgr) if device and device != 'all': device_status, device_state = _check_device(module, device) if device_status: # run cfgmgr on specific device result['changed'], result['msg'] = discover_device(module, device) else: result['msg'] = "Device %s does not exist." % device else: result['changed'], result['msg'] = discover_device(module, device) elif state == 'removed' or state == 'absent' or state == 'defined': if not device: result['msg'] = "device is required to removed or defined state." else: # Remove device check_device, device_state = _check_device(module, device) if check_device: if state == 'defined' and device_state == 'Defined': result['changed'] = False result['msg'] = 'Device %s already in Defined' % device else: result['changed'], result['msg'] = remove_device(module, device, force, recursive, state) else: result['msg'] = "Device %s does not exist." % device else: result['msg'] = "Unexpected state %s." % state module.fail_json(**result) module.exit_json(**result) if __name__ == '__main__': main()
washort/gelato.models
refs/heads/master
gelato/__init__.py
56
try: import pkg_resources pkg_resources.declare_namespace(__name__) except ImportError: import pkgutil __path__ = pkgutil.extend_path(__path__, __name__)
home-assistant/home-assistant
refs/heads/dev
homeassistant/components/lifx/const.py
28
"""Const for LIFX.""" DOMAIN = "lifx"
paynejacob/points-tracker
refs/heads/master
migrations/versions/496d97235057_.py
2
"""empty message Revision ID: 496d97235057 Revises: ea80674a1e2 Create Date: 2015-07-15 22:29:11.424610 """ # revision identifiers, used by Alembic. revision = '496d97235057' down_revision = 'ea80674a1e2' from alembic import op import sqlalchemy as sa def upgrade(): ### commands auto generated by Alembic - please adjust! ### op.alter_column('audio_tags', 'audio', new_column_name="audio_id") # op.drop_constraint(u'audio_tags_audio_fkey', 'audio_tags', type_='foreignkey') # op.create_foreign_key(None, 'audio_tags', 'audio', ['audio_id'], ['id']) # op.drop_column('audio_tags', 'audio') ### end Alembic commands ### def downgrade(): ### commands auto generated by Alembic - please adjust! ### op.alter_column('audio_tags', 'audio_id', new_column_name="audio") ### end Alembic commands ###
hoatle/odoo
refs/heads/8.0
openerp/addons/base/res/res_config.py
243
# -*- coding: utf-8 -*- ############################################################################## # # OpenERP, Open Source Management Solution # Copyright (C) 2004-2009 Tiny SPRL (<http://tiny.be>). # # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU Affero General Public License as # published by the Free Software Foundation, either version 3 of the # License, or (at your option) any later version. # # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU Affero General Public License for more details. # # You should have received a copy of the GNU Affero General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. # ############################################################################## import logging from operator import attrgetter import re import openerp from openerp import SUPERUSER_ID from openerp.osv import osv, fields from openerp.tools import ustr from openerp.tools.translate import _ from openerp import exceptions from lxml import etree _logger = logging.getLogger(__name__) class res_config_module_installation_mixin(object): def _install_modules(self, cr, uid, modules, context): """Install the requested modules. return the next action to execute modules is a list of tuples (mod_name, browse_record | None) """ ir_module = self.pool.get('ir.module.module') to_install_ids = [] to_install_missing_names = [] for name, module in modules: if not module: to_install_missing_names.append(name) elif module.state == 'uninstalled': to_install_ids.append(module.id) result = None if to_install_ids: result = ir_module.button_immediate_install(cr, uid, to_install_ids, context=context) #FIXME: if result is not none, the corresponding todo will be skipped because it was just marked done if to_install_missing_names: return { 'type': 'ir.actions.client', 'tag': 'apps', 'params': {'modules': to_install_missing_names}, } return result class res_config_configurable(osv.osv_memory): ''' Base classes for new-style configuration items Configuration items should inherit from this class, implement the execute method (and optionally the cancel one) and have their view inherit from the related res_config_view_base view. ''' _name = 'res.config' def _next_action(self, cr, uid, context=None): Todos = self.pool['ir.actions.todo'] _logger.info('getting next %s', Todos) active_todos = Todos.browse(cr, uid, Todos.search(cr, uid, ['&', ('type', '=', 'automatic'), ('state','=','open')]), context=context) user_groups = set(map( lambda g: g.id, self.pool['res.users'].browse(cr, uid, [uid], context=context)[0].groups_id)) valid_todos_for_user = [ todo for todo in active_todos if not todo.groups_id or bool(user_groups.intersection(( group.id for group in todo.groups_id))) ] if valid_todos_for_user: return valid_todos_for_user[0] return None def _next(self, cr, uid, context=None): _logger.info('getting next operation') next = self._next_action(cr, uid, context=context) _logger.info('next action is %s', next) if next: res = next.action_launch(context=context) res['nodestroy'] = False return res return { 'type': 'ir.actions.client', 'tag': 'reload', } def start(self, cr, uid, ids, context=None): return self.next(cr, uid, ids, context) def next(self, cr, uid, ids, context=None): """ Returns the next todo action to execute (using the default sort order) """ return self._next(cr, uid, context=context) def execute(self, cr, uid, ids, context=None): """ Method called when the user clicks on the ``Next`` button. Execute *must* be overloaded unless ``action_next`` is overloaded (which is something you generally don't need to do). If ``execute`` returns an action dictionary, that action is executed rather than just going to the next configuration item. """ raise NotImplementedError( 'Configuration items need to implement execute') def cancel(self, cr, uid, ids, context=None): """ Method called when the user click on the ``Skip`` button. ``cancel`` should be overloaded instead of ``action_skip``. As with ``execute``, if it returns an action dictionary that action is executed in stead of the default (going to the next configuration item) The default implementation is a NOOP. ``cancel`` is also called by the default implementation of ``action_cancel``. """ pass def action_next(self, cr, uid, ids, context=None): """ Action handler for the ``next`` event. Sets the status of the todo the event was sent from to ``done``, calls ``execute`` and -- unless ``execute`` returned an action dictionary -- executes the action provided by calling ``next``. """ next = self.execute(cr, uid, ids, context=context) if next: return next return self.next(cr, uid, ids, context=context) def action_skip(self, cr, uid, ids, context=None): """ Action handler for the ``skip`` event. Sets the status of the todo the event was sent from to ``skip``, calls ``cancel`` and -- unless ``cancel`` returned an action dictionary -- executes the action provided by calling ``next``. """ next = self.cancel(cr, uid, ids, context=context) if next: return next return self.next(cr, uid, ids, context=context) def action_cancel(self, cr, uid, ids, context=None): """ Action handler for the ``cancel`` event. That event isn't generated by the res.config.view.base inheritable view, the inherited view has to overload one of the buttons (or add one more). Sets the status of the todo the event was sent from to ``cancel``, calls ``cancel`` and -- unless ``cancel`` returned an action dictionary -- executes the action provided by calling ``next``. """ next = self.cancel(cr, uid, ids, context=context) if next: return next return self.next(cr, uid, ids, context=context) class res_config_installer(osv.osv_memory, res_config_module_installation_mixin): """ New-style configuration base specialized for addons selection and installation. Basic usage ----------- Subclasses can simply define a number of _columns as fields.boolean objects. The keys (column names) should be the names of the addons to install (when selected). Upon action execution, selected boolean fields (and those only) will be interpreted as addons to install, and batch-installed. Additional addons ----------------- It is also possible to require the installation of an additional addon set when a specific preset of addons has been marked for installation (in the basic usage only, additionals can't depend on one another). These additionals are defined through the ``_install_if`` property. This property is a mapping of a collection of addons (by name) to a collection of addons (by name) [#]_, and if all the *key* addons are selected for installation, then the *value* ones will be selected as well. For example:: _install_if = { ('sale','crm'): ['sale_crm'], } This will install the ``sale_crm`` addon if and only if both the ``sale`` and ``crm`` addons are selected for installation. You can define as many additionals as you wish, and additionals can overlap in key and value. For instance:: _install_if = { ('sale','crm'): ['sale_crm'], ('sale','project'): ['sale_service'], } will install both ``sale_crm`` and ``sale_service`` if all of ``sale``, ``crm`` and ``project`` are selected for installation. Hook methods ------------ Subclasses might also need to express dependencies more complex than that provided by additionals. In this case, it's possible to define methods of the form ``_if_%(name)s`` where ``name`` is the name of a boolean field. If the field is selected, then the corresponding module will be marked for installation *and* the hook method will be executed. Hook methods take the usual set of parameters (cr, uid, ids, context) and can return a collection of additional addons to install (if they return anything, otherwise they should not return anything, though returning any "falsy" value such as None or an empty collection will have the same effect). Complete control ---------------- The last hook is to simply overload the ``modules_to_install`` method, which implements all the mechanisms above. This method takes the usual set of parameters (cr, uid, ids, context) and returns a ``set`` of addons to install (addons selected by the above methods minus addons from the *basic* set which are already installed) [#]_ so an overloader can simply manipulate the ``set`` returned by ``res_config_installer.modules_to_install`` to add or remove addons. Skipping the installer ---------------------- Unless it is removed from the view, installers have a *skip* button which invokes ``action_skip`` (and the ``cancel`` hook from ``res.config``). Hooks and additionals *are not run* when skipping installation, even for already installed addons. Again, setup your hooks accordingly. .. [#] note that since a mapping key needs to be hashable, it's possible to use a tuple or a frozenset, but not a list or a regular set .. [#] because the already-installed modules are only pruned at the very end of ``modules_to_install``, additionals and hooks depending on them *are guaranteed to execute*. Setup your hooks accordingly. """ _name = 'res.config.installer' _inherit = 'res.config' _install_if = {} def already_installed(self, cr, uid, context=None): """ For each module, check if it's already installed and if it is return its name :returns: a list of the already installed modules in this installer :rtype: [str] """ return map(attrgetter('name'), self._already_installed(cr, uid, context=context)) def _already_installed(self, cr, uid, context=None): """ For each module (boolean fields in a res.config.installer), check if it's already installed (either 'to install', 'to upgrade' or 'installed') and if it is return the module's record :returns: a list of all installed modules in this installer :rtype: recordset (collection of Record) """ modules = self.pool['ir.module.module'] selectable = [field for field in self._columns if type(self._columns[field]) is fields.boolean] return modules.browse( cr, uid, modules.search(cr, uid, [('name','in',selectable), ('state','in',['to install', 'installed', 'to upgrade'])], context=context), context=context) def modules_to_install(self, cr, uid, ids, context=None): """ selects all modules to install: * checked boolean fields * return values of hook methods. Hook methods are of the form ``_if_%(addon_name)s``, and are called if the corresponding addon is marked for installation. They take the arguments cr, uid, ids and context, and return an iterable of addon names * additionals, additionals are setup through the ``_install_if`` class variable. ``_install_if`` is a dict of {iterable:iterable} where key and value are iterables of addon names. If all the addons in the key are selected for installation (warning: addons added through hooks don't count), then the addons in the value are added to the set of modules to install * not already installed """ base = set(module_name for installer in self.read(cr, uid, ids, context=context) for module_name, to_install in installer.iteritems() if module_name != 'id' if type(self._columns.get(module_name)) is fields.boolean if to_install) hooks_results = set() for module in base: hook = getattr(self, '_if_%s'% module, None) if hook: hooks_results.update(hook(cr, uid, ids, context=None) or set()) additionals = set( module for requirements, consequences \ in self._install_if.iteritems() if base.issuperset(requirements) for module in consequences) return (base | hooks_results | additionals).difference( self.already_installed(cr, uid, context)) def default_get(self, cr, uid, fields_list, context=None): ''' If an addon is already installed, check it by default ''' defaults = super(res_config_installer, self).default_get( cr, uid, fields_list, context=context) return dict(defaults, **dict.fromkeys( self.already_installed(cr, uid, context=context), True)) def fields_get(self, cr, uid, fields=None, context=None, write_access=True, attributes=None): """ If an addon is already installed, set it to readonly as res.config.installer doesn't handle uninstallations of already installed addons """ fields = super(res_config_installer, self).fields_get( cr, uid, fields, context, write_access, attributes) for name in self.already_installed(cr, uid, context=context): if name not in fields: continue fields[name].update( readonly=True, help= ustr(fields[name].get('help', '')) + _('\n\nThis addon is already installed on your system')) return fields def execute(self, cr, uid, ids, context=None): to_install = list(self.modules_to_install( cr, uid, ids, context=context)) _logger.info('Selecting addons %s to install', to_install) ir_module = self.pool.get('ir.module.module') modules = [] for name in to_install: mod_ids = ir_module.search(cr, uid, [('name', '=', name)]) record = ir_module.browse(cr, uid, mod_ids[0], context) if mod_ids else None modules.append((name, record)) return self._install_modules(cr, uid, modules, context=context) class res_config_settings(osv.osv_memory, res_config_module_installation_mixin): """ Base configuration wizard for application settings. It provides support for setting default values, assigning groups to employee users, and installing modules. To make such a 'settings' wizard, define a model like:: class my_config_wizard(osv.osv_memory): _name = 'my.settings' _inherit = 'res.config.settings' _columns = { 'default_foo': fields.type(..., default_model='my.model'), 'group_bar': fields.boolean(..., group='base.group_user', implied_group='my.group'), 'module_baz': fields.boolean(...), 'other_field': fields.type(...), } The method ``execute`` provides some support based on a naming convention: * For a field like 'default_XXX', ``execute`` sets the (global) default value of the field 'XXX' in the model named by ``default_model`` to the field's value. * For a boolean field like 'group_XXX', ``execute`` adds/removes 'implied_group' to/from the implied groups of 'group', depending on the field's value. By default 'group' is the group Employee. Groups are given by their xml id. The attribute 'group' may contain several xml ids, separated by commas. * For a boolean field like 'module_XXX', ``execute`` triggers the immediate installation of the module named 'XXX' if the field has value ``True``. * For the other fields, the method ``execute`` invokes all methods with a name that starts with 'set_'; such methods can be defined to implement the effect of those fields. The method ``default_get`` retrieves values that reflect the current status of the fields like 'default_XXX', 'group_XXX' and 'module_XXX'. It also invokes all methods with a name that starts with 'get_default_'; such methods can be defined to provide current values for other fields. """ _name = 'res.config.settings' def copy(self, cr, uid, id, values, context=None): raise osv.except_osv(_("Cannot duplicate configuration!"), "") def fields_view_get(self, cr, user, view_id=None, view_type='form', context=None, toolbar=False, submenu=False): ret_val = super(res_config_settings, self).fields_view_get( cr, user, view_id=view_id, view_type=view_type, context=context, toolbar=toolbar, submenu=submenu) doc = etree.XML(ret_val['arch']) for field in ret_val['fields']: if not field.startswith("module_"): continue for node in doc.xpath("//field[@name='%s']" % field): if 'on_change' not in node.attrib: node.set("on_change", "onchange_module(%s, '%s')" % (field, field)) ret_val['arch'] = etree.tostring(doc) return ret_val def onchange_module(self, cr, uid, ids, field_value, module_name, context={}): module_pool = self.pool.get('ir.module.module') module_ids = module_pool.search( cr, uid, [('name', '=', module_name.replace("module_", '')), ('state','in', ['to install', 'installed', 'to upgrade'])], context=context) if module_ids and not field_value: dep_ids = module_pool.downstream_dependencies(cr, uid, module_ids, context=context) dep_name = [x.shortdesc for x in module_pool.browse( cr, uid, dep_ids + module_ids, context=context)] message = '\n'.join(dep_name) return { 'warning': { 'title': _('Warning!'), 'message': _('Disabling this option will also uninstall the following modules \n%s') % message, } } return {} def _get_classified_fields(self, cr, uid, context=None): """ return a dictionary with the fields classified by category:: { 'default': [('default_foo', 'model', 'foo'), ...], 'group': [('group_bar', [browse_group], browse_implied_group), ...], 'module': [('module_baz', browse_module), ...], 'other': ['other_field', ...], } """ ir_model_data = self.pool['ir.model.data'] ir_module = self.pool['ir.module.module'] def ref(xml_id): mod, xml = xml_id.split('.', 1) return ir_model_data.get_object(cr, uid, mod, xml, context=context) defaults, groups, modules, others = [], [], [], [] for name, field in self._columns.items(): if name.startswith('default_') and hasattr(field, 'default_model'): defaults.append((name, field.default_model, name[8:])) elif name.startswith('group_') and isinstance(field, fields.boolean) and hasattr(field, 'implied_group'): field_groups = getattr(field, 'group', 'base.group_user').split(',') groups.append((name, map(ref, field_groups), ref(field.implied_group))) elif name.startswith('module_') and isinstance(field, fields.boolean): mod_ids = ir_module.search(cr, uid, [('name', '=', name[7:])]) record = ir_module.browse(cr, uid, mod_ids[0], context) if mod_ids else None modules.append((name, record)) else: others.append(name) return {'default': defaults, 'group': groups, 'module': modules, 'other': others} def default_get(self, cr, uid, fields, context=None): ir_values = self.pool['ir.values'] classified = self._get_classified_fields(cr, uid, context) res = super(res_config_settings, self).default_get(cr, uid, fields, context) # defaults: take the corresponding default value they set for name, model, field in classified['default']: value = ir_values.get_default(cr, uid, model, field) if value is not None: res[name] = value # groups: which groups are implied by the group Employee for name, groups, implied_group in classified['group']: res[name] = all(implied_group in group.implied_ids for group in groups) # modules: which modules are installed/to install for name, module in classified['module']: res[name] = module and module.state in ('installed', 'to install', 'to upgrade') # other fields: call all methods that start with 'get_default_' for method in dir(self): if method.startswith('get_default_'): res.update(getattr(self, method)(cr, uid, fields, context)) return res def execute(self, cr, uid, ids, context=None): if context is None: context = {} context = dict(context, active_test=False) if uid != SUPERUSER_ID and not self.pool['res.users'].has_group(cr, uid, 'base.group_erp_manager'): raise openerp.exceptions.AccessError(_("Only administrators can change the settings")) ir_values = self.pool['ir.values'] ir_module = self.pool['ir.module.module'] res_groups = self.pool['res.groups'] classified = self._get_classified_fields(cr, uid, context=context) config = self.browse(cr, uid, ids[0], context) # default values fields for name, model, field in classified['default']: ir_values.set_default(cr, SUPERUSER_ID, model, field, config[name]) # group fields: modify group / implied groups for name, groups, implied_group in classified['group']: gids = map(int, groups) if config[name]: res_groups.write(cr, uid, gids, {'implied_ids': [(4, implied_group.id)]}, context=context) else: res_groups.write(cr, uid, gids, {'implied_ids': [(3, implied_group.id)]}, context=context) uids = set() for group in groups: uids.update(map(int, group.users)) implied_group.write({'users': [(3, u) for u in uids]}) # other fields: execute all methods that start with 'set_' for method in dir(self): if method.startswith('set_'): getattr(self, method)(cr, uid, ids, context) # module fields: install/uninstall the selected modules to_install = [] to_uninstall_ids = [] lm = len('module_') for name, module in classified['module']: if config[name]: to_install.append((name[lm:], module)) else: if module and module.state in ('installed', 'to upgrade'): to_uninstall_ids.append(module.id) if to_uninstall_ids: ir_module.button_immediate_uninstall(cr, uid, to_uninstall_ids, context=context) action = self._install_modules(cr, uid, to_install, context=context) if action: return action # After the uninstall/install calls, the self.pool is no longer valid. # So we reach into the RegistryManager directly. res_config = openerp.modules.registry.RegistryManager.get(cr.dbname)['res.config'] config = res_config.next(cr, uid, [], context=context) or {} if config.get('type') not in ('ir.actions.act_window_close',): return config # force client-side reload (update user menu and current view) return { 'type': 'ir.actions.client', 'tag': 'reload', } def cancel(self, cr, uid, ids, context=None): # ignore the current record, and send the action to reopen the view act_window = self.pool['ir.actions.act_window'] action_ids = act_window.search(cr, uid, [('res_model', '=', self._name)]) if action_ids: return act_window.read(cr, uid, action_ids[0], [], context=context) return {} def name_get(self, cr, uid, ids, context=None): """ Override name_get method to return an appropriate configuration wizard name, and not the generated name.""" if not ids: return [] # name_get may receive int id instead of an id list if isinstance(ids, (int, long)): ids = [ids] act_window = self.pool['ir.actions.act_window'] action_ids = act_window.search(cr, uid, [('res_model', '=', self._name)], context=context) name = self._name if action_ids: name = act_window.read(cr, uid, action_ids[0], ['name'], context=context)['name'] return [(record.id, name) for record in self.browse(cr, uid , ids, context=context)] def get_option_path(self, cr, uid, menu_xml_id, context=None): """ Fetch the path to a specified configuration view and the action id to access it. :param string menu_xml_id: the xml id of the menuitem where the view is located, structured as follows: module_name.menuitem_xml_id (e.g.: "base.menu_sale_config") :return tuple: - t[0]: string: full path to the menuitem (e.g.: "Settings/Configuration/Sales") - t[1]: int or long: id of the menuitem's action """ module_name, menu_xml_id = menu_xml_id.split('.') dummy, menu_id = self.pool['ir.model.data'].get_object_reference(cr, uid, module_name, menu_xml_id) ir_ui_menu = self.pool['ir.ui.menu'].browse(cr, uid, menu_id, context=context) return (ir_ui_menu.complete_name, ir_ui_menu.action.id) def get_option_name(self, cr, uid, full_field_name, context=None): """ Fetch the human readable name of a specified configuration option. :param string full_field_name: the full name of the field, structured as follows: model_name.field_name (e.g.: "sale.config.settings.fetchmail_lead") :return string: human readable name of the field (e.g.: "Create leads from incoming mails") """ model_name, field_name = full_field_name.rsplit('.', 1) return self.pool[model_name].fields_get(cr, uid, allfields=[field_name], context=context)[field_name]['string'] def get_config_warning(self, cr, msg, context=None): """ Helper: return a Warning exception with the given message where the %(field:xxx)s and/or %(menu:yyy)s are replaced by the human readable field's name and/or menuitem's full path. Usage: ------ Just include in your error message %(field:model_name.field_name)s to obtain the human readable field's name, and/or %(menu:module_name.menuitem_xml_id)s to obtain the menuitem's full path. Example of use: --------------- from openerp.addons.base.res.res_config import get_warning_config raise get_warning_config(cr, _("Error: this action is prohibited. You should check the field %(field:sale.config.settings.fetchmail_lead)s in %(menu:base.menu_sale_config)s."), context=context) This will return an exception containing the following message: Error: this action is prohibited. You should check the field Create leads from incoming mails in Settings/Configuration/Sales. What if there is another substitution in the message already? ------------------------------------------------------------- You could have a situation where the error message you want to upgrade already contains a substitution. Example: Cannot find any account journal of %s type for this company.\n\nYou can create one in the menu: \nConfiguration\Journals\Journals. What you want to do here is simply to replace the path by %menu:account.menu_account_config)s, and leave the rest alone. In order to do that, you can use the double percent (%%) to escape your new substitution, like so: Cannot find any account journal of %s type for this company.\n\nYou can create one in the %%(menu:account.menu_account_config)s. """ res_config_obj = openerp.registry(cr.dbname)['res.config.settings'] regex_path = r'%\(((?:menu|field):[a-z_\.]*)\)s' # Process the message # 1/ find the menu and/or field references, put them in a list references = re.findall(regex_path, msg, flags=re.I) # 2/ fetch the menu and/or field replacement values (full path and # human readable field's name) and the action_id if any values = {} action_id = None for item in references: ref_type, ref = item.split(':') if ref_type == 'menu': values[item], action_id = res_config_obj.get_option_path(cr, SUPERUSER_ID, ref, context=context) elif ref_type == 'field': values[item] = res_config_obj.get_option_name(cr, SUPERUSER_ID, ref, context=context) # 3/ substitute and return the result if (action_id): return exceptions.RedirectWarning(msg % values, action_id, _('Go to the configuration panel')) return exceptions.Warning(msg % values) # vim:expandtab:smartindent:tabstop=4:softtabstop=4:shiftwidth=4:
kubeflow/pipelines
refs/heads/master
backend/api/python_http_client/kfp_server_api/models/api_run.py
2
# coding: utf-8 """ Kubeflow Pipelines API This file contains REST API specification for Kubeflow Pipelines. The file is autogenerated from the swagger definition. Contact: kubeflow-pipelines@google.com Generated by: https://openapi-generator.tech """ import pprint import re # noqa: F401 import six from kfp_server_api.configuration import Configuration class ApiRun(object): """NOTE: This class is auto generated by OpenAPI Generator. Ref: https://openapi-generator.tech Do not edit the class manually. """ """ Attributes: openapi_types (dict): The key is attribute name and the value is attribute type. attribute_map (dict): The key is attribute name and the value is json key in definition. """ openapi_types = { 'id': 'str', 'name': 'str', 'storage_state': 'ApiRunStorageState', 'description': 'str', 'pipeline_spec': 'ApiPipelineSpec', 'resource_references': 'list[ApiResourceReference]', 'service_account': 'str', 'created_at': 'datetime', 'scheduled_at': 'datetime', 'finished_at': 'datetime', 'status': 'str', 'error': 'str', 'metrics': 'list[ApiRunMetric]' } attribute_map = { 'id': 'id', 'name': 'name', 'storage_state': 'storage_state', 'description': 'description', 'pipeline_spec': 'pipeline_spec', 'resource_references': 'resource_references', 'service_account': 'service_account', 'created_at': 'created_at', 'scheduled_at': 'scheduled_at', 'finished_at': 'finished_at', 'status': 'status', 'error': 'error', 'metrics': 'metrics' } def __init__(self, id=None, name=None, storage_state=None, description=None, pipeline_spec=None, resource_references=None, service_account=None, created_at=None, scheduled_at=None, finished_at=None, status=None, error=None, metrics=None, local_vars_configuration=None): # noqa: E501 """ApiRun - a model defined in OpenAPI""" # noqa: E501 if local_vars_configuration is None: local_vars_configuration = Configuration() self.local_vars_configuration = local_vars_configuration self._id = None self._name = None self._storage_state = None self._description = None self._pipeline_spec = None self._resource_references = None self._service_account = None self._created_at = None self._scheduled_at = None self._finished_at = None self._status = None self._error = None self._metrics = None self.discriminator = None if id is not None: self.id = id if name is not None: self.name = name if storage_state is not None: self.storage_state = storage_state if description is not None: self.description = description if pipeline_spec is not None: self.pipeline_spec = pipeline_spec if resource_references is not None: self.resource_references = resource_references if service_account is not None: self.service_account = service_account if created_at is not None: self.created_at = created_at if scheduled_at is not None: self.scheduled_at = scheduled_at if finished_at is not None: self.finished_at = finished_at if status is not None: self.status = status if error is not None: self.error = error if metrics is not None: self.metrics = metrics @property def id(self): """Gets the id of this ApiRun. # noqa: E501 Output. Unique run ID. Generated by API server. # noqa: E501 :return: The id of this ApiRun. # noqa: E501 :rtype: str """ return self._id @id.setter def id(self, id): """Sets the id of this ApiRun. Output. Unique run ID. Generated by API server. # noqa: E501 :param id: The id of this ApiRun. # noqa: E501 :type id: str """ self._id = id @property def name(self): """Gets the name of this ApiRun. # noqa: E501 Required input field. Name provided by user, or auto generated if run is created by scheduled job. Not unique. # noqa: E501 :return: The name of this ApiRun. # noqa: E501 :rtype: str """ return self._name @name.setter def name(self, name): """Sets the name of this ApiRun. Required input field. Name provided by user, or auto generated if run is created by scheduled job. Not unique. # noqa: E501 :param name: The name of this ApiRun. # noqa: E501 :type name: str """ self._name = name @property def storage_state(self): """Gets the storage_state of this ApiRun. # noqa: E501 :return: The storage_state of this ApiRun. # noqa: E501 :rtype: ApiRunStorageState """ return self._storage_state @storage_state.setter def storage_state(self, storage_state): """Sets the storage_state of this ApiRun. :param storage_state: The storage_state of this ApiRun. # noqa: E501 :type storage_state: ApiRunStorageState """ self._storage_state = storage_state @property def description(self): """Gets the description of this ApiRun. # noqa: E501 :return: The description of this ApiRun. # noqa: E501 :rtype: str """ return self._description @description.setter def description(self, description): """Sets the description of this ApiRun. :param description: The description of this ApiRun. # noqa: E501 :type description: str """ self._description = description @property def pipeline_spec(self): """Gets the pipeline_spec of this ApiRun. # noqa: E501 :return: The pipeline_spec of this ApiRun. # noqa: E501 :rtype: ApiPipelineSpec """ return self._pipeline_spec @pipeline_spec.setter def pipeline_spec(self, pipeline_spec): """Sets the pipeline_spec of this ApiRun. :param pipeline_spec: The pipeline_spec of this ApiRun. # noqa: E501 :type pipeline_spec: ApiPipelineSpec """ self._pipeline_spec = pipeline_spec @property def resource_references(self): """Gets the resource_references of this ApiRun. # noqa: E501 Optional input field. Specify which resource this run belongs to. When creating a run from a particular pipeline version, the pipeline version can be specified here. # noqa: E501 :return: The resource_references of this ApiRun. # noqa: E501 :rtype: list[ApiResourceReference] """ return self._resource_references @resource_references.setter def resource_references(self, resource_references): """Sets the resource_references of this ApiRun. Optional input field. Specify which resource this run belongs to. When creating a run from a particular pipeline version, the pipeline version can be specified here. # noqa: E501 :param resource_references: The resource_references of this ApiRun. # noqa: E501 :type resource_references: list[ApiResourceReference] """ self._resource_references = resource_references @property def service_account(self): """Gets the service_account of this ApiRun. # noqa: E501 Optional input field. Specify which Kubernetes service account this run uses. # noqa: E501 :return: The service_account of this ApiRun. # noqa: E501 :rtype: str """ return self._service_account @service_account.setter def service_account(self, service_account): """Sets the service_account of this ApiRun. Optional input field. Specify which Kubernetes service account this run uses. # noqa: E501 :param service_account: The service_account of this ApiRun. # noqa: E501 :type service_account: str """ self._service_account = service_account @property def created_at(self): """Gets the created_at of this ApiRun. # noqa: E501 Output. The time that the run created. # noqa: E501 :return: The created_at of this ApiRun. # noqa: E501 :rtype: datetime """ return self._created_at @created_at.setter def created_at(self, created_at): """Sets the created_at of this ApiRun. Output. The time that the run created. # noqa: E501 :param created_at: The created_at of this ApiRun. # noqa: E501 :type created_at: datetime """ self._created_at = created_at @property def scheduled_at(self): """Gets the scheduled_at of this ApiRun. # noqa: E501 Output. When this run is scheduled to run. This could be different from created_at. For example, if a run is from a backfilling job that was supposed to run 2 month ago, the scheduled_at is 2 month ago, v.s. created_at is the current time. # noqa: E501 :return: The scheduled_at of this ApiRun. # noqa: E501 :rtype: datetime """ return self._scheduled_at @scheduled_at.setter def scheduled_at(self, scheduled_at): """Sets the scheduled_at of this ApiRun. Output. When this run is scheduled to run. This could be different from created_at. For example, if a run is from a backfilling job that was supposed to run 2 month ago, the scheduled_at is 2 month ago, v.s. created_at is the current time. # noqa: E501 :param scheduled_at: The scheduled_at of this ApiRun. # noqa: E501 :type scheduled_at: datetime """ self._scheduled_at = scheduled_at @property def finished_at(self): """Gets the finished_at of this ApiRun. # noqa: E501 Output. The time this run is finished. # noqa: E501 :return: The finished_at of this ApiRun. # noqa: E501 :rtype: datetime """ return self._finished_at @finished_at.setter def finished_at(self, finished_at): """Sets the finished_at of this ApiRun. Output. The time this run is finished. # noqa: E501 :param finished_at: The finished_at of this ApiRun. # noqa: E501 :type finished_at: datetime """ self._finished_at = finished_at @property def status(self): """Gets the status of this ApiRun. # noqa: E501 :return: The status of this ApiRun. # noqa: E501 :rtype: str """ return self._status @status.setter def status(self, status): """Sets the status of this ApiRun. :param status: The status of this ApiRun. # noqa: E501 :type status: str """ self._status = status @property def error(self): """Gets the error of this ApiRun. # noqa: E501 In case any error happens retrieving a run field, only run ID and the error message is returned. Client has the flexibility of choosing how to handle error. This is especially useful during listing call. # noqa: E501 :return: The error of this ApiRun. # noqa: E501 :rtype: str """ return self._error @error.setter def error(self, error): """Sets the error of this ApiRun. In case any error happens retrieving a run field, only run ID and the error message is returned. Client has the flexibility of choosing how to handle error. This is especially useful during listing call. # noqa: E501 :param error: The error of this ApiRun. # noqa: E501 :type error: str """ self._error = error @property def metrics(self): """Gets the metrics of this ApiRun. # noqa: E501 Output. The metrics of the run. The metrics are reported by ReportMetrics API. # noqa: E501 :return: The metrics of this ApiRun. # noqa: E501 :rtype: list[ApiRunMetric] """ return self._metrics @metrics.setter def metrics(self, metrics): """Sets the metrics of this ApiRun. Output. The metrics of the run. The metrics are reported by ReportMetrics API. # noqa: E501 :param metrics: The metrics of this ApiRun. # noqa: E501 :type metrics: list[ApiRunMetric] """ self._metrics = metrics def to_dict(self): """Returns the model properties as a dict""" result = {} for attr, _ in six.iteritems(self.openapi_types): value = getattr(self, attr) if isinstance(value, list): result[attr] = list(map( lambda x: x.to_dict() if hasattr(x, "to_dict") else x, value )) elif hasattr(value, "to_dict"): result[attr] = value.to_dict() elif isinstance(value, dict): result[attr] = dict(map( lambda item: (item[0], item[1].to_dict()) if hasattr(item[1], "to_dict") else item, value.items() )) else: result[attr] = value return result def to_str(self): """Returns the string representation of the model""" return pprint.pformat(self.to_dict()) def __repr__(self): """For `print` and `pprint`""" return self.to_str() def __eq__(self, other): """Returns true if both objects are equal""" if not isinstance(other, ApiRun): return False return self.to_dict() == other.to_dict() def __ne__(self, other): """Returns true if both objects are not equal""" if not isinstance(other, ApiRun): return True return self.to_dict() != other.to_dict()
siosio/intellij-community
refs/heads/master
python/testData/formatter/alignInGenerators.py
83
def supprice(): if True: if True: agdrn = sum(VARS[drn + price] * md.c[drn][1] * md.c[drn][3] * exp(md.c[drn][2] * VARS['SEEPAGE'] - md.c[drn][3] * pmp) for drn in md.agdrn_nodes if drn in md.c)
ekivemark/BlueButtonFHIR_API
refs/heads/master
fhir/views/create.py
1
import sys from django.shortcuts import render from fhir.models import SupportedResourceType from collections import OrderedDict from django.http import HttpResponse from django.views.decorators.csrf import csrf_exempt import json, uuid from jsonschema import validate, ValidationError import datetime from fhir.settings import DF_EXTRA_INFO from fhir.utils import (kickout_404, kickout_403, kickout_400, kickout_500) from fhir.views.hello import hello from fhir.views.search import search @csrf_exempt def create(request, resource_type): """Create FHIR Interaction""" # Example client use in curl: # curl -H "Content-Type: application/json" --data @test.json http://127.0.0.1:8000/fhir/Practitioner interaction_type = 'create' # re-route to hello if no resource type is given: if not resource_type: return hello(request) try: rt = SupportedResourceType.objects.get(resource_name=resource_type) if interaction_type not in rt.get_supported_interaction_types() and request.method == "GET": # GET means that this is a search so re-route return search(request, resource_type) elif interaction_type not in rt.get_supported_interaction_types() : msg = "The interaction %s is not permitted on %s FHIR resources on this FHIR sever." % (interaction_type, resource_type) return kickout_403(msg) except SupportedResourceType.DoesNotExist: msg = "%s is not a supported resource type on this FHIR server." % (resource_type) return kickout_404(msg) # Catch all for GETs to re-direct to search if CREATE permission is valid if request.method == "GET": return search(request, resource_type) if request.method == 'POST': # Check if request body is JSON ------------------------ try: j =json.loads(request.body, object_pairs_hook=OrderedDict) if type(j) != type({}): kickout_400("The request body did not contain a JSON object i.e. {}.") except: return kickout_400("The request body did not contain valid JSON.") if j.has_key('id'): return kickout_400("Create cannot have an id. Perhaps you meant to perform an update?") # check json_schema is valid try: json_schema = json.loads(rt.json_schema, object_pairs_hook=OrderedDict) except: return kickout_500("The JSON Schema on the server did not contain valid JSON.") # Check jsonschema if json_schema: try: validate(j, json_schema) except ValidationError: msg = "JSON Schema Conformance Error. %s" % (str(sys.exc_info()[1][0])) return kickout_400(msg) # write_to_mongo - TBD response = OrderedDict() response['id'] = str(uuid.uuid4()) meta = OrderedDict() if j.get('meta').get('versionId'): meta['versionId'] = j.get('meta').get('versionId') else: meta['versionId'] = 1 if j.get('meta').get('lastUpdated'): meta['lastUpdated'] = j.get('meta').get('lastUpdated') else: meta['lastUpdated'] = "%sZ" % (datetime.datetime.utcnow().isoformat()) meta['id'] = response['id'] response['meta'] = meta hr = HttpResponse(json.dumps(response, indent=4), status=201, content_type="application/json") hr['Location'] = "%s/%s/%s/_history/%s" % ( "http://127.0.0.1:8000/fhir", resource_type, meta['id'], meta['versionId']) return hr #This is something other than GET or POST (i.e. a GET) if request.method not in ("GET", "POST"): od = OrderedDict() if DF_EXTRA_INFO: od['request_method']= request.method od['interaction_type'] = "create" od['resource_type'] = resource_type if DF_EXTRA_INFO: od['note'] = "Perform an HTTP POST to this URL with the JSON resource as the request body." return HttpResponse(json.dumps(od, indent=4), content_type="application/json")
HatsuneMiku0309/2014cdag10
refs/heads/master
wsgi/static/Brython2.1.0-20140419-113919/Lib/unittest/__main__.py
737
"""Main entry point""" import sys if sys.argv[0].endswith("__main__.py"): import os.path # We change sys.argv[0] to make help message more useful # use executable without path, unquoted # (it's just a hint anyway) # (if you have spaces in your executable you get what you deserve!) executable = os.path.basename(sys.executable) sys.argv[0] = executable + " -m unittest" del os __unittest = True from .main import main, TestProgram, USAGE_AS_MAIN TestProgram.USAGE = USAGE_AS_MAIN main(module=None)
patrickdw123/ParanoiDF
refs/heads/master
makeEmbedded.py
1
# ParanoiDF. A combination of several PDF analysis/manipulation tools to # produce one of the most technically useful PDF analysis tools. # # Idea proposed by Julio Hernandez-Castro, University of Kent, UK. # By Patrick Wragg # University of Kent # 21/07/2014 # # With thanks to: # Julio Hernandez-Castro, my supervisor. # Jose Miguel Esparza for writing PeePDF (the basis of this tool). # Didier Stevens for his "make-PDF" tools. # Blake Hartstein for Jsunpack-n. # Yusuke Shinyama for Pdf2txt.py (PDFMiner) # Nacho Barrientos Arias for Pdfcrack. # Kovid Goyal for Calibre (DRM removal). # Jay Berkenbilt for QPDF. # # Copyright (C) 2014-2018 Patrick Wragg # # This file is part of ParanoiDF. # # ParanoiDF is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # # ParanoiDF is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # # You should have received a copy of the GNU General Public License # along with ParanoiDF. If not, see <http://www.gnu.org/licenses/>. # # This was written by Didier Stevens, but has been modified by Patrick # Wragg to be used inside ParanoiDF Interactive Console, 22/07/2014. __description__ = 'tool to create a PDF document with an embedded file' __author__ = 'Didier Stevens' __version__ = '0.5.0' __date__ = '2011/07/01' """ Source code put in public domain by Didier Stevens, no Copyright https://DidierStevens.com Use at your own risk History: 2008/05/18: V0.1 2008/05/19: Refactoring 2008/05/23: Refactoring 2008/05/27: Refactoring 2008/06/27: V0.2, Refactoring, options, cleanup 2008/11/09: V0.3, added autostart and button 2009/06/15: V0.4.0: added stego 2011/07/01: V0.5.0: added support for Python 3 Todo: """ import mPDF import optparse def ReadBinaryFile(name): """Read a binary file and return the content, return None if error occured """ try: fBinary = open(name, 'rb') except: return None try: content = fBinary.read() except: return None finally: fBinary.close() return content def CreatePDFWithEmbeddedFile(pdfFileName, embeddedFileName, embeddedFileContent, filters, nobinary, autoopen, button, stego, text): """Create a PDF document with an embedded file """ oPDF = mPDF.cPDF(pdfFileName) oPDF.header() if not nobinary: oPDF.binary() if stego: embeddedFiles = 'Embeddedfiles' else: embeddedFiles = 'EmbeddedFiles' if autoopen: openAction = ' /OpenAction 9 0 R\n' else: openAction = '' oPDF.indirectobject(1, 0, '<<\n /Type /Catalog\n /Outlines 2 0 R\n /Pages 3 0 R\n /Names << /%s << /Names [(%s) 7 0 R] >> >>\n%s>>' % (embeddedFiles, embeddedFileName, openAction)) oPDF.indirectobject(2, 0, '<<\n /Type /Outlines\n /Count 0\n>>') oPDF.indirectobject(3, 0, '<<\n /Type /Pages\n /Kids [4 0 R]\n /Count 1\n>>') if button: annots = ' /Annots [10 0 R]\n' else: annots = '' oPDF.indirectobject(4, 0, '<<\n /Type /Page\n /Parent 3 0 R\n /MediaBox [0 0 612 792]\n /Contents 5 0 R\n /Resources <<\n /ProcSet [/PDF /Text]\n /Font << /F1 6 0 R >>\n >>\n%s>>' % annots) if text == '': text = 'This PDF document embeds file %s' % embeddedFileName textCommands = '/F1 12 Tf 70 700 Td 15 TL (%s) Tj' % text if button: textCommands += " () ' () ' (Click inside the rectangle to save %s to a temporary folder and launch it.) ' () ' (Click here) '" % embeddedFileName oPDF.stream(5, 0, 'BT %s ET' % textCommands) oPDF.indirectobject(6, 0, '<<\n /Type /Font\n /Subtype /Type1\n /Name /F1\n /BaseFont /Helvetica\n /Encoding /MacRomanEncoding\n>>') oPDF.indirectobject(7, 0, '<<\n /Type /Filespec\n /F (%s)\n /EF << /F 8 0 R >>\n>>' % embeddedFileName) oPDF.stream2(8, 0, embeddedFileContent, '/Type /EmbeddedFile', filters) if autoopen or button: oPDF.indirectobject(9, 0, '<<\n /Type /Action\n /S /JavaScript\n /JS (this.exportDataObject({ cName: "%s", nLaunch: 2 });)\n>>' % embeddedFileName) if button: oPDF.indirectobject(10, 0, '<<\n /Type /Annot\n /Subtype /Link\n /Rect [65 620 130 640]\n /Border [16 16 1]\n /A 9 0 R\n>>') oPDF.xrefAndTrailer("1 0 R") def main(options, embeddedFileName, pdfFileName): nobinary = False autoopen = False button = False stego = False filters = 'f' message = '' if options != 0: optionList = options.split(",") for option in optionList: if option.lower() == "t": nobinary = True elif option.lower() == "a": autoopen = True elif option.lower() == "b": button = True elif option.lower() == "s": stego = True elif "f=" in option: filters = option[2:] elif "m=" in option: message = option[2:] else: print 'Option "%s" not recognised.' % option embeddedFileContent = ReadBinaryFile(embeddedFileName) if embeddedFileContent == None: print('Error opening/reading file %s' % embeddedFileName) else: CreatePDFWithEmbeddedFile(pdfFileName, embeddedFileName, embeddedFileContent, filters, nobinary, autoopen, button, stego, message) if __name__ == '__main__': Main()
ridfrustum/lettuce
refs/heads/master
tests/integration/lib/Django-1.3/tests/modeltests/fixtures/models.py
51
""" 37. Fixtures. Fixtures are a way of loading data into the database in bulk. Fixure data can be stored in any serializable format (including JSON and XML). Fixtures are identified by name, and are stored in either a directory named 'fixtures' in the application directory, or in one of the directories named in the ``FIXTURE_DIRS`` setting. """ from django.contrib.auth.models import Permission from django.contrib.contenttypes import generic from django.contrib.contenttypes.models import ContentType from django.db import models, DEFAULT_DB_ALIAS from django.conf import settings class Category(models.Model): title = models.CharField(max_length=100) description = models.TextField() def __unicode__(self): return self.title class Meta: ordering = ('title',) class Article(models.Model): headline = models.CharField(max_length=100, default='Default headline') pub_date = models.DateTimeField() def __unicode__(self): return self.headline class Meta: ordering = ('-pub_date', 'headline') class Blog(models.Model): name = models.CharField(max_length=100) featured = models.ForeignKey(Article, related_name='fixtures_featured_set') articles = models.ManyToManyField(Article, blank=True, related_name='fixtures_articles_set') def __unicode__(self): return self.name class Tag(models.Model): name = models.CharField(max_length=100) tagged_type = models.ForeignKey(ContentType, related_name="fixtures_tag_set") tagged_id = models.PositiveIntegerField(default=0) tagged = generic.GenericForeignKey(ct_field='tagged_type', fk_field='tagged_id') def __unicode__(self): return '<%s: %s> tagged "%s"' % (self.tagged.__class__.__name__, self.tagged, self.name) class PersonManager(models.Manager): def get_by_natural_key(self, name): return self.get(name=name) class Person(models.Model): objects = PersonManager() name = models.CharField(max_length=100) def __unicode__(self): return self.name class Meta: ordering = ('name',) def natural_key(self): return (self.name,) class SpyManager(PersonManager): def get_query_set(self): return super(SpyManager, self).get_query_set().filter(cover_blown=False) class Spy(Person): objects = SpyManager() cover_blown = models.BooleanField(default=False) class Visa(models.Model): person = models.ForeignKey(Person) permissions = models.ManyToManyField(Permission, blank=True) def __unicode__(self): return '%s %s' % (self.person.name, ', '.join(p.name for p in self.permissions.all())) class Book(models.Model): name = models.CharField(max_length=100) authors = models.ManyToManyField(Person) def __unicode__(self): return '%s by %s' % (self.name, ' and '.join(a.name for a in self.authors.all())) class Meta: ordering = ('name',)
gonzolino/heat
refs/heads/master
heat/db/sqlalchemy/migrate_repo/versions/045_stack_backup.py
13
# # Licensed under the Apache License, Version 2.0 (the "License"); you may # not use this file except in compliance with the License. You may obtain # a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, WITHOUT # WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. See the # License for the specific language governing permissions and limitations # under the License. import sqlalchemy def upgrade(migrate_engine): meta = sqlalchemy.MetaData(bind=migrate_engine) stack = sqlalchemy.Table('stack', meta, autoload=True) backup = sqlalchemy.Column('backup', sqlalchemy.Boolean(), default=False) backup.create(stack) # Set backup flag for backup stacks, which are the only ones named "foo*" not_deleted = None stmt = sqlalchemy.select([stack.c.id, stack.c.name] ).where(stack.c.deleted_at == not_deleted) stacks = migrate_engine.execute(stmt) for s in stacks: if s.name.endswith('*'): values = {'backup': True} update = stack.update().where( stack.c.id == s.id).values(values) migrate_engine.execute(update)
adrienbrault/home-assistant
refs/heads/dev
tests/components/onvif/__init__.py
21
"""Tests for the ONVIF integration."""
mfjb/scikit-learn
refs/heads/master
sklearn/neural_network/tests/test_rbm.py
142
import sys import re import numpy as np from scipy.sparse import csc_matrix, csr_matrix, lil_matrix from sklearn.utils.testing import (assert_almost_equal, assert_array_equal, assert_true) from sklearn.datasets import load_digits from sklearn.externals.six.moves import cStringIO as StringIO from sklearn.neural_network import BernoulliRBM from sklearn.utils.validation import assert_all_finite np.seterr(all='warn') Xdigits = load_digits().data Xdigits -= Xdigits.min() Xdigits /= Xdigits.max() def test_fit(): X = Xdigits.copy() rbm = BernoulliRBM(n_components=64, learning_rate=0.1, batch_size=10, n_iter=7, random_state=9) rbm.fit(X) assert_almost_equal(rbm.score_samples(X).mean(), -21., decimal=0) # in-place tricks shouldn't have modified X assert_array_equal(X, Xdigits) def test_partial_fit(): X = Xdigits.copy() rbm = BernoulliRBM(n_components=64, learning_rate=0.1, batch_size=20, random_state=9) n_samples = X.shape[0] n_batches = int(np.ceil(float(n_samples) / rbm.batch_size)) batch_slices = np.array_split(X, n_batches) for i in range(7): for batch in batch_slices: rbm.partial_fit(batch) assert_almost_equal(rbm.score_samples(X).mean(), -21., decimal=0) assert_array_equal(X, Xdigits) def test_transform(): X = Xdigits[:100] rbm1 = BernoulliRBM(n_components=16, batch_size=5, n_iter=5, random_state=42) rbm1.fit(X) Xt1 = rbm1.transform(X) Xt2 = rbm1._mean_hiddens(X) assert_array_equal(Xt1, Xt2) def test_small_sparse(): # BernoulliRBM should work on small sparse matrices. X = csr_matrix(Xdigits[:4]) BernoulliRBM().fit(X) # no exception def test_small_sparse_partial_fit(): for sparse in [csc_matrix, csr_matrix]: X_sparse = sparse(Xdigits[:100]) X = Xdigits[:100].copy() rbm1 = BernoulliRBM(n_components=64, learning_rate=0.1, batch_size=10, random_state=9) rbm2 = BernoulliRBM(n_components=64, learning_rate=0.1, batch_size=10, random_state=9) rbm1.partial_fit(X_sparse) rbm2.partial_fit(X) assert_almost_equal(rbm1.score_samples(X).mean(), rbm2.score_samples(X).mean(), decimal=0) def test_sample_hiddens(): rng = np.random.RandomState(0) X = Xdigits[:100] rbm1 = BernoulliRBM(n_components=2, batch_size=5, n_iter=5, random_state=42) rbm1.fit(X) h = rbm1._mean_hiddens(X[0]) hs = np.mean([rbm1._sample_hiddens(X[0], rng) for i in range(100)], 0) assert_almost_equal(h, hs, decimal=1) def test_fit_gibbs(): # Gibbs on the RBM hidden layer should be able to recreate [[0], [1]] # from the same input rng = np.random.RandomState(42) X = np.array([[0.], [1.]]) rbm1 = BernoulliRBM(n_components=2, batch_size=2, n_iter=42, random_state=rng) # you need that much iters rbm1.fit(X) assert_almost_equal(rbm1.components_, np.array([[0.02649814], [0.02009084]]), decimal=4) assert_almost_equal(rbm1.gibbs(X), X) return rbm1 def test_fit_gibbs_sparse(): # Gibbs on the RBM hidden layer should be able to recreate [[0], [1]] from # the same input even when the input is sparse, and test against non-sparse rbm1 = test_fit_gibbs() rng = np.random.RandomState(42) from scipy.sparse import csc_matrix X = csc_matrix([[0.], [1.]]) rbm2 = BernoulliRBM(n_components=2, batch_size=2, n_iter=42, random_state=rng) rbm2.fit(X) assert_almost_equal(rbm2.components_, np.array([[0.02649814], [0.02009084]]), decimal=4) assert_almost_equal(rbm2.gibbs(X), X.toarray()) assert_almost_equal(rbm1.components_, rbm2.components_) def test_gibbs_smoke(): # Check if we don't get NaNs sampling the full digits dataset. # Also check that sampling again will yield different results. X = Xdigits rbm1 = BernoulliRBM(n_components=42, batch_size=40, n_iter=20, random_state=42) rbm1.fit(X) X_sampled = rbm1.gibbs(X) assert_all_finite(X_sampled) X_sampled2 = rbm1.gibbs(X) assert_true(np.all((X_sampled != X_sampled2).max(axis=1))) def test_score_samples(): # Test score_samples (pseudo-likelihood) method. # Assert that pseudo-likelihood is computed without clipping. # See Fabian's blog, http://bit.ly/1iYefRk rng = np.random.RandomState(42) X = np.vstack([np.zeros(1000), np.ones(1000)]) rbm1 = BernoulliRBM(n_components=10, batch_size=2, n_iter=10, random_state=rng) rbm1.fit(X) assert_true((rbm1.score_samples(X) < -300).all()) # Sparse vs. dense should not affect the output. Also test sparse input # validation. rbm1.random_state = 42 d_score = rbm1.score_samples(X) rbm1.random_state = 42 s_score = rbm1.score_samples(lil_matrix(X)) assert_almost_equal(d_score, s_score) # Test numerical stability (#2785): would previously generate infinities # and crash with an exception. with np.errstate(under='ignore'): rbm1.score_samples(np.arange(1000) * 100) def test_rbm_verbose(): rbm = BernoulliRBM(n_iter=2, verbose=10) old_stdout = sys.stdout sys.stdout = StringIO() try: rbm.fit(Xdigits) finally: sys.stdout = old_stdout def test_sparse_and_verbose(): # Make sure RBM works with sparse input when verbose=True old_stdout = sys.stdout sys.stdout = StringIO() from scipy.sparse import csc_matrix X = csc_matrix([[0.], [1.]]) rbm = BernoulliRBM(n_components=2, batch_size=2, n_iter=1, random_state=42, verbose=True) try: rbm.fit(X) s = sys.stdout.getvalue() # make sure output is sound assert_true(re.match(r"\[BernoulliRBM\] Iteration 1," r" pseudo-likelihood = -?(\d)+(\.\d+)?," r" time = (\d|\.)+s", s)) finally: sys.stdout = old_stdout
comic/comic-django
refs/heads/master
app/tests/utils.py
1
from typing import Callable from urllib.parse import urlparse import pytest from django.conf import settings from django.contrib.auth.models import AnonymousUser from django.core.exceptions import PermissionDenied from django.test import RequestFactory, Client from django.views.generic import View from grandchallenge.challenges.models import Challenge from grandchallenge.subdomains.utils import reverse from tests.factories import SUPER_SECURE_TEST_PASSWORD, UserFactory def assert_redirect(uri: str, *args): request, response = assert_status(302, *args) redirect_url = list(urlparse(response.url))[2] assert uri == redirect_url return request, response def assert_status( code: int, user: settings.AUTH_USER_MODEL, view: View, challenge: Challenge, rf: RequestFactory, ): request = rf.get("/rand") request.challenge = challenge if user is not None: request.user = user view = view.as_view() if code == 403: with pytest.raises(PermissionDenied): view(request) response = None else: response = view(request) assert response.status_code == code return request, response def get_view_for_user( *, viewname: str = None, challenge: Challenge = None, client: Client, method: Callable = None, user: settings.AUTH_USER_MODEL = None, url: str = None, reverse_kwargs: dict = None, **kwargs, ): """ Returns the view for a particular user """ if url is None: extra_kwargs = {} if challenge: extra_kwargs.update({"challenge_short_name": challenge.short_name}) if reverse_kwargs is not None: for key, value in reverse_kwargs.items(): if value is not None: extra_kwargs.update({key: value}) url = reverse(viewname, kwargs=extra_kwargs) elif viewname: raise AttributeError( "You defined both a viewname and a url, only " "use one!" ) if user and not isinstance(user, AnonymousUser): client.login( username=user.username, password=SUPER_SECURE_TEST_PASSWORD ) if method is None: method = client.get url, kwargs = get_http_host(url=url, kwargs=kwargs) try: response = method(url, **kwargs) finally: if user: client.logout() return response def get_http_host(*, url, kwargs): """ Takes a url and splits out the http host, if found """ urlparts = urlparse(url) if urlparts[1]: kwargs.update({"HTTP_HOST": urlparts[1]}) url = urlparts[2] return url, kwargs def assert_viewname_status(*, code: int, **kwargs): """ Asserts that a viewname for challenge_short_name and pk returns status code `code` for a particular user """ response = get_view_for_user(**kwargs) assert response.status_code == code return response def assert_viewname_redirect(*, redirect_url: str, **kwargs): """ Asserts that a view redirects to the given url. See assert_viewname_status for kwargs details """ response = assert_viewname_status(code=302, **kwargs) assert list(urlparse(response.url))[2] == redirect_url return response def validate_admin_only_view(*, two_challenge_set, client: Client, **kwargs): """ Assert that a view is only accessible to administrators for that particular challenge """ # No user assert_viewname_redirect( redirect_url=settings.LOGIN_URL, challenge=two_challenge_set.ChallengeSet1.challenge, client=client, **kwargs, ) tests = [ (403, two_challenge_set.ChallengeSet1.non_participant), (403, two_challenge_set.ChallengeSet1.participant), (403, two_challenge_set.ChallengeSet1.participant1), (200, two_challenge_set.ChallengeSet1.creator), (200, two_challenge_set.ChallengeSet1.admin), (403, two_challenge_set.ChallengeSet2.non_participant), (403, two_challenge_set.ChallengeSet2.participant), (403, two_challenge_set.ChallengeSet2.participant1), (403, two_challenge_set.ChallengeSet2.creator), (403, two_challenge_set.ChallengeSet2.admin), (200, two_challenge_set.admin12), (403, two_challenge_set.participant12), (200, two_challenge_set.admin1participant2), ] for test in tests: assert_viewname_status( code=test[0], challenge=two_challenge_set.ChallengeSet1.challenge, client=client, user=test[1], **kwargs, ) def validate_admin_or_participant_view( *, two_challenge_set, client: Client, **kwargs ): """ Assert that a view is only accessible to administrators or participants of that particular challenge """ # No user assert_viewname_redirect( redirect_url=settings.LOGIN_URL, challenge=two_challenge_set.ChallengeSet1.challenge, client=client, **kwargs, ) tests = [ (403, two_challenge_set.ChallengeSet1.non_participant), (200, two_challenge_set.ChallengeSet1.participant), (200, two_challenge_set.ChallengeSet1.participant1), (200, two_challenge_set.ChallengeSet1.creator), (200, two_challenge_set.ChallengeSet1.admin), (403, two_challenge_set.ChallengeSet2.non_participant), (403, two_challenge_set.ChallengeSet2.participant), (403, two_challenge_set.ChallengeSet2.participant1), (403, two_challenge_set.ChallengeSet2.creator), (403, two_challenge_set.ChallengeSet2.admin), (200, two_challenge_set.admin12), (200, two_challenge_set.participant12), (200, two_challenge_set.admin1participant2), ] for test in tests: assert_viewname_status( code=test[0], challenge=two_challenge_set.ChallengeSet1.challenge, client=client, user=test[1], **kwargs, ) def validate_open_view(*, challenge_set, client: Client, **kwargs): tests = [ (200, None), (200, challenge_set.non_participant), (200, challenge_set.participant), (200, challenge_set.participant1), (200, challenge_set.creator), (200, challenge_set.admin), ] for test in tests: assert_viewname_status( code=test[0], challenge=challenge_set.challenge, client=client, user=test[1], **kwargs, ) def validate_logged_in_view(*, challenge_set, client: Client, **kwargs): assert_viewname_redirect( redirect_url=settings.LOGIN_URL, challenge=challenge_set.challenge, client=client, **kwargs, ) tests = [ (200, challenge_set.non_participant), (200, challenge_set.participant), (200, challenge_set.participant1), (200, challenge_set.creator), (200, challenge_set.admin), ] for test in tests: assert_viewname_status( code=test[0], challenge=challenge_set.challenge, client=client, user=test[1], **kwargs, ) def validate_staff_only_view(*, client: Client, **kwargs): assert_viewname_redirect( redirect_url=settings.LOGIN_URL, client=client, **kwargs ) user = UserFactory() staff_user = UserFactory(is_staff=True) tests = [(200, staff_user), (403, user)] for test in tests: assert_viewname_status( code=test[0], client=client, user=test[1], **kwargs ) def validate_admin_only_text_in_page( *, expected_text, two_challenge_set, client: Client, **kwargs ): tests = [ (False, None), (False, two_challenge_set.ChallengeSet1.non_participant), (False, two_challenge_set.ChallengeSet1.participant), (False, two_challenge_set.ChallengeSet1.participant1), (True, two_challenge_set.ChallengeSet1.creator), (True, two_challenge_set.ChallengeSet1.admin), (False, two_challenge_set.ChallengeSet2.non_participant), (False, two_challenge_set.ChallengeSet2.participant), (False, two_challenge_set.ChallengeSet2.participant1), (False, two_challenge_set.ChallengeSet2.creator), (False, two_challenge_set.ChallengeSet2.admin), (True, two_challenge_set.admin12), (False, two_challenge_set.participant12), (True, two_challenge_set.admin1participant2), ] for test in tests: response = assert_viewname_status( code=200, challenge=two_challenge_set.ChallengeSet1.challenge, client=client, user=test[1], **kwargs, ) assert (expected_text in str(response.content)) == test[0]
yueyongyue/saltshaker
refs/heads/master
files_manager/estimate.py
1
from django import template register = template.Library() @register.filter(is_safe=True) def to_str(value): if type(value) == list: value_str = '\n'.join(value) return value_str elif type(value) == dict: value_l = [] for key in value: tmp = str(key) + '\n ' + str(value[key]) value_l.append(tmp) value_str = '\n'.join(value_l) return value_str else: return value @register.filter(is_safe=True) def ListToStr(l,nu): return l[nu]
ESS-LLP/erpnext-medical
refs/heads/develop
erpnext/patches/v7_0/update_refdoc_in_landed_cost_voucher.py
54
from __future__ import unicode_literals import frappe def execute(): if "purchase_receipt" not in frappe.db.get_table_columns("Landed Cost Purchase Receipt"): return frappe.reload_doctype("Landed Cost Purchase Receipt") frappe.db.sql(""" update `tabLanded Cost Purchase Receipt` set receipt_document_type = 'Purchase Receipt', receipt_document = purchase_receipt where (receipt_document is null or receipt_document = '') and (purchase_receipt is not null and purchase_receipt != '') """)