rem stringlengths 1 226k | add stringlengths 0 227k | context stringlengths 6 326k | meta stringlengths 143 403 | input_ids listlengths 256 256 | attention_mask listlengths 256 256 | labels listlengths 128 128 |
|---|---|---|---|---|---|---|
def compare(self, xyzFilename1, xyzFilename2): L1, A1 = self.readXyz(xyzFilename1) L2, A2 = self.readXyz(xyzFilename2) | def compare(self, questionableXyzFile, knownGoodXyzFile): todo("handle multiple-frame xyz files from animations") L1, A1 = self.readXyz(questionableXyzFile) L2, A2 = self.readXyz(knownGoodXyzFile) | def compare(self, xyzFilename1, xyzFilename2): # TODO: handle multiple-frame xyz files from animations L1, A1 = self.readXyz(xyzFilename1) L2, A2 = self.readXyz(xyzFilename2) for xyz1, xyz2 in map(None, L1, L2): a1, a2, L = xyz1 a11, a22, LL = xyz2 if a1 != a11 or a2 != a22: raise LengthMismatch("Term (%d, %d) versus (%d, %d)" % (a11, a22, a1, a2)) if abs(L - LL) > LENGTH_TOLERANCE: raise LengthMismatch("(%d, %d) -> %f versus %f" % (a1, a2, LL, L)) for xyz1, xyz2 in map(None, A1, A2): a1, a2, a3, A = xyz1 a11, a22, a33, AA = xyz2 if a1 != a11 or a2 != a22 or a3 != a33: raise AngleMismatch("Term (%d, %d, %d) versus (%d, %d, %d)" % (a11, a22, a33, a1, a2, a3)) if abs(L - LL) > ANGLE_TOLERANCE: raise AngleMismatch("(%d, %d, %d) -> %f versus %f" % (a1, a2, a3, AA, A)) | d4548954f0b2ec61c249ce8e1b4af97c07a52a78 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/11221/d4548954f0b2ec61c249ce8e1b4af97c07a52a78/tests.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3400,
12,
2890,
16,
14779,
5359,
21,
16,
14779,
5359,
22,
4672,
468,
2660,
30,
1640,
3229,
17,
3789,
14779,
1390,
628,
25536,
511,
21,
16,
432,
21,
273,
365,
18,
896,
60,
93,
94,
12,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3400,
12,
2890,
16,
14779,
5359,
21,
16,
14779,
5359,
22,
4672,
468,
2660,
30,
1640,
3229,
17,
3789,
14779,
1390,
628,
25536,
511,
21,
16,
432,
21,
273,
365,
18,
896,
60,
93,
94,
12,... |
return Overlay( AppOuterLayer( expr), DrawInCorner( top_left_corner, (-1,1)), DrawInCorner( bottom_left_corner, (-1,-1)), | return AppOuterLayer( Overlay( expr, DrawInCorner( top_left_corner, (-1,1)), | def testbed(expr): "this turns the current testexpr into the actual expr to render" ## return Overlay(expr, Closer(Rect(1,1,black), 3.4)) #stub ## return Overlay(expr, If(1,DrawInCorner1,Closer)(Highlightable(Rect(1,1,black),Rect(1,1,green),projection=True))) ## return Overlay(expr, DrawInCorner(Highlightable(Rect(1,1,black),Rect(1,1,green)) )) return Overlay( AppOuterLayer( expr), | 0f6c5d19bf88c1380bfe4de9834fbb499df34df8 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/11221/0f6c5d19bf88c1380bfe4de9834fbb499df34df8/test.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
2992,
12,
8638,
4672,
315,
2211,
20748,
326,
783,
1842,
8638,
1368,
326,
3214,
3065,
358,
1743,
6,
7541,
327,
26834,
12,
8638,
16,
22442,
550,
12,
6120,
12,
21,
16,
21,
16,
11223... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
2992,
12,
8638,
4672,
315,
2211,
20748,
326,
783,
1842,
8638,
1368,
326,
3214,
3065,
358,
1743,
6,
7541,
327,
26834,
12,
8638,
16,
22442,
550,
12,
6120,
12,
21,
16,
21,
16,
11223... |
for pkg in conflicts: | for pkg in conflicts.keys(): | def checkConflicts(self): conflicts = self.getConflicts() if len(conflicts) == 0: return self.OK | 24eacb956168a0d35742ff85786a9dcc5f740370 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/1143/24eacb956168a0d35742ff85786a9dcc5f740370/resolver.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
866,
30897,
12,
2890,
4672,
14450,
273,
365,
18,
588,
30897,
1435,
309,
562,
12,
20340,
87,
13,
422,
374,
30,
327,
365,
18,
3141,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
866,
30897,
12,
2890,
4672,
14450,
273,
365,
18,
588,
30897,
1435,
309,
562,
12,
20340,
87,
13,
422,
374,
30,
327,
365,
18,
3141,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-10... |
yield (packageId, dep.name, classId, 'run') | if dep.name not in ignored_names: yield (packageId, classId, dep.name, 'run') | def lookupUsedByDeps(packages): for packageId, package in enumerate(packages): for classId in sorted(package.classes): classObj = ClassIdToObject[classId] classDeps = classObj.dependencies(variants) for dep in classDeps["load"]: yield (packageId, dep.name, classId, 'load') # the packageId is somewhat bogus here for dep in classDeps["run"]: yield (packageId, dep.name, classId, 'run') return | a5d704758da36b45ee452c5ee2d16861bf2b225e /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/5718/a5d704758da36b45ee452c5ee2d16861bf2b225e/Generator.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3689,
6668,
858,
14430,
12,
10308,
4672,
364,
2181,
548,
16,
2181,
316,
4241,
12,
10308,
4672,
364,
31181,
316,
3115,
12,
5610,
18,
4701,
4672,
667,
2675,
273,
1659,
548,
19121,
63,
1106... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3689,
6668,
858,
14430,
12,
10308,
4672,
364,
2181,
548,
16,
2181,
316,
4241,
12,
10308,
4672,
364,
31181,
316,
3115,
12,
5610,
18,
4701,
4672,
667,
2675,
273,
1659,
548,
19121,
63,
1106... |
for d in args: if d is None: continue if type(d) is type(''): l = len(string.split(d)) else: l = 1 if l == 0 or l > 1 and (d[0],d[-1]) not in (('(',')'),('"','"')): data = '%s "%s"' % (data, d) else: data = '%s %s' % (data, d) | for arg in args: if arg is None: continue data = '%s %s' % (data, self._checkquote(arg)) | def _command(self, name, *args): | 8c0622114bc95f9f5b85a8a7e4ed8619bf3ff024 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8546/8c0622114bc95f9f5b85a8a7e4ed8619bf3ff024/imaplib.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
3076,
12,
2890,
16,
508,
16,
380,
1968,
4672,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
3076,
12,
2890,
16,
508,
16,
380,
1968,
4672,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
raise "\nERROR: Invalid Tree Item. " | raise Exception("\nERROR: Invalid Tree Item. ") | def GetPyData(self, item): """Returns the data associated to an item.""" | e1463b9df091ad0e9d76292564d4389882b28de1 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12725/e1463b9df091ad0e9d76292564d4389882b28de1/customtreectrl.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
968,
9413,
751,
12,
2890,
16,
761,
4672,
3536,
1356,
326,
501,
3627,
358,
392,
761,
12123,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
968,
9413,
751,
12,
2890,
16,
761,
4672,
3536,
1356,
326,
501,
3627,
358,
392,
761,
12123,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-1... |
queryparser = 'release/bin/Lib/site-packages/QueryParser.py' | def generateDocs(options, outputDir): if options.verbose: verbosity = 4 else: verbosity = 1 if sys.platform == 'cygwin' or os.name == 'nt': chandlerdb = 'release/bin/Lib/site-packages/chandlerdb' queryparser = 'release/bin/Lib/site-packages/QueryParser.py' pyicu = 'release/bin/Lib/site-packages/PyICU.py' pylucene = 'release/bin/Lib/site-packages/PyLucene.py' elif sys.platform == 'darwin': chandlerdb = 'release/Library/Frameworks/Python.framework/Versions/2.4/lib/python2.4/site-packages/chandlerdb' queryparser = 'release/Library/Frameworks/Python.framework/Versions/2.4/lib/python2.4/site-packages/QueryParser.py' pyicu = 'release/Library/Frameworks/Python.framework/Versions/2.4/lib/python2.4/site-packages/PyICU.py' pylucene = 'release/Library/Frameworks/Python.framework/Versions/2.4/lib/python2.4/site-packages/PyLucene.py' else: chandlerdb = 'release/lib/python2.4/site-packages/chandlerdb' queryparser = 'release/lib/python2.4/site-packages/QueryParser.py' pyicu = 'release/lib/python2.4/site-packages/PyICU.py' pylucene = 'release/lib/python2.4/site-packages/PyLucene.py' chandlerBin = os.getenv('CHANDLERBIN') chandlerdb = os.path.join(chandlerBin, chandlerdb) queryparser = os.path.join(chandlerBin, queryparser) pyicu = os.path.join(chandlerBin, pyicu) pylucene = os.path.join(chandlerBin, pylucene) targetDir = os.path.join(outputDir, 'api') if not os.path.isdir(targetDir): _mkdirs(targetDir) # This is the options dictionary # It is used by most of the epydoc routines and # the contents were determined by examining epydoc/gui.py # and epydoc/cli.py e_options = { 'target': targetDir, 'verbosity': verbosity, 'prj_name': 'Chandler', 'action': 'html', 'tests': { 'basic': 1 }, 'show_imports': 0, 'frames': 1, 'private': 0, 'debug': 0, 'docformat': None, 'top': None, 'inheritance': "listed", 'alphabetical': 1, 'ignore_param_mismatch': 1, 'list_classes_separately': 0, 'modules': ['application', 'i18n', 'parcels/feeds', 'parcels/osaf', 'parcels/osaf/app', 'parcels/osaf/examples', 'parcels/osaf/framework', 'parcels/osaf/mail', 'parcels/osaf/pim', 'parcels/osaf/servlets', 'parcels/osaf/sharing', 'parcels/osaf/tests', 'parcels/osaf/views', 'parcels/photos', 'repository', 'samples/skeleton', 'tools', 'util', 'Chandler.py', 'version.py', chandlerdb, # This comes from internal queryparser, # This comes from external pyicu, # This comes from external pylucene, # This comes from external ], } # based on the code in epydoc's gui.py # with subtle changes made to make it work :) from epydoc.html import HTMLFormatter from epydoc.objdoc import DocMap, report_param_mismatches from epydoc.imports import import_module, find_modules from epydoc.objdoc import set_default_docformat set_default_docformat('epytext') try: modules = [] module_names = e_options['modules'] # walk thru list of modules and expand # any packages found for name in module_names[:]: if os.path.isdir(name): index = module_names.index(name) new_modules = find_modules(name) if new_modules: module_names[index:index+1] = new_modules elif options.verbose: print 'Error: %s is not a package' % name # basic regex to exclude directories from consideration exc = re.compile(".*tests.*|.*scripts.*") for name in module_names: if exc.match(name): continue if options.verbose: print 'IMPORT: %s' % name # try importing the module and # add it to the list if successful try: module = import_module(name) if module not in modules: modules.append(module) elif options.verbose: print ' (duplicate)' except ImportError, e: if options.verbose: print e if len(modules) == 0: print 'Error: no modules successfully loaded!' sys.exit(1) document_bases = 1 document_autogen_vars = 1 inheritance_groups = (e_options['inheritance'] == 'grouped') inherit_groups = (e_options['inheritance'] != 'grouped') # let epydoc create an empty document map d = DocMap(verbosity, document_bases, document_autogen_vars, inheritance_groups, inherit_groups) # walk the module list and let epydoc build the documentation for (module, num) in zip(modules, range(len(modules))): if options.verbose: print '\n***Building docs for %s***' % module.__name__ try: d.add(module) except Exception, e: print "Internal Error: %s" % e except: print "Internal Error" if not e_options['ignore_param_mismatch']: if not report_param_mismatches(d): print ' (To supress these warnings, use --ignore-param-mismatch)' htmldoc = HTMLFormatter(d, **e_options) numfiles = htmldoc.num_files() def progress_callback(path): (dir, file) = os.path.split(path) (root, d) = os.path.split(dir) if d in ('public', 'private'): fname = os.path.join(d, file) else: fname = file if options.verbose: print '\n***Writing %s***' % fname # Write the documentation. print "\n***Saving to %s" % targetDir htmldoc.write(targetDir, progress_callback) except Exception, e: print 'Internal error: ', e raise except: raise | fb002db03d47b9ce762eb7c6c81d990b789b7a03 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9228/fb002db03d47b9ce762eb7c6c81d990b789b7a03/gen_docs.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2103,
12656,
12,
2116,
16,
21412,
4672,
309,
702,
18,
11369,
30,
11561,
273,
1059,
469,
30,
11561,
273,
404,
225,
309,
2589,
18,
9898,
422,
296,
2431,
75,
8082,
11,
578,
1140,
18,
529,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2103,
12656,
12,
2116,
16,
21412,
4672,
309,
702,
18,
11369,
30,
11561,
273,
1059,
469,
30,
11561,
273,
404,
225,
309,
2589,
18,
9898,
422,
296,
2431,
75,
8082,
11,
578,
1140,
18,
529,... | |
if self._allowDrag and mouseDown and optionDown and found in self._selection: | if self._allowDrag and mouseDown and found in self._selection: | def _mouseSelection(self, event, mouseDown=False): if mouseDown: self._oldSelection = set(self._selection) | 2cdbb975cc9513653183dd9549ba6e63ccaaf1e6 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/8848/2cdbb975cc9513653183dd9549ba6e63ccaaf1e6/glyphCellView.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
11697,
6233,
12,
2890,
16,
871,
16,
7644,
4164,
33,
8381,
4672,
309,
7644,
4164,
30,
365,
6315,
1673,
6233,
273,
444,
12,
2890,
6315,
10705,
13,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
11697,
6233,
12,
2890,
16,
871,
16,
7644,
4164,
33,
8381,
4672,
309,
7644,
4164,
30,
365,
6315,
1673,
6233,
273,
444,
12,
2890,
6315,
10705,
13,
2,
-100,
-100,
-100,
-100,
-100,
-... |
the id-ref of the schema for the annotation content (can be empty) | the id-ref of the schema for the resource content (can be empty) | def create_resource(self, package_id, id, mimetype, schema, url): """Create a new resource and its associated content. | 14bb9e570257f9e83f9dfdf8b8543732f4823b65 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/11977/14bb9e570257f9e83f9dfdf8b8543732f4823b65/sqlite.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
752,
67,
3146,
12,
2890,
16,
2181,
67,
350,
16,
612,
16,
12595,
16,
1963,
16,
880,
4672,
3536,
1684,
279,
394,
1058,
471,
2097,
3627,
913,
18,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
752,
67,
3146,
12,
2890,
16,
2181,
67,
350,
16,
612,
16,
12595,
16,
1963,
16,
880,
4672,
3536,
1684,
279,
394,
1058,
471,
2097,
3627,
913,
18,
2,
-100,
-100,
-100,
-100,
-100,
-100,
... |
self.toneMarkLookup = {} | self._toneMarkLookup = {} | def getToneForToneMark(self, toneMark): """ Gets the tone for the given tone mark. | 01cc0f3b3089a2c824402c7d0c59f75b7c5af166 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/11128/01cc0f3b3089a2c824402c7d0c59f75b7c5af166/operator.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3181,
476,
1290,
56,
476,
3882,
12,
2890,
16,
30600,
3882,
4672,
3536,
11881,
326,
30600,
364,
326,
864,
30600,
2267,
18,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3181,
476,
1290,
56,
476,
3882,
12,
2890,
16,
30600,
3882,
4672,
3536,
11881,
326,
30600,
364,
326,
864,
30600,
2267,
18,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-... |
if self.args['DBSType'] != 'CGI': logging.error("Fatal error: only CGI DBS supported") sys.exit(1) | def __init__(self, **args): """ Arguments: args -- all arguments from StartComponent. Return: none | 87d7e4aaef9a55485525200d47dab0552b5b83de /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/8887/87d7e4aaef9a55485525200d47dab0552b5b83de/MergeSensorComponent.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
2826,
1968,
4672,
3536,
225,
13599,
30,
225,
833,
1493,
777,
1775,
628,
3603,
1841,
18,
225,
2000,
30,
225,
6555,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
2826,
1968,
4672,
3536,
225,
13599,
30,
225,
833,
1493,
777,
1775,
628,
3603,
1841,
18,
225,
2000,
30,
225,
6555,
2,
-100,
-100,
-100,
-100,
-100,
-100,
... | |
if Target == "NT386GNU" or (Target == "NT386" and GCC_BACKEND and OSType == "POSIX"): | if Host == "NT386GNU" or (Host == "NT386" and GCC_BACKEND and TargetOS == "POSIX"): | def SetupEnvironment(): SystemDrive = os.environ.get("SystemDrive", "") if os.environ.get("OS") == "Windows_NT": HostIsNT = True else: HostIsNT = False SystemDrive = os.environ.get("SYSTEMDRIVE") if SystemDrive: SystemDrive += os.path.sep # Do this earlier so that its link isn't a problem. # Looking in the registry HKEY_LOCAL_MACHINE\SOFTWARE\Cygnus Solutions\Cygwin\mounts v2 # would be reasonable here. if CM3IsCygwin: _SetupEnvironmentVariableAll( "PATH", ["cygwin1.dll"], os.path.join(SystemDrive, "cygwin", "bin")) # some host/target confusion here.. if Target == "NT386" and HostIsNT and Config == "NT386" and (not GCC_BACKEND) and OSType == "WIN32": VCBin = "" VCInc = "" VCLib = "" MspdbDir = "" # 4.0 e:\MSDEV # 5.0 E:\Program Files\DevStudio\SharedIDE MSDevDir = os.environ.get("MSDEVDIR") # 5.0 MSVCDir = os.environ.get("MSVCDIR") # E:\Program Files\DevStudio\VC # 7.1 Express VCToolkitInstallDir = os.environ.get("VCTOOLKITINSTALLDIR") # E:\Program Files\Microsoft Visual C++ Toolkit 2003 (not set by vcvars32) # 8.0 Express # E:\Program Files\Microsoft Visual Studio 8\VC # E:\Program Files\Microsoft Visual Studio 8\Common7\Tools DevEnvDir = os.environ.get("DevEnvDir") # E:\Program Files\Microsoft Visual Studio 8\Common7\IDE VSInstallDir = os.environ.get("VSINSTALLDIR") # E:\Program Files\Microsoft Visual Studio 8 # VS80CommonTools = os.environ.get("VS80COMNTOOLS") # E:\Program Files\Microsoft Visual Studio 8\Common7\Tools VCInstallDir = os.environ.get("VCINSTALLDIR") # E:\Program Files\Microsoft Visual Studio 8\VC # 9.0 Express # always, global #VS90COMNTOOLS=D:\msdev\90\Common7\Tools\ # after running the shortcut #VCINSTALLDIR=D:\msdev\90\VC #VSINSTALLDIR=D:\msdev\90 VSCommonTools = os.environ.get("VS90COMNTOOLS") if VSCommonTools and not VSInstallDir: VSInstallDir = RemoveLastPathElement(RemoveLastPathElement(VSCommonTools)) # The Windows SDK is carried with the express edition and tricky to find. # Best if folks just run the installed shortcut probably. # We do a pretty good job now of finding it, be need to encode # more paths to known versions. # This is not yet finished. # # Probe the partly version-specific less-polluting environment variables, # from newest to oldest. # That is, having setup alter PATH, INCLUDE, and LIB system-wide is not # a great idea, but having setup set DevEnvDir, VSINSTALLDIR, VS80COMNTOOLS, etc. # isn't so bad and we can temporarily establish the first set from the second # set. if VSInstallDir: # Visual C++ 2005/8.0, at least the Express Edition, free download # also Visual C++ 2008/9.0 Express Edition if not VCInstallDir: VCInstallDir = os.path.join(VSInstallDir, "VC") #print("VCInstallDir:" + VCInstallDir) if not DevEnvDir: DevEnvDir = os.path.join(VSInstallDir, "Common7", "IDE") #print("DevEnvDir:" + DevEnvDir) MspdbDir = DevEnvDir elif VCToolkitInstallDir: # free download Visual C++ 2003; no longer available VCInstallDir = VCToolkitInstallDir elif MSVCDir and MSDevDir: # Visual C++ 5.0 pass # do more research # VCInstallDir = MSVCDir elif MSDevDir: # Visual C++ 4.0, 5.0 pass # do more research # VCInstallDir = MSDevDir else: # This is what really happens on my machine, for 8.0. # It might be good to guide pylib.py to other versions, # however setting things up manually suffices and I have, um, # well automated. Msdev = os.path.join(SystemDrive, "msdev", "80") VCInstallDir = os.path.join(Msdev, "VC") DevEnvDir = os.path.join(Msdev, "Common7", "IDE") if VCInstallDir: VCBin = os.path.join(VCInstallDir, "bin") VCLib = os.path.join(VCInstallDir, "lib") VCInc = os.path.join(VCInstallDir, "include") if DevEnvDir: MspdbDir = DevEnvDir #elif VCBin: # MspdbDir = VCBin # Look for SDKs. # expand this as they are released/discovered # ordering is from newest to oldest PossibleSDKs = [os.path.join("Microsoft SDKs", "Windows", "v6.0A"), "Microsoft Platform SDK for Windows Server 2003 R2"] SDKs = [] for a in GetProgramFiles(): #print("checking " + a) for b in PossibleSDKs: c = os.path.join(a, b) #print("checking " + c) if isdir(c) and not (c in SDKs): SDKs.append(c) # Make sure %INCLUDE% contains errno.h and windows.h. # This doesn't work correctly for Cygwin Python, ok. if _CheckSetupEnvironmentVariableAll("INCLUDE", ["errno.h", "windows.h"], VCInc): for a in SDKs: b = os.path.join(a, "include") if isfile(os.path.join(b, "windows.h")): _SetupEnvironmentVariableAll("INCLUDE", ["errno.h", "windows.h"], VCInc + ";" + b, ";") break # Make sure %LIB% contains kernel32.lib and libcmt.lib. # We carry our own kernel32.lib so we don't look in the SDKs. # We usually use msvcrt.lib and not libcmt.lib, but Express 2003 had libcmt.lib and not msvcrt.lib # I think, and libcmt.lib is always present. _SetupEnvironmentVariableAll( "LIB", ["kernel32.lib", "libcmt.lib"], VCLib + ";" + os.path.join(InstallRoot, "lib")) # Check that cl.exe and link.exe are in path, and if not, add VCBin to it, # checking that they are in it. # # Do this before mspdb*dll because it sometimes gets it in the path. # (Why do we care?) _SetupEnvironmentVariableAll("PATH", ["cl", "link"], VCBin) # If none of mspdb*.dll are in PATH, add MpsdbDir to PATH, and check that one of them is in it. _SetupEnvironmentVariableAny( "PATH", ["mspdb80.dll", "mspdb71.dll", "mspdb70.dll", "mspdb60.dll", "mspdb50.dll", "mspdb41.dll", "mspdb40.dll", "dbi.dll"], MspdbDir) # Try to get mt.exe in %PATH% if it isn't already. # We only need this for certain toolsets. if not SearchPath("mt.exe", os.environ.get("PATH")): for a in SDKs: b = os.path.join(a, "bin") if isfile(os.path.join(b, "mt.exe")): SetEnvironmentVariable("PATH", os.environ.get("PATH") + os.pathsep + b) break # sys.exit(1) # The free Visual C++ 2003 has neither delayimp.lib nor msvcrt.lib. # Very old toolsets have no delayimp.lib. # The Quake config file checks these environment variables. Lib = os.environ.get("LIB") if not SearchPath("delayimp.lib", Lib): os.environ["USE_DELAYLOAD"] = "0" print("set USE_DELAYLOAD=0") if not SearchPath("msvcrt.lib", Lib): os.environ["USE_MSVCRT"] = "0" print("set USE_MSVCRT=0") # some host/target confusion here.. if Target == "NT386MINGNU" or (Target == "NT386" and GCC_BACKEND and OSType == "WIN32"): _ClearEnvironmentVariable("LIB") _ClearEnvironmentVariable("INCLUDE") _SetupEnvironmentVariableAll( "PATH", ["gcc", "as", "ld"], os.path.join(SystemDrive, "mingw", "bin")) # need to probe for ld that accepts response files. # For example, this version does not: # C:\dev2\cm3\scripts\python>ld -v # GNU ld version 2.15.91 20040904 # This comes with Qt I think (free Windows version) # # This version works: # C:\dev2\cm3\scripts\python>ld -v # GNU ld version 2.17.50 20060824 # Ensure msys make is ahead of mingwin make, by adding # msys to the start of the path after adding mingw to the # start of the path. Modula-3 does not generally use # make, but this might matter when building m3cg, and # is usually the right thing. _SetupEnvironmentVariableAll( "PATH", ["sh", "sed", "gawk", "make"], os.path.join(SystemDrive, "msys", "1.0", "bin")) # some host/target confusion here.. if Target == "NT386GNU" or (Target == "NT386" and GCC_BACKEND and OSType == "POSIX"): #_ClearEnvironmentVariable("LIB") #_ClearEnvironmentVariable("INCLUDE") #if HostIsNT: # _SetupEnvironmentVariableAll( # "PATH", # ["cygX11-6.dll"], # os.path.join(SystemDrive, "cygwin", "usr", "X11R6", "bin")) _SetupEnvironmentVariableAll( "PATH", ["gcc", "as", "ld"], os.path.join(SystemDrive, "cygwin", "bin")) | 2d4d2df3c53195cedf60628bf047247a16f565eb /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9328/2d4d2df3c53195cedf60628bf047247a16f565eb/pylib.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
10939,
5494,
13332,
2332,
24681,
273,
1140,
18,
28684,
18,
588,
2932,
3163,
24681,
3113,
1408,
13,
309,
1140,
18,
28684,
18,
588,
2932,
4618,
7923,
422,
315,
10399,
67,
1784,
6877,
4893,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
10939,
5494,
13332,
2332,
24681,
273,
1140,
18,
28684,
18,
588,
2932,
3163,
24681,
3113,
1408,
13,
309,
1140,
18,
28684,
18,
588,
2932,
4618,
7923,
422,
315,
10399,
67,
1784,
6877,
4893,
... |
os.renames(old_dir, new_dir) | atexit.register(os.renames, old_dir, new_dir) | def moveCollectedTorrents(self, old_dir, new_dir): #physical move old_dir = os.path.join(old_dir, 'collected_torrent_files') new_dir = os.path.join(new_dir, 'collected_torrent_files') os.renames(old_dir, new_dir) old_dir = os.path.join(old_dir, 'collected_subtitles_files') new_dir = os.path.join(new_dir, 'collected_subtitles_files') os.renames(old_dir, new_dir) old_dir = os.path.join(old_dir, 'downloadhelp') new_dir = os.path.join(new_dir, 'downloadhelp') os.renames(old_dir, new_dir) #update db self.guiUtility.torrentsearch_manager.torrent_db.updateTorrentDir(new_dir) | 6ec2fda1ea91397aad70681809c0e0a9781bc1e5 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9686/6ec2fda1ea91397aad70681809c0e0a9781bc1e5/settingsDialog.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3635,
10808,
329,
27266,
547,
87,
12,
2890,
16,
1592,
67,
1214,
16,
394,
67,
1214,
4672,
468,
20441,
3635,
1592,
67,
1214,
273,
1140,
18,
803,
18,
5701,
12,
1673,
67,
1214,
16,
296,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3635,
10808,
329,
27266,
547,
87,
12,
2890,
16,
1592,
67,
1214,
16,
394,
67,
1214,
4672,
468,
20441,
3635,
1592,
67,
1214,
273,
1140,
18,
803,
18,
5701,
12,
1673,
67,
1214,
16,
296,
... |
child = self.childAt(e.pos(), False) | child = self.childAt(e.pos()) | def mousePressEvent(self, e): if e is None: return child = self.childAt(e.pos(), False) if child is None: return if not isinstance(child, QLabel): return shortname = str(child.name()) nameidx = 1 while nameidx < len(shortname): if shortname[nameidx] < 'a' or shortname[nameidx] > 'z': shorttype = shortname[nameidx:] shortname = shortname[0:nameidx] nameidx += 1 if shortname in ['Gem', 'Points', 'Cost', 'Name']: slot = child.name()[-2:] if str(slot[0:1]) == '_': slot = slot[1:] if self.PlayerMade.isChecked(): self.gemClicked(self.currentTabLabel, int(slot)) return if shortname in ['', 'Label', 'Total', 'Item']: return if child.parent().name() == 'GroupResists': self.DelveItemsDialog(shortname, 'Resist') else: self.DelveItemsDialog(shortname) | b02b1b3c3ce350ce2866d80a832fd61c5ae3b6c0 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/6671/b02b1b3c3ce350ce2866d80a832fd61c5ae3b6c0/ScWindow.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
7644,
11840,
1133,
12,
2890,
16,
425,
4672,
309,
425,
353,
599,
30,
327,
1151,
273,
365,
18,
3624,
861,
12,
73,
18,
917,
10756,
309,
1151,
353,
599,
30,
327,
309,
486,
1549,
12,
3624... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
7644,
11840,
1133,
12,
2890,
16,
425,
4672,
309,
425,
353,
599,
30,
327,
1151,
273,
365,
18,
3624,
861,
12,
73,
18,
917,
10756,
309,
1151,
353,
599,
30,
327,
309,
486,
1549,
12,
3624... |
if len(stdout) > 1 and stdout[-1] == '\n': | if stdout.endswith('\n'): | def resolve(self, makefile, variables, fd, setting): #TODO: call this once up-front somewhere and save the result? shell, msys = util.checkmsyscompat() cline = self._arguments[0].resolvestr(makefile, variables, setting) | f7304aa55be56feac24cd669687661a6391ce8aa /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7005/f7304aa55be56feac24cd669687661a6391ce8aa/functions.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2245,
12,
2890,
16,
1221,
768,
16,
3152,
16,
5194,
16,
3637,
4672,
468,
6241,
30,
745,
333,
3647,
731,
17,
10211,
22234,
471,
1923,
326,
563,
35,
5972,
16,
4086,
1900,
273,
1709,
18,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2245,
12,
2890,
16,
1221,
768,
16,
3152,
16,
5194,
16,
3637,
4672,
468,
6241,
30,
745,
333,
3647,
731,
17,
10211,
22234,
471,
1923,
326,
563,
35,
5972,
16,
4086,
1900,
273,
1709,
18,
... |
print cmd | print yellow_text(cmd) | def f_compile(self,compiler,switches, source_files, module_dirs=None, temp_dir=''): | 40006b76c878853d4b2ed3be9c264071e6cea220 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/14925/40006b76c878853d4b2ed3be9c264071e6cea220/build_flib.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
284,
67,
11100,
12,
2890,
16,
9576,
16,
9610,
281,
16,
1084,
67,
2354,
16,
1605,
67,
8291,
33,
7036,
16,
1906,
67,
1214,
2218,
11,
4672,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
284,
67,
11100,
12,
2890,
16,
9576,
16,
9610,
281,
16,
1084,
67,
2354,
16,
1605,
67,
8291,
33,
7036,
16,
1906,
67,
1214,
2218,
11,
4672,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
if useCompiledSize: fsize = self._treeCompiler.getCompiledSize(classId, variants, optimize, recompile=True) mode = 'compiled' else: fsize = self._classes[classId]['size'] mode = 'source' if fsize > sizes[mode][0]: color = "red" fontsize = 15 elif fsize > sizes[mode][1]: color = "green" fontsize = 13 else: color = "blue" fontsize = 10 | if classId in self._classes: if useCompiledSize: fsize = self._treeCompiler.getCompiledSize(classId, variants, optimize, recompile=True) mode = 'compiled' else: fsize = self._classes[classId]['size'] mode = 'source' if fsize > sizes[mode][0]: color = "red" fontsize = 15 elif fsize > sizes[mode][1]: color = "green" fontsize = 13 else: color = "blue" fontsize = 10 | def getNodeAttribs(classId, useCompiledSize=False, optimize=[]): # return color according to size attribs = [] color = fontsize = None sizes = { # (big-threshold, medium-threshold) 'compiled' : (8000, 2000), 'source' : (20000, 5000) } if useCompiledSize: fsize = self._treeCompiler.getCompiledSize(classId, variants, optimize, recompile=True) mode = 'compiled' else: fsize = self._classes[classId]['size'] mode = 'source' | d1325a630dd674037a9e9c9ee82994269849d10d /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/5718/d1325a630dd674037a9e9c9ee82994269849d10d/Generator.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
5973,
31292,
12,
1106,
548,
16,
999,
20733,
1225,
33,
8381,
16,
10979,
33,
8526,
4672,
468,
327,
2036,
4888,
358,
963,
11460,
273,
5378,
2036,
273,
14869,
273,
599,
8453,
273,
288,
1377,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
5973,
31292,
12,
1106,
548,
16,
999,
20733,
1225,
33,
8381,
16,
10979,
33,
8526,
4672,
468,
327,
2036,
4888,
358,
963,
11460,
273,
5378,
2036,
273,
14869,
273,
599,
8453,
273,
288,
1377,... |
enable_periodic_save() gtk.quit_add(1, widgets.save_library) | if "--debug" not in sys.argv: enable_periodic_save() gtk.quit_add(1, widgets.save_library) | def main(): import util; util.mkdir(const.DIR) import signal, gtk, widgets SIGNALS = [signal.SIGINT, signal.SIGTERM, signal.SIGHUP] try: ratings = config.getint("settings", "ratings") except (ValueError, TypeError): pass else: util.RATING_PRECISION = 1.0/ratings widgets.init() enable_periodic_save() gtk.quit_add(1, widgets.save_library) for sig in SIGNALS: signal.signal(sig, gtk.main_quit) gtk.threads_init() gtk.main() | b6cfaf30c3010cf5741c4233cbb0f6c00e25b596 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/4764/b6cfaf30c3010cf5741c4233cbb0f6c00e25b596/quodlibet.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2774,
13332,
1930,
1709,
31,
1709,
18,
26686,
12,
10248,
18,
4537,
13,
225,
1930,
4277,
16,
22718,
16,
10965,
12057,
1013,
55,
273,
306,
10420,
18,
18513,
3217,
16,
4277,
18,
18513,
1517... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2774,
13332,
1930,
1709,
31,
1709,
18,
26686,
12,
10248,
18,
4537,
13,
225,
1930,
4277,
16,
22718,
16,
10965,
12057,
1013,
55,
273,
306,
10420,
18,
18513,
3217,
16,
4277,
18,
18513,
1517... |
" font-size='%.1f' x='-2' y='%.1f'>%.0f</text>\n" % (fs, -y, y+lowest)) | " font-size='%.1f' x='-8' y='%.1f'>%.0f</text>\n" % (fs, -y, (y+ybottom)/float(vs))) | def plot(arg, inp, out, meta): """Read data from `inp` and create a plot of the stations specified in the list `arg`. Plot is written to `out`. Metadata (station name, location) is takem from the `meta` file (usually v2.inv). """ import struct BAD = -9999 table = asdict(arg, inp) if meta: meta = get_meta(table, meta) title = [] for id11,d in meta.items(): title.append('%s %+06.2f%+07.2f %s' % (id11, d['lat'], d['lon'], d['name'])) title = '\n'.join(title) minyear = 9999 maxyear = -9999 highest = -9999 lowest = 9999 for _,lines in table.items(): for row in lines: year = int(row[12:16]) minyear = min(minyear, year) maxyear = max(maxyear, year) data = struct.unpack('5s'*12, row[16:-1]) for datum in map(int, data): if datum == BAD: continue highest = max(highest, datum) lowest = min(lowest, datum) if highest == -9999: raise Error('No data found for %s' % (', '.join(table))) # The data should be such that a station cannot have entirely # invalid data. At least one year should have at least one valid # datum. assert highest > -9999 assert lowest < 9999 highest /= 10.0 lowest /= 10.0 limyear = maxyear + 1 # Bounds of the box that displays data. In SVG viewBox format. databox = (minyear, lowest, limyear-minyear, highest-lowest) plotwidth = databox[2] plotheight = databox[3] out.write("""<svg width='1000px' height='750px' viewBox='0 0 %d %d' xmlns="http://www.w3.org/2000/svg" version="1.1">\n""" % (plotwidth+12, plotheight+20)) # Style out.write("""<defs> <style type="text/css"> path { stroke-width: 0.1; fill: none } path.singleton { stroke-width: 0.2; stroke-linecap: round } g#axes path { stroke-width:0.1; fill:none; stroke: #888 } g#axes text { fill: black; font-family: Verdana } g#title text { fill: black; font-family: Verdana } | 56f4bc366ef894f5205cbf82b6b8d1525a03fba4 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6890/56f4bc366ef894f5205cbf82b6b8d1525a03fba4/stationplot.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3207,
12,
3175,
16,
12789,
16,
596,
16,
2191,
4672,
3536,
1994,
501,
628,
1375,
31647,
68,
471,
752,
279,
3207,
434,
326,
29719,
1269,
316,
326,
666,
1375,
3175,
8338,
225,
15211,
353,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3207,
12,
3175,
16,
12789,
16,
596,
16,
2191,
4672,
3536,
1994,
501,
628,
1375,
31647,
68,
471,
752,
279,
3207,
434,
326,
29719,
1269,
316,
326,
666,
1375,
3175,
8338,
225,
15211,
353,
... |
return self._rangeObj.text | text=self._rangeObj.text if not text: text="" return text | def _get_text(self): return self._rangeObj.text | 893f641b159891f8a2d595ee8faf561bf6cceb58 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9340/893f641b159891f8a2d595ee8faf561bf6cceb58/MSHTML.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
588,
67,
955,
12,
2890,
4672,
327,
365,
6315,
3676,
2675,
18,
955,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
588,
67,
955,
12,
2890,
4672,
327,
365,
6315,
3676,
2675,
18,
955,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-10... |
if pathname[-1] == '/': | if pathname and pathname[-1] == '/': | def convert_path (pathname): """Return 'pathname' as a name that will work on the native filesystem, i.e. split it on '/' and put it back together again using the current directory separator. Needed because filenames in the setup script are always supplied in Unix style, and have to be converted to the local convention before we can actually use them in the filesystem. Raises ValueError on non-Unix-ish systems if 'pathname' either starts or ends with a slash. """ if os.sep == '/': return pathname if pathname[0] == '/': raise ValueError, "path '%s' cannot be absolute" % pathname if pathname[-1] == '/': raise ValueError, "path '%s' cannot end with '/'" % pathname paths = string.split(pathname, '/') while '.' in paths: paths.remove('.') if not paths: return os.curdir return apply(os.path.join, paths) | 9d8150fdae9a4a1dfa8f0a1f9c992042065f423e /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8125/9d8150fdae9a4a1dfa8f0a1f9c992042065f423e/util.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1765,
67,
803,
261,
28336,
4672,
3536,
990,
296,
28336,
11,
487,
279,
508,
716,
903,
1440,
603,
326,
6448,
6496,
16,
277,
18,
73,
18,
1416,
518,
603,
2023,
471,
1378,
518,
1473,
9475,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1765,
67,
803,
261,
28336,
4672,
3536,
990,
296,
28336,
11,
487,
279,
508,
716,
903,
1440,
603,
326,
6448,
6496,
16,
277,
18,
73,
18,
1416,
518,
603,
2023,
471,
1378,
518,
1473,
9475,
... |
else | else: | def __init__(data = None) if data == None: quickfix.DoubleField.__init__(self, 231) else quickfix.DoubleField.__init__(self, 231, data) | 484890147d4b23aac4b9d0e85e84fceab7e137c3 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8819/484890147d4b23aac4b9d0e85e84fceab7e137c3/quickfix_fields.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
892,
273,
599,
13,
309,
501,
422,
599,
30,
9549,
904,
18,
5265,
974,
16186,
2738,
972,
12,
2890,
16,
576,
6938,
13,
469,
30,
9549,
904,
18,
5265,
974,
16186,
273... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
892,
273,
599,
13,
309,
501,
422,
599,
30,
9549,
904,
18,
5265,
974,
16186,
2738,
972,
12,
2890,
16,
576,
6938,
13,
469,
30,
9549,
904,
18,
5265,
974,
16186,
273... |
("script", self.startTagScript), (("base", "link", "meta", "style", "title"), | (("script", "style"), self.startTagScriptStyle), (("base", "link", "meta", "title"), | def __init__(self, parser, tree): Phase.__init__(self, parser, tree) self.startTagHandler = utils.MethodDispatcher([ ("html", self.startTagHtml), ("script", self.startTagScript), (("base", "link", "meta", "style", "title"), self.startTagFromHead), ("body", self.startTagBody), (("address", "blockquote", "center", "dir", "div", "dl", "fieldset", "listing", "menu", "ol", "p", "pre", "ul"), self.startTagCloseP), ("form", self.startTagForm), (("li", "dd", "dt"), self.startTagListItem), ("plaintext",self.startTagPlaintext), (headingElements, self.startTagHeading), ("a", self.startTagA), (("b", "big", "em", "font", "i", "nobr", "s", "small", "strike", "strong", "tt", "u"),self.startTagFormatting), ("button", self.startTagButton), (("marquee", "object"), self.startTagMarqueeObject), ("xmp", self.startTagXmp), ("table", self.startTagTable), (("area", "basefont", "bgsound", "br", "embed", "img", "param", "spacer", "wbr"), self.startTagVoidFormatting), ("hr", self.startTagHr), ("image", self.startTagImage), ("input", self.startTagInput), ("isindex", self.startTagIsIndex), ("textarea", self.startTagTextarea), (("iframe", "noembed", "noframes", "noscript"), self.startTagCdata), ("select", self.startTagSelect), (("caption", "col", "colgroup", "frame", "frameset", "head", "option", "optgroup", "tbody", "td", "tfoot", "th", "thead", "tr"), self.startTagMisplaced), (("event-source", "section", "nav", "article", "aside", "header", "footer", "datagrid", "command"), self.startTagNew) ]) self.startTagHandler.default = self.startTagOther | 68e5292908c342cac5224497fa257ce9542cb3ba /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/10463/68e5292908c342cac5224497fa257ce9542cb3ba/html5parser.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
2082,
16,
2151,
4672,
19459,
16186,
2738,
972,
12,
2890,
16,
2082,
16,
2151,
13,
365,
18,
1937,
1805,
1503,
273,
2990,
18,
1305,
6681,
3816,
7566,
2620,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
2082,
16,
2151,
4672,
19459,
16186,
2738,
972,
12,
2890,
16,
2082,
16,
2151,
13,
365,
18,
1937,
1805,
1503,
273,
2990,
18,
1305,
6681,
3816,
7566,
2620,
... |
client = SystemAdministratorClient(self.host) | client = SystemAdministratorClient( self.host ) | def do_show(self,args): """ Add a record to the File Catalog usage: show software - show components for which software is available show installed - show components installed in the host show status - show status of the installed components show database - show the status of the databases show mysql - show the status of the MySQL server show log <system> <service|agent> show info - show version of software and setup """ argss = args.split() option = argss[0] del argss[0] if option == 'software': client = SystemAdministratorClient(self.host) result = client.getSoftwareComponents() if not result['OK']: print "ERROR:",result['Message'] else: serviceCount = 0 agentCount = 0 for compType in result['Value']: for system in result['Value'][compType]: for component in result['Value'][compType][system]: print compType.ljust(8),system.ljust(28),component.ljust(28) if compType == 'Services': serviceCount += 1 if compType == 'Agents': agentCount += 1 print "Total: %d services, %d agents" % (serviceCount,agentCount) elif option == 'installed': client = SystemAdministratorClient(self.host) result = client.getSetupComponents() if not result['OK']: print "ERROR:",result['Message'] else: serviceCount = 0 agentCount = 0 for compType in result['Value']: for system in result['Value'][compType]: for component in result['Value'][compType][system]: print compType.ljust(8),system.ljust(28),component.ljust(28) if compType == 'Services': serviceCount += 1 if compType == 'Agents': agentCount += 1 print "Total: %d services, %d agents" % (serviceCount,agentCount) elif option == 'status': client = SystemAdministratorClient(self.host) result = client.getOverallStatus() if not result['OK']: print "ERROR:",result['Message'] else: rDict = result['Value'] print " System",' '*20,'Name',' '*5,'Type',' '*23,'Setup Installed Runit Uptime PID' print '-'*116 for compType in rDict: for system in rDict[compType]: for component in rDict[compType][system]: if rDict[compType][system][component]['Installed']: print system.ljust(28),component.ljust(28),compType.lower()[:-1].ljust(7), if rDict[compType][system][component]['Setup']: print 'SetUp'.rjust(12), else: print 'NotSetup'.rjust(12), if rDict[compType][system][component]['Installed']: print 'Installed'.rjust(12), else: print 'NotInstalled'.rjust(12), print str(rDict[compType][system][component]['RunitStatus']).ljust(7), print str(rDict[compType][system][component]['Timeup']).rjust(7), print str(rDict[compType][system][component]['PID']).rjust(8), print elif option == 'database' or option == 'databases': client = SystemAdministratorClient(self.host) result = client.getDatabases(self.rootPwd) if not result['OK']: print "ERROR:",result['Message'] return if result.has_key('MySQLPassword'): self.rootPwd = result['MySQLPassword'] resultSW = client.getSoftwareDatabases() if not resultSW['OK']: print "ERROR:",resultSW['Message'] return sw = resultSW['Value'] installed = result['Value'] print for db in sw: if db in installed: print db.rjust(25),': Installed' else: print db.rjust(25),': Not installed' if not sw: print "No database found" elif option == 'mysql': client = SystemAdministratorClient(self.host) result = client.getMySQLStatus() if not result['OK']: print "ERROR:",result['Message'] elif result['Value']: for par,value in result['Value'].items(): print par.rjust(28),':',value else: print "No MySQL database found" elif option == "log": self.getLog(argss) elif option == "info": client = SystemAdministratorClient(self.host) result = client.getInfo() if not result['OK']: print "ERROR:",result['Message'] print "Setup:", result['Value']['Setup'] print "DIRAC version:",result['Value']['DIRAC'] if result['Value']['Extensions']: for e,v in result['Value']['Extensions'].items(): print "%s version" % e,v else: print "Unknown option:",option | b68208f4ea74e3c61aead9e641ff73054212c46c /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12864/b68208f4ea74e3c61aead9e641ff73054212c46c/SystemAdministratorClientCLI.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
741,
67,
4500,
12,
2890,
16,
1968,
4672,
3536,
1436,
279,
1409,
358,
326,
1387,
14953,
225,
4084,
30,
225,
2405,
17888,
1377,
300,
2405,
4085,
364,
1492,
17888,
353,
2319,
2405,
5876,
37... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
741,
67,
4500,
12,
2890,
16,
1968,
4672,
3536,
1436,
279,
1409,
358,
326,
1387,
14953,
225,
4084,
30,
225,
2405,
17888,
1377,
300,
2405,
4085,
364,
1492,
17888,
353,
2319,
2405,
5876,
37... |
class Track(BaseObject, Cacheable, Taggable): | class Track(_BaseObject, _Cacheable, _Taggable): | def toStr(self): """Returns a string representation of the object.""" return self.getArtist().getName().encode('utf-8') + ' - ' + self.getTitle().encode('utf-8') | 903c9b1622fe56617e5099f20abdc69f2090f8e0 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9926/903c9b1622fe56617e5099f20abdc69f2090f8e0/pylast.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
358,
1585,
12,
2890,
4672,
3536,
1356,
279,
533,
4335,
434,
326,
733,
12123,
225,
327,
365,
18,
588,
4411,
376,
7675,
17994,
7675,
3015,
2668,
3158,
17,
28,
6134,
397,
296,
300,
296,
3... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
358,
1585,
12,
2890,
4672,
3536,
1356,
279,
533,
4335,
434,
326,
733,
12123,
225,
327,
365,
18,
588,
4411,
376,
7675,
17994,
7675,
3015,
2668,
3158,
17,
28,
6134,
397,
296,
300,
296,
3... |
psout = ps.stdout.read() psout.strip() | psout = ps.communicate()[0] | def main(): # Initialize the update url to just be the ip address of this machine. ip = getmyip() updateurl = 'http://' + ip + ':12345/' if len(sys.argv) == 1: sys.argv.append(updateurl) else: sys.argv[1] = updateurl # Create a temp directory to serve the updates from that we can # automatically clean up when we are done. tmpserver = tempfile.mkdtemp() try: test_updater.create_folders(tmpserver) ############################ # Run the rsync only tests # ############################ # Run the noup test (Nothing should fail, nothing should be updated) runRsyncTest('-x', tmpserver + '/noup/') # Run the wronghash test(There should be an RsyncError, and no updates) runRsyncTest('-e', tmpserver + '/wronghash/') # Run the badkeysig test (There should be no updates) runRsyncTest('-x', tmpserver + '/badkeysig/') # Run the corruptmeta test (there should be an RsyncError, and no updates) runRsyncTest('-e', tmpserver + '/corruptmeta/') # Run the updatenmmain test (only nmmain should get updated) runRsyncTest('-u', tmpserver + '/updatenmmain/', ['nmmain.py', 'metainfo']) # Run an update that should get us into a state where the softwareupdater has # a different key than what the metainfo is signed with. The next test will # ensure that our method of dealing with this situation works. runRsyncTest('-u', tmpserver + '/updater/', ['softwareupdater.py', 'metainfo']) # Run an update that should successfully update from the strange state from # the previous test. runRsyncTest('-u', tmpserver + '/updater_new/', ['nmmain.py', 'metainfo']) ##################################### # Finished running rsync only tests # ##################################### # Copy back everything from noup so the restart tests start with a # clean slate. for originalfile in glob.glob(tmpserver + '/noup/*'): shutil.copy(originalfile, os.path.basename(originalfile)) ################################## # Software updater restart tests # ################################## # Start the web server for the first update webserver = run_webserver(tmpserver + '/updater/') # Keep track of whether ps is there (it isn't on Windows) no_ps = False if nonportable.ostype == 'Windows': # If we are running on windows, disable the ps calls. no_ps = True # ps works different on a mac, where we need to use 'ps -aww' instead of # 'ps -ef'. if nonportable.ostype == 'Darwin': pscommand = 'ps -aww' else: pscommand = 'ps -ef' updateprocess = subprocess.Popen(['python', 'softwareupdater.py']) if not no_ps: # Only do the ps check if ps is available ps = subprocess.Popen(pscommand + ' | grep "softwareupdater.py" | grep -v grep', shell=True, stdout=subprocess.PIPE) psout = ps.stdout.read() print 'Initial ps out:' print psout if psout == '': print 'Failure to start initially' # Wait for 2 minutes for the update to happen and the # process to die. for junk in range(60): if updateprocess.poll() != None: break time.sleep(2) ret = updateprocess.returncode if ret != 10: if ret == None: raise Exception("First software updater never died.") elif ret == 55: raise Exception("Software updater failed to get the process lock.") elif ret == 1: raise Exception("Softwareupdater failed with an uncaught exception. \n\ See end of softwareupdater.old in v2 for details.") else: raise Exception("Unknown return code from the software updater (" + str(ret) + ")") else: print "Old softwareupdater returned correctly" if not no_ps: # Only do the ps check if ps is available ps = subprocess.Popen(pscommand + ' | grep "softwareupdater.py" | grep -v grep', shell=True, stdout=subprocess.PIPE) psout = ps.stdout.read() psout.strip() print 'After ps out:' print psout if psout == '': print 'New updater failed to start!' else: print 'softwareupdater restart success!' # We need to kill the webserver serving from /updater, and start one serving # from updater_new kill_webserver(webserver.pid, updateurl) webserver = run_webserver(tmpserver + '/updater_new/') print "Waiting 2 minutes for the second update to happen" # Wait 2 minutes for the second update to happen. # Is there a way to get a handle for the new softwareupdater? time.sleep(120) # If nmmain's version has been updated, the second update was a success! nmmainfile = file('nmmain.py', 'r') nmmaindata = nmmainfile.read() nmmainfile.close() if 'version = "1234"' in nmmaindata: print 'Second update a success!' else: print 'Second update failed to happen within 2 minutes' # Kill the webserver again now that we are all done with it. kill_webserver(webserver.pid, updateurl) ###################################### # End Software updater restart tests # ###################################### except: traceback.print_exc() finally: # Clean up the temporary server directory. # This throws an exception on Windows if a file is still in use (for example, # if the webserver started in a test couldn't be shut down). shutil.rmtree(tmpserver) | c8fdaf6a7ab121cac0290ff7cc511ff8a9996737 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7995/c8fdaf6a7ab121cac0290ff7cc511ff8a9996737/test_updater_local.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2774,
13332,
468,
9190,
326,
1089,
880,
358,
2537,
506,
326,
2359,
1758,
434,
333,
5228,
18,
2359,
273,
336,
4811,
625,
1435,
1089,
718,
273,
296,
2505,
8931,
397,
2359,
397,
4290,
2138,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2774,
13332,
468,
9190,
326,
1089,
880,
358,
2537,
506,
326,
2359,
1758,
434,
333,
5228,
18,
2359,
273,
336,
4811,
625,
1435,
1089,
718,
273,
296,
2505,
8931,
397,
2359,
397,
4290,
2138,... |
if str[4] <= str[3] and str[4] != None: | if str[4] != None and str[4] <= str[3]: | def string_segmentation(args, cluster, ccs, string, groups, phrases, H_a, theta, R, hough_image, image): """This function should perform the string segmentation.""" | 69b5bb9017bd106b070f14f72121ba350eb10df5 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2794/69b5bb9017bd106b070f14f72121ba350eb10df5/main.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
533,
67,
9273,
367,
12,
1968,
16,
2855,
16,
276,
2143,
16,
533,
16,
3252,
16,
31413,
16,
670,
67,
69,
16,
7338,
16,
534,
16,
366,
4966,
67,
2730,
16,
1316,
4672,
3536,
2503,
445,
1... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
533,
67,
9273,
367,
12,
1968,
16,
2855,
16,
276,
2143,
16,
533,
16,
3252,
16,
31413,
16,
670,
67,
69,
16,
7338,
16,
534,
16,
366,
4966,
67,
2730,
16,
1316,
4672,
3536,
2503,
445,
1... |
cmd_text = 'cd ' + self.workdir + '; ' + command.command | cmd_text = 'cd ' + self.workdir + '; ' if command.errors_to > '': cmd_text += ' rm -f ' + command.errors_to + '; ' cmd_text = cmd_text + command.command | def make(self, name='all'): """ Runs the makefile specified target (defaulting to all) and records errors, results and other status flags against the document. If the document not been mirrored, this routine will not attempt to build it. | e85410e8074f79705a34ba9e936a469b56cc2427 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/1319/e85410e8074f79705a34ba9e936a469b56cc2427/Makefile.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1221,
12,
2890,
16,
508,
2218,
454,
11,
4672,
3536,
1939,
87,
326,
1221,
768,
1269,
1018,
261,
1886,
310,
358,
777,
13,
471,
3853,
1334,
16,
1686,
471,
1308,
1267,
2943,
5314,
326,
166... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1221,
12,
2890,
16,
508,
2218,
454,
11,
4672,
3536,
1939,
87,
326,
1221,
768,
1269,
1018,
261,
1886,
310,
358,
777,
13,
471,
3853,
1334,
16,
1686,
471,
1308,
1267,
2943,
5314,
326,
166... |
self.onReplyOrForWardEvent(Mail.forwardMessage) | self.onReplyOrForwardEvent(Mail.forwardMessage) | def onForwardEvent(self, event): self.onReplyOrForWardEvent(Mail.forwardMessage) | 8fc24be64a9f8e190acdaf8331a4329984eed270 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9228/8fc24be64a9f8e190acdaf8331a4329984eed270/PimBlocks.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
603,
8514,
1133,
12,
2890,
16,
871,
4672,
365,
18,
265,
7817,
1162,
1290,
59,
1060,
1133,
12,
6759,
18,
11565,
1079,
13,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
603,
8514,
1133,
12,
2890,
16,
871,
4672,
365,
18,
265,
7817,
1162,
1290,
59,
1060,
1133,
12,
6759,
18,
11565,
1079,
13,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
if not compile_dir(dir, maxlevels, ddir, force, rx): | if not compile_dir(dir, maxlevels, ddir, force, rx, quiet): | def main(): """Script main program.""" import getopt try: opts, args = getopt.getopt(sys.argv[1:], 'lfd:x:') except getopt.error, msg: print msg print "usage: python compileall.py [-l] [-f] [-d destdir] " \ "[-s regexp] [directory ...]" print "-l: don't recurse down" print "-f: force rebuild even if timestamps are up-to-date" print "-d destdir: purported directory name for error messages" print " if no directory arguments, -l sys.path is assumed" print "-x regexp: skip files matching the regular expression regexp" print " the regexp is search for in the full path of the file" sys.exit(2) maxlevels = 10 ddir = None force = 0 rx = None for o, a in opts: if o == '-l': maxlevels = 0 if o == '-d': ddir = a if o == '-f': force = 1 if o == '-x': import re rx = re.compile(a) if ddir: if len(args) != 1: print "-d destdir require exactly one directory argument" sys.exit(2) success = 1 try: if args: for dir in args: if not compile_dir(dir, maxlevels, ddir, force, rx): success = 0 else: success = compile_path() except KeyboardInterrupt: print "\n[interrupt]" success = 0 return success | 0f9b0fd3c2a89f138e2ebe25a20280803a090026 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8125/0f9b0fd3c2a89f138e2ebe25a20280803a090026/compileall.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2774,
13332,
3536,
3651,
2774,
5402,
12123,
1930,
336,
3838,
775,
30,
1500,
16,
833,
273,
336,
3838,
18,
588,
3838,
12,
9499,
18,
19485,
63,
21,
30,
6487,
296,
80,
8313,
30,
92,
2497,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2774,
13332,
3536,
3651,
2774,
5402,
12123,
1930,
336,
3838,
775,
30,
1500,
16,
833,
273,
336,
3838,
18,
588,
3838,
12,
9499,
18,
19485,
63,
21,
30,
6487,
296,
80,
8313,
30,
92,
2497,
... |
Inf = Decimal('Inf') negInf = Decimal('-Inf') NaN = Decimal('NaN') Dec_0 = Decimal(0) Dec_p1 = Decimal(1) Dec_n1 = Decimal(-1) Infsign = (Inf, negInf) | _Inf = Decimal('Inf') _negInf = Decimal('-Inf') _NaN = Decimal('NaN') _Dec_0 = Decimal(0) _Dec_p1 = Decimal(1) _Dec_n1 = Decimal(-1) _Infsign = (_Inf, _negInf) | def _format_align(body, spec_dict): """Given an unpadded, non-aligned numeric string, add padding and aligment to conform with the given format specifier dictionary (as output from parse_format_specifier). It's assumed that if body is negative then it starts with '-'. Any leading sign ('-' or '+') is stripped from the body before applying the alignment and padding rules, and replaced in the appropriate position. """ # figure out the sign; we only examine the first character, so if # body has leading whitespace the results may be surprising. if len(body) > 0 and body[0] in '-+': sign = body[0] body = body[1:] else: sign = '' if sign != '-': if spec_dict['sign'] in ' +': sign = spec_dict['sign'] else: sign = '' # how much extra space do we have to play with? minimumwidth = spec_dict['minimumwidth'] fill = spec_dict['fill'] padding = fill*(max(minimumwidth - (len(sign+body)), 0)) align = spec_dict['align'] if align == '<': result = padding + sign + body elif align == '>': result = sign + body + padding elif align == '=': result = sign + padding + body else: #align == '^' half = len(padding)//2 result = padding[:half] + sign + body + padding[half:] # make sure that result is unicode if necessary if spec_dict['unicode']: result = unicode(result) return result | 335c6f295295ab5b62a80d25d26d89243eb80fb0 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/3187/335c6f295295ab5b62a80d25d26d89243eb80fb0/decimal.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2139,
67,
7989,
12,
3432,
16,
857,
67,
1576,
4672,
3536,
6083,
392,
640,
6982,
785,
16,
1661,
17,
20677,
6389,
533,
16,
527,
4992,
471,
524,
360,
475,
358,
20156,
598,
326,
864,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2139,
67,
7989,
12,
3432,
16,
857,
67,
1576,
4672,
3536,
6083,
392,
640,
6982,
785,
16,
1661,
17,
20677,
6389,
533,
16,
527,
4992,
471,
524,
360,
475,
358,
20156,
598,
326,
864,
... |
a = _castCopyAndTranspose(t, a) | a = _fastCopyAndTranspose(t, a) | def cholesky(a): _assertRank2(a) _assertSquareness(a) t, result_t = _commonType(a) a = _castCopyAndTranspose(t, a) m = a.shape[0] n = a.shape[1] if isComplexType(t): lapack_routine = lapack_lite.zpotrf else: lapack_routine = lapack_lite.dpotrf results = lapack_routine('L', n, a, m, 0) if results['info'] > 0: raise LinAlgError, 'Matrix is not positive definite - Cholesky decomposition cannot be computed' s = triu(a, k=0).transpose() return array(s, dtype=result_t, copy=True) | 958601cb55c15930a443057d61230a6aa5a45c02 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/14925/958601cb55c15930a443057d61230a6aa5a45c02/linalg.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
462,
9112,
18465,
12,
69,
4672,
389,
11231,
12925,
22,
12,
69,
13,
389,
11231,
22255,
4496,
12,
69,
13,
268,
16,
563,
67,
88,
273,
389,
6054,
559,
12,
69,
13,
279,
273,
389,
8076,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
462,
9112,
18465,
12,
69,
4672,
389,
11231,
12925,
22,
12,
69,
13,
389,
11231,
22255,
4496,
12,
69,
13,
268,
16,
563,
67,
88,
273,
389,
6054,
559,
12,
69,
13,
279,
273,
389,
8076,
... |
file.write (zcfgdata) crc32 = zlib.crc32 (cfgdata) header = struct.pack ("<iiiiiiii", 0x12345678, comp_method, crc32, len (cfgdata), len (zcfgdata), 0, 0, 0) | file.write (cfgdata) header = struct.pack ("<ii", 0x12345679, len (cfgdata)) | def create_exe (self, arcname, fullname): import struct, zlib | 8a718307e8760854ee5452cdbef49b10400b0ced /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3187/8a718307e8760854ee5452cdbef49b10400b0ced/bdist_wininst.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
752,
67,
14880,
261,
2890,
16,
8028,
529,
16,
13321,
4672,
1930,
1958,
16,
24220,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
752,
67,
14880,
261,
2890,
16,
8028,
529,
16,
13321,
4672,
1930,
1958,
16,
24220,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
opt['allowedTypes'] = [types.ListType, types.TupleType] | opt['allowedTypes'] = [list, tuple] | def addOption(self, pos=-1, **kwargs): ''' Append an entry to the parameter specification list. Dictionary entries should be specified as keyword arguments such as ``name='option'``. More specifically, you can specify parameters ``name`` (required), ``label``, ``default`` (required), ``description``, ``validator``, ``type``, and ``separator``. This option will have a name specified by ``name`` and an initial default value specified by ``default``. | 0808dd362a5f2dd5d90d950500fd972461718062 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/401/0808dd362a5f2dd5d90d950500fd972461718062/simuOpt.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
12966,
12,
2890,
16,
949,
29711,
21,
16,
2826,
4333,
4672,
9163,
6181,
392,
1241,
358,
326,
1569,
7490,
666,
18,
16447,
3222,
1410,
506,
1269,
487,
4932,
1775,
4123,
487,
12176,
529,
221... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
12966,
12,
2890,
16,
949,
29711,
21,
16,
2826,
4333,
4672,
9163,
6181,
392,
1241,
358,
326,
1569,
7490,
666,
18,
16447,
3222,
1410,
506,
1269,
487,
4932,
1775,
4123,
487,
12176,
529,
221... |
raise ReadError, "not a gzip file" | raise ReadError("not a gzip file") | def gzopen(cls, name, mode="r", fileobj=None, compresslevel=9): """Open gzip compressed tar archive name for reading or writing. Appending is not allowed. """ if len(mode) > 1 or mode not in "rw": raise ValueError, "mode must be 'r' or 'w'" | e4751e3cdc8c271f24e46a6155f255b6e33da158 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8546/e4751e3cdc8c271f24e46a6155f255b6e33da158/tarfile.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
14136,
3190,
12,
6429,
16,
508,
16,
1965,
1546,
86,
3113,
17041,
33,
7036,
16,
8099,
2815,
33,
29,
4672,
3536,
3678,
10331,
8968,
8232,
5052,
508,
364,
6453,
578,
7410,
18,
6181,
310,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
14136,
3190,
12,
6429,
16,
508,
16,
1965,
1546,
86,
3113,
17041,
33,
7036,
16,
8099,
2815,
33,
29,
4672,
3536,
3678,
10331,
8968,
8232,
5052,
508,
364,
6453,
578,
7410,
18,
6181,
310,
... |
imagefilename = getThemeFile(themeName, node.attributes["filename"].value) if not doesFileExist(imagefilename): fatalError("Cannot find image for next button (%s)." % imagefilename) maskimagefilename = getThemeFile(themeName, node.attributes["mask"].value) if not doesFileExist(maskimagefilename): fatalError("Cannot find mask image for next button (%s)." % maskimagefilename) picture = Image.open(imagefilename,"r").resize((getScaledAttribute(node, "w"), getScaledAttribute(node, "h"))) picture = picture.convert("RGBA") bgimage.paste(picture, (getScaledAttribute(node, "x"), getScaledAttribute(node, "y")), picture) del picture write( "Added next button image %s " % imagefilename) picture=Image.open(maskimagefilename,"r").resize((getScaledAttribute(node, "w"), getScaledAttribute(node, "h"))) picture=picture.convert("RGBA") bgimagemask.paste(picture, (getScaledAttribute(node, "x"), getScaledAttribute(node, "y")), picture) del picture write( "Added next button mask image %s" % imagefilename) | paintBackground(bgimage, node) paintButton(draw, bgimage, bgimagemask, node, infoDOM, itemnum, page, itemsonthispage, chapternumber, chapterlist) | def drawThemeItem(page, itemsonthispage, itemnum, menuitem, bgimage, draw, bgimagemask, drawmask, highlightcolor, spumuxdom, spunode, numberofitems, chapternumber, chapterlist): """Draws text and graphics onto a dvd menu, called by createMenu and createChapterMenu""" #Get the XML containing information about this item infoDOM = xml.dom.minidom.parse( os.path.join(getItemTempPath(itemnum),"info.xml") ) #Error out if its the wrong XML if infoDOM.documentElement.tagName != "fileinfo": fatalError("The info.xml file (%s) doesn't look right" % os.path.join(getItemTempPath(itemnum),"info.xml")) #boundarybox holds the max and min dimensions for this item so we can auto build a menu highlight box boundarybox=9999,9999,0,0 wantHighlightBox = True #Loop through all the nodes inside this menu item for node in menuitem.childNodes: #Process each type of item to add it onto the background image if node.nodeName=="graphic": #Overlay graphic image onto background #if this graphic item is a movie thumbnail then we dont process it here if node.attributes["filename"].value == "%movie": #this is a movie item but we must still update the boundary box boundarybox=checkBoundaryBox(boundarybox, node) else: imagefilename = expandItemText(infoDOM,node.attributes["filename"].value, itemnum, page, itemsonthispage, chapternumber, chapterlist) if doesFileExist(imagefilename) == False: if imagefilename == node.attributes["filename"].value: imagefilename = getThemeFile(themeName, node.attributes["filename"].value) if doesFileExist(imagefilename): picture = Image.open(imagefilename,"r").resize((getScaledAttribute(node, "w"), getScaledAttribute(node, "h"))) picture = picture.convert("RGBA") #see if an image mask exists imagemaskfilename = None if node.hasAttribute("mask"): if node.attribute["mask"].value <> "": imagemaskfilename = getThemeFile(themeName, node.attributes["mask"].value) if imagemaskfilename <> None and doesFileExist(imagemaskfilename): maskpicture = Image.open(imagemaskfilename,"r").resize((getScaledAttribute(node, "w"), getScaledAttribute(node, "h"))) maskpicture = maskpicture.convert("RGBA") bgimage.paste(picture, (getScaledAttribute(node, "x"), getScaledAttribute(node, "y")), maskpicture) del maskpicture else: bgimage.paste(picture, (getScaledAttribute(node, "x"), getScaledAttribute(node, "y")), picture) del picture write( "Added image %s" % imagefilename) boundarybox=checkBoundaryBox(boundarybox, node) else: write( "Image file does not exist '%s'" % imagefilename) elif node.nodeName=="text": #Apply some text to the background, including wordwrap if required. text=expandItemText(infoDOM,node.attributes["value"].value, itemnum, page, itemsonthispage,chapternumber,chapterlist) if node.hasAttribute("colour"): color = node.attributes["colour"].value elif node.hasAttribute("color"): color = node.attributes["color"].value else: color = None if text>"": paintText( draw, bgimage, getScaledAttribute(node, "x"), getScaledAttribute(node, "y"), getScaledAttribute(node, "w"), getScaledAttribute(node, "h"), text, themeFonts[node.attributes["font"].value], color, node.attributes["align"].value ) boundarybox=checkBoundaryBox(boundarybox, node) del text elif node.nodeName=="previous": if page>1: #Overlay previous graphic button onto background imagefilename = getThemeFile(themeName, node.attributes["filename"].value) if not doesFileExist(imagefilename): fatalError("Cannot find image for previous button (%s)." % imagefilename) maskimagefilename = getThemeFile(themeName, node.attributes["mask"].value) if not doesFileExist(maskimagefilename): fatalError("Cannot find mask image for previous button (%s)." % maskimagefilename) picture=Image.open(imagefilename,"r").resize((getScaledAttribute(node, "w"), getScaledAttribute(node, "h"))) picture=picture.convert("RGBA") bgimage.paste(picture, (getScaledAttribute(node, "x"), getScaledAttribute(node, "y")), picture) del picture write( "Added previous button image %s" % imagefilename) picture=Image.open(maskimagefilename,"r").resize((getScaledAttribute(node, "w"), getScaledAttribute(node, "h"))) picture=picture.convert("RGBA") bgimagemask.paste(picture, (getScaledAttribute(node, "x"), getScaledAttribute(node, "y")), picture) del picture write( "Added previous button mask image %s" % imagefilename) button = spumuxdom.createElement("button") button.setAttribute("name","previous") button.setAttribute("x0","%s" % getScaledAttribute(node, "x")) button.setAttribute("y0","%s" % getScaledAttribute(node, "y")) button.setAttribute("x1","%s" % (getScaledAttribute(node, "x") + getScaledAttribute(node, "w"))) button.setAttribute("y1","%s" % (getScaledAttribute(node, "y") + getScaledAttribute(node, "h"))) spunode.appendChild(button) elif node.nodeName=="next": if itemnum < numberofitems: #Overlay next graphic button onto background imagefilename = getThemeFile(themeName, node.attributes["filename"].value) if not doesFileExist(imagefilename): fatalError("Cannot find image for next button (%s)." % imagefilename) maskimagefilename = getThemeFile(themeName, node.attributes["mask"].value) if not doesFileExist(maskimagefilename): fatalError("Cannot find mask image for next button (%s)." % maskimagefilename) picture = Image.open(imagefilename,"r").resize((getScaledAttribute(node, "w"), getScaledAttribute(node, "h"))) picture = picture.convert("RGBA") bgimage.paste(picture, (getScaledAttribute(node, "x"), getScaledAttribute(node, "y")), picture) del picture write( "Added next button image %s " % imagefilename) picture=Image.open(maskimagefilename,"r").resize((getScaledAttribute(node, "w"), getScaledAttribute(node, "h"))) picture=picture.convert("RGBA") bgimagemask.paste(picture, (getScaledAttribute(node, "x"), getScaledAttribute(node, "y")), picture) del picture write( "Added next button mask image %s" % imagefilename) button = spumuxdom.createElement("button") button.setAttribute("name","next") button.setAttribute("x0","%s" % getScaledAttribute(node, "x")) button.setAttribute("y0","%s" % getScaledAttribute(node, "y")) button.setAttribute("x1","%s" % (getScaledAttribute(node, "x") + getScaledAttribute(node, "w"))) button.setAttribute("y1","%s" % (getScaledAttribute(node, "y") + getScaledAttribute(node, "h"))) spunode.appendChild(button) elif node.nodeName=="playall": #Overlay playall graphic button onto background imagefilename = getThemeFile(themeName, node.attributes["filename"].value) if not doesFileExist(imagefilename): fatalError("Cannot find image for playall button (%s)." % imagefilename) maskimagefilename = getThemeFile(themeName, node.attributes["mask"].value) if not doesFileExist(maskimagefilename): fatalError("Cannot find mask image for playall button (%s)." % maskimagefilename) picture = Image.open(imagefilename,"r").resize((getScaledAttribute(node, "w"), getScaledAttribute(node, "h"))) picture = picture.convert("RGBA") bgimage.paste(picture, (getScaledAttribute(node, "x"), getScaledAttribute(node, "y")), picture) del picture write( "Added playall button image %s " % imagefilename) picture=Image.open(maskimagefilename,"r").resize((getScaledAttribute(node, "w"), getScaledAttribute(node, "h"))) picture=picture.convert("RGBA") bgimagemask.paste(picture, (getScaledAttribute(node, "x"), getScaledAttribute(node, "y")), picture) del picture write( "Added playall button mask image %s" % imagefilename) button = spumuxdom.createElement("button") button.setAttribute("name","playall") button.setAttribute("x0","%s" % getScaledAttribute(node, "x")) button.setAttribute("y0","%s" % getScaledAttribute(node, "y")) button.setAttribute("x1","%s" % (getScaledAttribute(node, "x") + getScaledAttribute(node, "w"))) button.setAttribute("y1","%s" % (getScaledAttribute(node, "y") + getScaledAttribute(node, "h"))) spunode.appendChild(button) elif node.nodeName=="titlemenu": if itemnum < numberofitems: #Overlay next graphic button onto background imagefilename = getThemeFile(themeName, node.attributes["filename"].value) if not doesFileExist(imagefilename): fatalError("Cannot find image for titlemenu button (%s)." % imagefilename) maskimagefilename = getThemeFile(themeName, node.attributes["mask"].value) if not doesFileExist(maskimagefilename): fatalError("Cannot find mask image for titlemenu button (%s)." % maskimagefilename) picture = Image.open(imagefilename,"r").resize((getScaledAttribute(node, "w"), getScaledAttribute(node, "h"))) picture = picture.convert("RGBA") bgimage.paste(picture, (getScaledAttribute(node, "x"), getScaledAttribute(node, "y")), picture) del picture write( "Added titlemenu button image %s " % imagefilename) picture=Image.open(maskimagefilename,"r").resize((getScaledAttribute(node, "w"), getScaledAttribute(node, "h"))) picture=picture.convert("RGBA") bgimagemask.paste(picture, (getScaledAttribute(node, "x"), getScaledAttribute(node, "y")), picture) del picture write( "Added titlemenu button mask image %s" % imagefilename) button = spumuxdom.createElement("button") button.setAttribute("name","titlemenu") button.setAttribute("x0","%s" % getScaledAttribute(node, "x")) button.setAttribute("y0","%s" % getScaledAttribute(node, "y")) button.setAttribute("x1","%s" % (getScaledAttribute(node, "x") + getScaledAttribute(node, "w"))) button.setAttribute("y1","%s" % (getScaledAttribute(node, "y") + getScaledAttribute(node, "h"))) spunode.appendChild(button) elif node.nodeName=="button": wantHighlightBox = False #Overlay item graphic/text button onto background imagefilename = getThemeFile(themeName, node.attributes["filename"].value) if not doesFileExist(imagefilename): fatalError("Cannot find image for menu button (%s)." % imagefilename) maskimagefilename = getThemeFile(themeName, node.attributes["mask"].value) if not doesFileExist(maskimagefilename): fatalError("Cannot find mask image for menu button (%s)." % maskimagefilename) picture=Image.open(imagefilename,"r").resize((getScaledAttribute(node, "w"), getScaledAttribute(node, "h"))) picture=picture.convert("RGBA") bgimage.paste(picture, (getScaledAttribute(node, "x"), getScaledAttribute(node, "y")), picture) del picture # if we have some text paint that over the image textnode = node.getElementsByTagName("textnormal") if textnode.length > 0: textnode = textnode[0] text=expandItemText(infoDOM,textnode.attributes["value"].value, itemnum, page, itemsonthispage,chapternumber,chapterlist) if node.hasAttribute("colour"): color = node.attributes["colour"].value elif node.hasAttribute("color"): color = node.attributes["color"].value else: color = None if text > "": paintText( draw, bgimage, getScaledAttribute(textnode, "x"), getScaledAttribute(textnode, "y"), getScaledAttribute(textnode, "w"), getScaledAttribute(textnode, "h"), text, themeFonts[textnode.attributes["font"].value], color, textnode.attributes["align"].value ) boundarybox=checkBoundaryBox(boundarybox, node) del text write( "Added button image %s" % imagefilename) picture=Image.open(maskimagefilename,"r").resize((getScaledAttribute(node, "w"), getScaledAttribute(node, "h"))) picture=picture.convert("RGBA") bgimagemask.paste(picture, (getScaledAttribute(node, "x"), getScaledAttribute(node, "y")),picture) #del picture # if we have some text paint that over the image textnode = node.getElementsByTagName("textselected") if textnode.length > 0: textnode = textnode[0] text = expandItemText(infoDOM,textnode.attributes["value"].value, itemnum, page, itemsonthispage,chapternumber,chapterlist) textImage = Image.new("RGBA",picture.size) textDraw = ImageDraw.Draw(textImage) if text > "": paintText(textDraw, textImage, getScaledAttribute(node, "x") - getScaledAttribute(textnode, "x"), getScaledAttribute(node, "y") - getScaledAttribute(textnode, "y"), getScaledAttribute(textnode, "w"), getScaledAttribute(textnode, "h"), text, themeFonts[textnode.attributes["font"].value], "white", textnode.attributes["align"].value ) #convert the RGB image to a 1 bit image (width, height) = textImage.size for y in range(height): for x in range(width): if textImage.getpixel((x,y)) < (100, 100, 100, 255): textImage.putpixel((x,y), (0, 0, 0, 0)) else: textImage.putpixel((x,y), (255, 255, 255, 255)) if textnode.hasAttribute("colour"): color = textnode.attributes["colour"].value elif textnode.hasAttribute("color"): color = textnode.attributes["color"].value else: color = "white" bgimagemask.paste(color, (getScaledAttribute(textnode, "x"), getScaledAttribute(textnode, "y")), textImage) boundarybox=checkBoundaryBox(boundarybox, node) del text, textImage, textDraw del picture elif node.nodeName=="#text" or node.nodeName=="#comment": #Do nothing assert True else: write( "Dont know how to process %s" % node.nodeName) if drawmask == None: return #Draw the mask for this item if wantHighlightBox == True: # Make the boundary box bigger than the content to avoid over writing it boundarybox=boundarybox[0]-1,boundarybox[1]-1,boundarybox[2]+1,boundarybox[3]+1 drawmask.rectangle(boundarybox,outline=highlightcolor) # Draw another line to make the box thicker - PIL does not support linewidth boundarybox=boundarybox[0]-1,boundarybox[1]-1,boundarybox[2]+1,boundarybox[3]+1 drawmask.rectangle(boundarybox,outline=highlightcolor) node = spumuxdom.createElement("button") #Fiddle this for chapter marks.... if chapternumber>0: node.setAttribute("name","%s" % chapternumber) else: node.setAttribute("name","%s" % itemnum) node.setAttribute("x0","%d" % int(boundarybox[0])) node.setAttribute("y0","%d" % int(boundarybox[1])) node.setAttribute("x1","%d" % int(boundarybox[2] + 1)) node.setAttribute("y1","%d" % int(boundarybox[3] + 1)) spunode.appendChild(node) | 3b621623397009015c837134fe56f0ca3f353bed /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/13713/3b621623397009015c837134fe56f0ca3f353bed/mythburn.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3724,
8335,
1180,
12,
2433,
16,
761,
816,
2211,
2433,
16,
761,
2107,
16,
3824,
1726,
16,
7611,
2730,
16,
3724,
16,
7611,
15374,
351,
835,
16,
3724,
4455,
16,
8839,
3266,
16,
1694,
379,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3724,
8335,
1180,
12,
2433,
16,
761,
816,
2211,
2433,
16,
761,
2107,
16,
3824,
1726,
16,
7611,
2730,
16,
3724,
16,
7611,
15374,
351,
835,
16,
3724,
4455,
16,
8839,
3266,
16,
1694,
379,... |
backend, client_name, in_ports, out_ports, debug) | backend, client_name, in_ports, out_ports) | def run(patches, control=None, preprocess=None, postprocess=None, default_patch=0, backend='alsa', client_name='mididings', in_ports=1, out_ports=1, debug=False): s = Setup(patches, control, preprocess, postprocess, default_patch, backend, client_name, in_ports, out_ports, debug) try: s.run() except KeyboardInterrupt: return | b1df734bc550a40eb6b9d7e89dddd54b2ef14a28 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/7503/b1df734bc550a40eb6b9d7e89dddd54b2ef14a28/mididings.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
12,
30278,
16,
3325,
33,
7036,
16,
17942,
33,
7036,
16,
1603,
2567,
33,
7036,
16,
805,
67,
2272,
33,
20,
16,
4221,
2218,
1031,
69,
2187,
1004,
67,
529,
2218,
13138,
350,
899,
2... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
12,
30278,
16,
3325,
33,
7036,
16,
17942,
33,
7036,
16,
1603,
2567,
33,
7036,
16,
805,
67,
2272,
33,
20,
16,
4221,
2218,
1031,
69,
2187,
1004,
67,
529,
2218,
13138,
350,
899,
2... |
authmethod = None for method in preferred_auths: if method in authlist: authmethod = method break if authmethod == AUTH_CRAM_MD5: (code, resp) = self.docmd("AUTH", AUTH_CRAM_MD5) if code == 503: return (code, resp) (code, resp) = self.docmd(encode_cram_md5(resp, user, password)) elif authmethod == AUTH_PLAIN: (code, resp) = self.docmd("AUTH", AUTH_PLAIN + " " + encode_plain(user, password)) elif authmethod == AUTH_LOGIN: (code, resp) = self.docmd("AUTH", "%s %s" % (AUTH_LOGIN, encode_base64(user.encode('ascii'), eol=''))) if code != 334: raise SMTPAuthenticationError(code, resp) (code, resp) = self.docmd(encode_base64(password.encode('ascii'), eol='')) elif authmethod is None: | authlist = [auth for auth in preferred_auths if auth in advertised_authlist] if not authlist: | def encode_plain(user, password): s = "\0%s\0%s" % (user, password) return encode_base64(s.encode('ascii'), eol='') | 1c5471f319a4e5c7bbb56c7ae349b5fe455dc857 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/8546/1c5471f319a4e5c7bbb56c7ae349b5fe455dc857/smtplib.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2017,
67,
7446,
12,
1355,
16,
2201,
4672,
272,
273,
1548,
20,
9,
87,
64,
20,
9,
87,
6,
738,
261,
1355,
16,
2201,
13,
327,
2017,
67,
1969,
1105,
12,
87,
18,
3015,
2668,
9184,
19899,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2017,
67,
7446,
12,
1355,
16,
2201,
4672,
272,
273,
1548,
20,
9,
87,
64,
20,
9,
87,
6,
738,
261,
1355,
16,
2201,
13,
327,
2017,
67,
1969,
1105,
12,
87,
18,
3015,
2668,
9184,
19899,... |
e = reactor.callLater(delay, d.callback, None) | e = reactor.callLater(self.state('tempRecordDelay'), d.callback, None) | def tempCycle(self): """ this function should be called when the temperature recording starts, and will return when it stops. it will loop every either 10s or 10 min and then call record temp. this also checks each time to see if we should continue recording. """ while self.state('recordTemp'): # make the record self.recordTemp() # check if we should still be recording fast. if self.state('recordFast') and self.state('autoRecord') and self.shouldStopFastRecording(): self.stopFastRecording() # check if we should stop recording altogether if self.state('autoRecord') and (yield self.shouldStopRecording()): self.stopRecording() # set up a deferred to run in either 10 min or 10 s if self.state('recordFast'): delay = self.state('fastRecordingRecordDelay') else: delay = self.state('tempRecordDelay') d = defer.Deferred() # we use a blank deferred, so nothing will actually happen when we finish e = reactor.callLater(delay, d.callback, None) self.state('tempDelayedCall', e) # and now, we wait. yield d # note that we can interrupt the waiting by messing with the e object (saved in a state variable) | ecc133167eefb5fe6d901da9bd5da2284e919371 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/7016/ecc133167eefb5fe6d901da9bd5da2284e919371/ADR.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1906,
13279,
12,
2890,
4672,
3536,
333,
445,
1410,
506,
2566,
1347,
326,
12843,
14949,
2542,
16,
471,
903,
327,
1347,
518,
12349,
18,
518,
903,
2798,
3614,
3344,
1728,
87,
578,
1728,
113... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1906,
13279,
12,
2890,
4672,
3536,
333,
445,
1410,
506,
2566,
1347,
326,
12843,
14949,
2542,
16,
471,
903,
327,
1347,
518,
12349,
18,
518,
903,
2798,
3614,
3344,
1728,
87,
578,
1728,
113... |
p -- a prime (default: 0); just the p-new subspace if given | p -- a prime (default: 0); just the $p$-new subspace if given | def dimension_new_cusp_forms_eps(eps, k=2, p=0): """ Dimension of the new subspace (or p-new subspace) of cusp forms of weight k and character eps. INPUT: eps -- a Dirichlet character k -- an integer (ddefault: 2) p -- a prime (default: 0); just the p-new subspace if given OUTPUT: Integer EXAMPLES: sage: G = DirichletGroup(9) sage: eps = G.0^3 sage: eps.conductor() 3 sage: [sage.modular.dims.dimension_new_cusp_forms_eps(eps, k) for k in [2..10]] [0, 0, 0, 2, 0, 2, 0, 2, 0] sage: [sage.modular.dims.dimension_cusp_forms_eps(eps, k) for k in [2..10]] [0, 0, 0, 2, 0, 4, 0, 6, 0] sage: [sage.modular.dims.dimension_new_cusp_forms_eps(eps, k, 3) for k in [2..10]] [0, 0, 0, 2, 0, 2, 0, 2, 0] Double check using modular symbols (independent calculation): sage: [ModularSymbols(eps,k,sign=1).cuspidal_subspace().new_subspace().dimension() for k in [2..10]] [0, 0, 0, 2, 0, 2, 0, 2, 0] sage: [ModularSymbols(eps,k,sign=1).cuspidal_subspace().new_subspace(3).dimension() for k in [2..10]] [0, 0, 0, 2, 0, 2, 0, 2, 0] Another example at level 33: sage: G = DirichletGroup(33) sage: eps = G.1 sage: eps.conductor() 11 sage: [sage.modular.dims.dimension_new_cusp_forms_eps(G.1, k) for k in [2..4]] [0, 4, 0] sage: [sage.modular.dims.dimension_new_cusp_forms_eps(G.1^2, k) for k in [2..4]] [2, 0, 6] sage: [sage.modular.dims.dimension_new_cusp_forms_eps(G.1^2, k, 3) for k in [2..4]] [2, 0, 6] """ if not isinstance(eps, dirichlet.DirichletCharacter): raise TypeError, "eps = (%s) must be a DirichletCharacter"%eps if k < 0: return Z(0) if eps.is_even(): if k % 2 == 1: return Z(0) else: # odd if k % 2 == 0: return Z(0) if k == 0: return Z(0) elif k == 1: raise NotImplementedError, "Computation of dimensions of spaces of weight 1 modular forms not implemented." N = eps.modulus() if p == 0 or N%p != 0 or valuation(eps.conductor(),p) == valuation(N,p): D = [eps.conductor()*d for d in divisors(N//eps.conductor())] return sum([dimension_cusp_forms_eps(eps.restrict(M), k)*mumu(N//M) for M in D]) eps_p = eps.restrict(N//p) old = dimension_cusp_forms(eps_p, k) return dimension_cusp_forms(eps, k) - 2*old | 231200694940d44e11baa531020b171931c64488 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9417/231200694940d44e11baa531020b171931c64488/dims.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4968,
67,
2704,
67,
71,
407,
84,
67,
9741,
67,
13058,
12,
13058,
16,
417,
33,
22,
16,
293,
33,
20,
4672,
3536,
13037,
434,
326,
394,
720,
2981,
261,
280,
293,
17,
2704,
720,
2981,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4968,
67,
2704,
67,
71,
407,
84,
67,
9741,
67,
13058,
12,
13058,
16,
417,
33,
22,
16,
293,
33,
20,
4672,
3536,
13037,
434,
326,
394,
720,
2981,
261,
280,
293,
17,
2704,
720,
2981,
... |
string = 'X' * self.__padding + string + 'X' * self.__padding | string = u'\xa0' * self.__padding + string.replace(u'\xa0', ' ') + u'\xa0' * self.__padding | def getSimilarStrings(self, string): """ Retrieves all strings that have a similarity higher than min_sim with "string" | 24bba282fb5583348dc65021ea3127639cdefbc7 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/12095/24bba282fb5583348dc65021ea3127639cdefbc7/ngram.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1322,
381,
5611,
7957,
12,
2890,
16,
533,
4672,
3536,
20507,
777,
2064,
716,
1240,
279,
16416,
10478,
2353,
1131,
67,
9812,
598,
315,
1080,
6,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1322,
381,
5611,
7957,
12,
2890,
16,
533,
4672,
3536,
20507,
777,
2064,
716,
1240,
279,
16416,
10478,
2353,
1131,
67,
9812,
598,
315,
1080,
6,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
>>> check(y.new('D')); | >>> check(y.new('D').getshape()); >>> check(y.new('D').type()); | def _numarray_tests(): ''' >>> from numpy_ext import * >>> x = new_array() >>> y = x.copy() >>> p = _printer() >>> check = p.check >>> exercise_numarray(x, p) >>> check(y.astype()); >>> check(y.argmax()); >>> check(y.argmax(0)); >>> check(y.argmin()); >>> check(y.argmin(0)); >>> check(y.argsort()); >>> check(y.argsort(1)); >>> y.byteswap(); >>> check(y); >>> check(y.diagonal()); >>> check(y.diagonal(1)); >>> check(y.diagonal(0, 1)); >>> check(y.diagonal(0, 1, 0)); >>> check(y.is_c_array()); >>> check(y.isbyteswapped()); >>> check(y.trace()); >>> check(y.trace(1)); >>> check(y.trace(0, 1)); >>> check(y.trace(0, 1, 0)); >>> check(y.new('D')); >>> y.sort(); >>> check(y); >>> check(y.type()); >>> check(y.factory((1.2, 3.4))); >>> check(y.factory((1.2, 3.4), "Double")); >>> check(y.factory((1.2, 3.4), "Double", (1,2,1))); >>> check(y.factory((1.2, 3.4), "Double", (2,1,1), false)); >>> check(y.factory((1.2, 3.4), "Double", (2,), true, true)); >>> p.results [] >>> del p ''' pass | d78836b82844e73bca9310699a0c2339eb90f5df /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9978/d78836b82844e73bca9310699a0c2339eb90f5df/numpy.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2107,
1126,
67,
16341,
13332,
9163,
4080,
628,
3972,
67,
408,
1930,
380,
4080,
619,
273,
394,
67,
1126,
1435,
4080,
677,
273,
619,
18,
3530,
1435,
4080,
293,
273,
389,
30439,
1435,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2107,
1126,
67,
16341,
13332,
9163,
4080,
628,
3972,
67,
408,
1930,
380,
4080,
619,
273,
394,
67,
1126,
1435,
4080,
677,
273,
619,
18,
3530,
1435,
4080,
293,
273,
389,
30439,
1435,
... |
infoline(' infoline(' infoline(' infoline(' infoline(' | infoline('!!!!! installing without freetype no ttf, sorry!') infoline('!!!!! You need to install a static library version of the freetype2 software') infoline('!!!!! If you need truetype support in renderPM') infoline('!!!!! You may need to edit setup.cfg (win32)') infoline('!!!!! or edit this file to access the library if it is installed') | def libart_version(): K = ('LIBART_MAJOR_VERSION','LIBART_MINOR_VERSION','LIBART_MICRO_VERSION') D = {} for l in open(pjoin(LIBART_DIR,'configure.in'),'r').readlines(): l = l.strip().split('=') if len(l)>1 and l[0].strip() in K: D[l[0].strip()] = l[1].strip() if len(D)==3: break return (platform == 'win32' and '\\"%s\\"' or '"%s"') % '.'.join(map(lambda k,D=D: D.get(k,'?'),K)) | dcb4863a83dff1a07b67404dd90989089cdc1127 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3878/dcb4863a83dff1a07b67404dd90989089cdc1127/setup.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2561,
485,
67,
1589,
13332,
1475,
273,
7707,
14484,
4928,
67,
26976,
916,
67,
5757,
17023,
14484,
4928,
67,
6236,
916,
67,
5757,
17023,
14484,
4928,
67,
22972,
1457,
67,
5757,
6134,
463,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2561,
485,
67,
1589,
13332,
1475,
273,
7707,
14484,
4928,
67,
26976,
916,
67,
5757,
17023,
14484,
4928,
67,
6236,
916,
67,
5757,
17023,
14484,
4928,
67,
22972,
1457,
67,
5757,
6134,
463,
... |
self.tk.call(self._w, 'delete', 'offsprings', entry) | self.tk.call(self._w, 'delete', 'offsprings', entry) | def delete_offsprings(self, entry): | 86af7ef7e3f4448abc89aa941517a84075d99a38 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3187/86af7ef7e3f4448abc89aa941517a84075d99a38/Tix.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1430,
67,
26600,
683,
899,
12,
2890,
16,
1241,
4672,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1430,
67,
26600,
683,
899,
12,
2890,
16,
1241,
4672,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
ErrorMsg = "Pcd '%s.%s' defined in module '%s' is not found in any platform" % (Guid, Name, ModuleName) | ErrorMsg = "Pcd '%s.%s' defined in module '%s' is not found in any platform for %s" % (Guid, Name, ModuleName, Arch) | def FindPcd(self, Arch, ModuleName, Name, Guid, Type, PcdsSet): DatumType = '' Value = '' Token = '' MaxDatumSize = '' SkuInfoList = {} IsOverrided = False IsFoundInDsc = False IsFoundInDec = False # # First get information from platform database # for Dsc in self.Build[Arch].PlatformDatabase.keys(): Pcds = self.Build[Arch].PlatformDatabase[Dsc].Pcds if (Name, Guid) in Pcds: Type = Pcds[(Name, Guid)].Type DatumType = Pcds[(Name, Guid)].DatumType Value = Pcds[(Name, Guid)].DefaultValue Token = Pcds[(Name, Guid)].TokenValue MaxDatumSize = Pcds[(Name, Guid)].MaxDatumSize SkuInfoList = Pcds[(Name, Guid)].SkuInfoList IsOverrided = True IsFoundInDsc = True break | 4cac6be7a76351ac3176cce2f79bfc50ee2ccdb1 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/914/4cac6be7a76351ac3176cce2f79bfc50ee2ccdb1/EdkIIWorkspaceBuild.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4163,
52,
4315,
12,
2890,
16,
16959,
16,
5924,
461,
16,
1770,
16,
611,
1911,
16,
1412,
16,
453,
71,
2377,
694,
4672,
20102,
559,
273,
875,
1445,
273,
875,
3155,
273,
875,
4238,
18268,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4163,
52,
4315,
12,
2890,
16,
16959,
16,
5924,
461,
16,
1770,
16,
611,
1911,
16,
1412,
16,
453,
71,
2377,
694,
4672,
20102,
559,
273,
875,
1445,
273,
875,
3155,
273,
875,
4238,
18268,
... |
if sys.platform == 'win32': pydname = soname.new(purebasename=modname, ext='.pyd') else: pydname = soname.new(purebasename=modname, ext='.so') | from pypy.module.imp.importing import get_so_extension pydname = soname.new(purebasename=modname, ext=get_so_extension(space)) | def compile_module(modname, **kwds): """ Build an extension module and return the filename of the resulting native code file. modname is the name of the module, possibly including dots if it is a module inside a package. Any extra keyword arguments are passed on to ExternalCompilationInfo to build the module (so specify your source with one of those). """ modname = modname.split('.')[-1] eci = ExternalCompilationInfo( export_symbols=['init%s' % (modname,)], include_dirs=api.include_dirs, **kwds ) eci = eci.convert_sources_to_files() dirname = (udir/uniquemodulename('module')).ensure(dir=1) soname = platform.platform.compile( [], eci, outputfilename=str(dirname/modname), standalone=False) if sys.platform == 'win32': pydname = soname.new(purebasename=modname, ext='.pyd') else: pydname = soname.new(purebasename=modname, ext='.so') soname.rename(pydname) return str(pydname) | 6db6e56beb9c0629ec40a5175dda35025d0ea45c /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6934/6db6e56beb9c0629ec40a5175dda35025d0ea45c/test_cpyext.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4074,
67,
2978,
12,
1711,
529,
16,
2826,
25577,
4672,
3536,
3998,
392,
2710,
1605,
471,
327,
326,
1544,
434,
326,
8156,
6448,
981,
585,
18,
225,
16037,
353,
326,
508,
434,
326,
1605,
1... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4074,
67,
2978,
12,
1711,
529,
16,
2826,
25577,
4672,
3536,
3998,
392,
2710,
1605,
471,
327,
326,
1544,
434,
326,
8156,
6448,
981,
585,
18,
225,
16037,
353,
326,
508,
434,
326,
1605,
1... |
def asString(self, format, verbose=None): | def asString(self, format, verbose=None, preview=0): | def asString(self, format, verbose=None): """Converts to an 8 bit string in given format.""" assert format in ['pdf','ps','eps','gif','png','jpg','jpeg','bmp','ppm','tiff','tif','py','pict','pct'], 'Unknown file format "%s"' % format from reportlab import rl_config #verbose = verbose is not None and (verbose,) or (getattr(self,'verbose',verbose),)[0] if format == 'pdf': from reportlab.graphics import renderPDF return renderPDF.drawToString(self) elif format in ['gif','png','tif','jpg','pct','pict','bmp','ppm']: from reportlab.graphics import renderPM return renderPM.drawToString(self, fmt=format) elif format == 'eps': from rlextra.graphics import renderPS_SEP return renderPS_SEP.drawToString(self, preview = getattr(self,'preview',1), showBoundary=getattr(self,'showBorder',rl_config.showBoundary)) elif format == 'ps': from reportlab.graphics import renderPS return renderPS.drawToString(self, showBoundary=getattr(self,'showBorder',rl_config.showBoundary)) elif format == 'py': return self._renderPy() | d4a4f180f66bc31cee317ef36218c5e8b53b5ac0 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3878/d4a4f180f66bc31cee317ef36218c5e8b53b5ac0/shapes.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
16771,
12,
2890,
16,
740,
16,
3988,
33,
7036,
16,
10143,
33,
20,
4672,
3536,
5692,
358,
392,
1725,
2831,
533,
316,
864,
740,
12123,
1815,
740,
316,
10228,
7699,
17023,
1121,
17023,
13058... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
16771,
12,
2890,
16,
740,
16,
3988,
33,
7036,
16,
10143,
33,
20,
4672,
3536,
5692,
358,
392,
1725,
2831,
533,
316,
864,
740,
12123,
1815,
740,
316,
10228,
7699,
17023,
1121,
17023,
13058... |
5.0099999999999998 | 5.00999999999999801 | def jacobi_P(n,a,b,x): """ Returns the Jacobi polynomial $P_n^{(a,b)}(x)$ for integers n > -1 and a and b symbolic or a > -1 and b > -1. The Jacobi polynomials are actually defined for all a and b. However, the Jacobi polynomial weight $(1-x)^a(1+x)^b$ isn't integrable for $a \leq -1$ or $b \leq -1$. EXAMPLES: sage: x = PolynomialRing(QQ, 'x').gen() sage: jacobi_P(2,0,0,x) 3/2*x^2 - 1/2 sage: jacobi_P(2,1,2,1.2) 5.0099999999999998 """ _init() n0 = ZZ(n) # n must be an integer if not(is_Polynomial(x)): return sage_eval(maxima.eval("jacobi_p(%s,%s,%s,%s)"%(n0,a,b,RR(x)))) R = x.parent() y = R.gen() return sage_eval(maxima.eval("jacobi_p(%s,%s,%s,%s)"%(n0,a,b,y)),locals={str(y):y}) | 07b997b0f30cf3292afd8a442aae7d9ff8238bcc /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9417/07b997b0f30cf3292afd8a442aae7d9ff8238bcc/orthogonal_polys.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
20138,
30875,
67,
52,
12,
82,
16,
69,
16,
70,
16,
92,
4672,
3536,
2860,
326,
804,
1077,
30875,
16991,
271,
52,
67,
82,
66,
95,
12,
69,
16,
70,
16869,
12,
92,
21877,
364,
12321,
290... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
20138,
30875,
67,
52,
12,
82,
16,
69,
16,
70,
16,
92,
4672,
3536,
2860,
326,
804,
1077,
30875,
16991,
271,
52,
67,
82,
66,
95,
12,
69,
16,
70,
16869,
12,
92,
21877,
364,
12321,
290... |
"""Parse a sequence of RFC-822 phrases. A phrase is a sequence of words, which are in turn either RFC-822 atoms or quoted-strings. Phrases are canonicalized by squeezing all runs of continuous whitespace into one space. | """Parse a sequence of RFC 2822 phrases. A phrase is a sequence of words, which are in turn either RFC 2822 atoms or quoted-strings. Phrases are canonicalized by squeezing all runs of continuous whitespace into one space. | def getphraselist(self): """Parse a sequence of RFC-822 phrases. | b6f6fad5b96d38181611f28a27ba278b112ea9c9 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3187/b6f6fad5b96d38181611f28a27ba278b112ea9c9/rfc822.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
844,
21636,
5449,
12,
2890,
4672,
3536,
3201,
279,
3102,
434,
8372,
17,
24532,
31413,
18,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
844,
21636,
5449,
12,
2890,
4672,
3536,
3201,
279,
3102,
434,
8372,
17,
24532,
31413,
18,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
cr.execute('select distinct product_id, location_id from stock_move where location_id in %s or location_dest_id in %s', (tuple(ids), tuple(ids))) res_products_by_location = sorted(cr.dictfetchall(), key=itemgetter('location_id')) | cr.execute('select distinct product_id, location_id from stock_move where location_id in %s', (tuple(ids), )) dict1 = cr.dictfetchall() cr.execute('select distinct product_id, location_dest_id as location_id from stock_move where location_dest_id in %s', (tuple(ids), )) dict2 = cr.dictfetchall() res_products_by_location = sorted(dict1+dict2, key=itemgetter('location_id')) | def _product_value(self, cr, uid, ids, field_names, arg, context=None): """Computes stock value (real and virtual) for a product, as well as stock qty (real and virtual). @param field_names: Name of field @return: Dictionary of values """ prod_id = context and context.get('product_id', False) | b1f8aa98da905cef5bbb9c3ccc63cede1165e1a7 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/7397/b1f8aa98da905cef5bbb9c3ccc63cede1165e1a7/stock.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
5896,
67,
1132,
12,
2890,
16,
4422,
16,
4555,
16,
3258,
16,
652,
67,
1973,
16,
1501,
16,
819,
33,
7036,
4672,
3536,
10743,
12480,
460,
261,
7688,
471,
5024,
13,
364,
279,
3017,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
5896,
67,
1132,
12,
2890,
16,
4422,
16,
4555,
16,
3258,
16,
652,
67,
1973,
16,
1501,
16,
819,
33,
7036,
4672,
3536,
10743,
12480,
460,
261,
7688,
471,
5024,
13,
364,
279,
3017,
... |
Makefile.write("all: cm3\n\n") | Makefile.write(".SUFFIXES:\nall: cm3\n\n") | def Boot(): global BuildLocal BuildLocal += " -boot -keep " Version = "1" # This information is duplicated from the config files. # TBD: put it only in one place. # The older bootstraping method does get that right. Compile = ("gcc -g -fPIC " + ({ "AMD64_LINUX" : " -m64 -mno-align-double ", "LINUXLIBC6" : " -m32 -mno-align-double ", "MIPS64_OPENBSD" : " -mabi=64 ", "SPARC32_LINUX" : " -m32 -munaligned-doubles ", "SPARC64_LINUX" : " -m64 -munaligned-doubles ", }.get(Target) or "")) Link = Compile + " -lm -lpthread " SunCompile = "cc -g -mt -xcode=pic32 -xldscope=symbolic" Compile = { "SOLsun" : SunCompile + " -xarch=v8plus ", "SPARC64_SOLARIS" : SunCompile + " -xarch=v9 ", }.get(Target) or Compile SunLink = " -lrt -lm -lnsl -lsocket " Link = { "SOLsun" : Compile + SunLink, "SPARC64_SOLARIS" : Compile + SunLink, }.get(Target) or Link # not in Link Compile += " -c " Assemble = ("as " + ({ "AMD64_LINUX" : " --64 ", "LINUXLIBC6" : " --32 ", "SPARC32_LINUX" : " -32 ", "SPARC64_LINUX" : " -64 ", "SOLsun" : " -s -K PIC -xarch=v8plus ", "SPARC64_SOLARIS" : " -s -K PIC -xarch=v9 ", }.get(Target) or "")) # # squeeze runs of spaces # Compile = re.sub(" +", " ", Compile) Link = re.sub(" +", " ", Link) Assemble = re.sub(" +", " ", Assemble) BootDir = "/cm3-boot-" + Target + "-" + Version P = [ "import-libs", "m3core", "libm3", "sysutils", "m3middle", "m3quake", "m3objfile", "m3linker", "m3back", "m3front", "cm3" ] if Target == "NT386": P += ["mklib"] #DoPackage(["", "realclean"] + P) or sys.exit(1) DoPackage(["", "buildlocal"] + P) or sys.exit(1) try: shutil.rmtree(BootDir) except: pass try: os.mkdir(BootDir) except: pass # # This would probably be a good use of XSL (xml style sheets) # Make = open(os.path.join(BootDir, "make.sh"), "wb") Makefile = open(os.path.join(BootDir, "Makefile"), "wb") UpdateSource = open(os.path.join(BootDir, "updatesource.sh"), "wb") Makefile.write("all: cm3\n\n") for a in [UpdateSource, Make]: a.write("#!/bin/sh\n\nset -e\nset -x\n\n") for a in [Makefile]: a.write("Assemble=" + Assemble + "\nCompile=" + Compile + "\nLink=" + Link + "\n") for a in [Make]: a.write("Assemble=\"" + Assemble + "\"\nCompile=\"" + Compile + "\"\nLink=\"" + Link + "\"\n") for q in P: dir = GetPackagePath(q) for a in os.listdir(os.path.join(Root, dir, Config)): if (a.endswith(".ms") or a.endswith(".is") or a.endswith(".s") or a.endswith(".c")): CopyFile(os.path.join(Root, dir, Config, a), BootDir) Makefile.write("Objects += " + a + ".o\n" + a + ".o: " + a + "\n\t") if a.endswith(".c"): Command = "Compile" else: Command = "Assemble" for b in [Make, Makefile]: b.write("${" + Command + "} " + a + " -o " + a + ".o\n") if a.endswith(".h"): CopyFile(os.path.join(Root, dir, Config, a), BootDir) Makefile.write("cm3: $(Objects)\n\t") for a in [Make, Makefile]: a.write("$(Link) -o cm3 *.o\n") Common = "Common" for a in [ # # Add to this list as needed. # Adding more than necessary is ok -- assume the target system has no changes, # so we can replace whatever is there. # "m3-libs/libm3/src/os/POSIX/OSConfigPosix.m3", "m3-libs/libm3/src/random/m3makefile", "m3-libs/m3core/src/m3makefile", "m3-libs/m3core/src/C/m3makefile", "m3-libs/m3core/src/C/" + Target + "/m3makefile", "m3-libs/m3core/src/C/" + Common + "/m3makefile", "m3-libs/m3core/src/Csupport/m3makefile", "m3-libs/m3core/src/float/m3makefile", "m3-libs/m3core/src/runtime/m3makefile", "m3-libs/m3core/src/runtime/common/Compiler.tmpl", "m3-libs/m3core/src/runtime/" + Target + "/m3makefile", "m3-libs/m3core/src/runtime/" + Target + "/RTMachine.m3", "m3-libs/m3core/src/runtime/" + Target + "/RTSignal.m3", "m3-libs/m3core/src/runtime/" + Target + "/RTThread.m3", "m3-libs/m3core/src/text/TextLiteral.i3", "m3-libs/m3core/src/thread/m3makefile", "m3-libs/m3core/src/time/POSIX/m3makefile", "m3-libs/m3core/src/unix/m3makefile", "m3-libs/m3core/src/unix/linux-32/m3makefile", "m3-libs/m3core/src/unix/linux-64/m3makefile", "m3-libs/m3core/src/unix/Common/m3makefile", "m3-libs/m3core/src/unix/Common/m3unix.h", "m3-libs/m3core/src/unix/Common/Udir.i3", "m3-libs/m3core/src/unix/Common/UdirC.c", "m3-libs/m3core/src/unix/Common/Ustat.i3", "m3-libs/m3core/src/unix/Common/UstatC.c", "m3-libs/m3core/src/unix/Common/UtimeC.c", "m3-sys/cminstall/src/config-no-install/SOLgnu", "m3-sys/cminstall/src/config-no-install/SOLsun", "m3-sys/cminstall/src/config-no-install/Solaris.common", "m3-sys/cminstall/src/config-no-install/Unix.common", "m3-sys/cminstall/src/config-no-install/cm3cfg.common", "m3-sys/cminstall/src/config-no-install/" + Target, "m3-sys/m3cc/src/m3makefile", "m3-sys/m3cc/src/gcc/m3cg/parse.c", "m3-sys/m3middle/src/Target.i3", "m3-sys/m3middle/src/Target.m3", "scripts/python/pylib.py", "m3-libs/m3core/src/C/" + Target + "/Csetjmp.i3", "m3-libs/m3core/src/C/" + Target + "/Csignal.i3", "m3-libs/m3core/src/C/" + Target + "/Cstdio.i3", "m3-libs/m3core/src/C/" + Target + "/Cstring.i3", "m3-libs/m3core/src/C/" + Target + "/m3makefile", "m3-libs/m3core/src/C/" + Common + "/Csetjmp.i3", "m3-libs/m3core/src/C/" + Common + "/Csignal.i3", "m3-libs/m3core/src/C/" + Common + "/Cstdio.i3", "m3-libs/m3core/src/C/" + Common + "/Cstring.i3", "m3-libs/m3core/src/C/" + Common + "/m3makefile", "m3-libs/m3core/src/C/" + Common + "/Udir.i3", "m3-libs/m3core/src/C/" + Common + "/UdirC.c", "m3-libs/m3core/src/C/" + Common + "/Uerror.i3.cpp", "m3-libs/m3core/src/C/" + Common + "/UnixC.c", "m3-libs/m3core/src/C/" + Common + "/Ustat.i3", "m3-libs/m3core/src/C/" + Common + "/UstatC.c", "m3-libs/m3core/src/C/" + Common + "/UtimeC.c", "m3-libs/m3core/src/C/" + Common + "/Uutsname.i3.cpp", ]: source = os.path.join(Root, a) if FileExists(source): name = GetLastPathElement(a) reldir = RemoveLastPathElement(a) destdir = os.path.join(BootDir, reldir) dest = os.path.join(destdir, name) try: os.makedirs(destdir) except: pass CopyFile(source, dest) for b in [UpdateSource, Make]: b.write("mkdir -p /dev2/cm3/" + reldir + "\n") b.write("cp " + a + " /dev2/cm3/" + a + "\n") for a in [UpdateSource, Make, Makefile]: a.close() os.chdir("/") _MakeArchive(BootDir[1:]) | 510ccd943f091afb67ca192fe4502fb96c7a185b /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/9328/510ccd943f091afb67ca192fe4502fb96c7a185b/pylib.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
26254,
13332,
225,
2552,
3998,
2042,
3998,
2042,
1011,
315,
300,
7137,
300,
10102,
315,
225,
4049,
273,
315,
21,
6,
225,
468,
1220,
1779,
353,
16975,
628,
326,
642,
1390,
18,
468,
399,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
26254,
13332,
225,
2552,
3998,
2042,
3998,
2042,
1011,
315,
300,
7137,
300,
10102,
315,
225,
4049,
273,
315,
21,
6,
225,
468,
1220,
1779,
353,
16975,
628,
326,
642,
1390,
18,
468,
399,
... |
self.key_file = x509.get('key_file') self.cert_file = x509.get('cert_file') | self.key_file = key_file self.cert_file = cert_file | def __init__(self, host='', port=None, **x509): # provide a default host, pass the X509 cert info | d46aa37d35811a37397104f02074c8a44e7dbec1 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8546/d46aa37d35811a37397104f02074c8a44e7dbec1/httplib.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
1479,
2218,
2187,
1756,
33,
7036,
16,
2826,
92,
5995,
4672,
468,
5615,
279,
805,
1479,
16,
1342,
326,
1139,
5995,
3320,
1123,
2,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
1479,
2218,
2187,
1756,
33,
7036,
16,
2826,
92,
5995,
4672,
468,
5615,
279,
805,
1479,
16,
1342,
326,
1139,
5995,
3320,
1123,
2,
-100,
-100,
-100,
-100,
... |
dumpFile(name, tag, size, fcf, | s += dumpFile(name, tag, size, fcf, | def sizeCallback(dumpString, name, tag, size): hdr = struct.pack("!HHIH%ds%ds" % (len(name), len(tag)), SUBFILE_MAGIC, len(name), size, len(tag), name, tag) dumpString(hdr) | b4704a77be9e37187ce9af73a9dd23d85d80751e /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8747/b4704a77be9e37187ce9af73a9dd23d85d80751e/filecontainer.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
963,
2428,
12,
8481,
780,
16,
508,
16,
1047,
16,
963,
4672,
7723,
273,
1958,
18,
2920,
2932,
5,
17500,
45,
44,
9,
2377,
9,
2377,
6,
738,
261,
1897,
12,
529,
3631,
562,
12,
2692,
13... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
963,
2428,
12,
8481,
780,
16,
508,
16,
1047,
16,
963,
4672,
7723,
273,
1958,
18,
2920,
2932,
5,
17500,
45,
44,
9,
2377,
9,
2377,
6,
738,
261,
1897,
12,
529,
3631,
562,
12,
2692,
13... |
user =models.OneToOneField(User, primary_key=True) warehouses =models.ForeignKey(DispatchPoint, blank=True,null=True) receptionPoints =models.ForeignKey(ReceptionPoint, blank=True,null=True) isCompasUser =models.BooleanField() isDispatcher =models.BooleanField() isReciever =models.BooleanField() compasUser =models.OneToOneField(EpicPerson, blank=True,null=True) | user =models.OneToOneField(User, primary_key=True) warehouses =models.ForeignKey(DispatchPoint, blank=True,null=True) receptionPoints =models.ForeignKey(ReceptionPoint, blank=True,null=True) isCompasUser =models.BooleanField() isDispatcher =models.BooleanField() isReciever =models.BooleanField() compasUser =models.OneToOneField(EpicPerson, blank=True,null=True) superUser =models.BooleanField() | def __unicode__(self): return self.LOC_NAME + ' ' + self.CONSEGNEE_CODE + ' - ' + self.CONSEGNEE_NAME | 64a9e5d268ba2528ef04dcf2d1a9058cc8d90f71 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12559/64a9e5d268ba2528ef04dcf2d1a9058cc8d90f71/models.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
225,
1001,
9124,
972,
12,
2890,
4672,
327,
365,
18,
29450,
67,
1985,
397,
296,
296,
397,
365,
18,
2248,
1090,
43,
5407,
41,
67,
5572,
397,
296,
300,
296,
397,
365,
18,
2248,
1090,
43... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
225,
1001,
9124,
972,
12,
2890,
4672,
327,
365,
18,
29450,
67,
1985,
397,
296,
296,
397,
365,
18,
2248,
1090,
43,
5407,
41,
67,
5572,
397,
296,
300,
296,
397,
365,
18,
2248,
1090,
43... |
ObjectField('visible_ids', | BooleanField('visible_ids', | def modify_fti(fti): # fix up the factory type information generated by archetypes fti['global_allow'] = 0 # only allow Members to be added where explicitly allowed | 311f261570e72b35c95f1fa2cd3c6d0df8f0b05f /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/1807/311f261570e72b35c95f1fa2cd3c6d0df8f0b05f/Member.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
5612,
67,
1222,
77,
12,
1222,
77,
4672,
468,
2917,
731,
326,
3272,
618,
1779,
4374,
635,
6637,
15180,
284,
9096,
3292,
6347,
67,
5965,
3546,
273,
374,
225,
468,
1338,
1699,
17524,
358,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
5612,
67,
1222,
77,
12,
1222,
77,
4672,
468,
2917,
731,
326,
3272,
618,
1779,
4374,
635,
6637,
15180,
284,
9096,
3292,
6347,
67,
5965,
3546,
273,
374,
225,
468,
1338,
1699,
17524,
358,
... |
entry_name = self.get_string_u_at_rva(ustr_offset) | entry_name = UnicodeStringWrapperPostProcessor(self, ustr_offset) strings_to_postprocess.append(entry_name) | def parse_resources_directory(self, rva, size=0, base_rva = None, level = 0): """Parse the resources directory. Given the rva of the resources directory, it will process all its entries. The root will have the corresponding member of its structure, IMAGE_RESOURCE_DIRECTORY plus 'entries', a list of all the entries in the directory. Those entries will have, correspondingly, all the structure's members (IMAGE_RESOURCE_DIRECTORY_ENTRY) and an additional one, "directory", pointing to the IMAGE_RESOURCE_DIRECTORY structure representing upper layers of the tree. This one will also have an 'entries' attribute, pointing to the 3rd, and last, level. Another directory with more entries. Those last entries will have a new atribute (both 'leaf' or 'data_entry' can be used to access it). This structure finally points to the resource data. All the members of this structure, IMAGE_RESOURCE_DATA_ENTRY, are available as its attributes. """ # OC Patch: original_rva = rva if base_rva is None: base_rva = rva resources_section = self.get_section_by_rva(rva) try: # If the RVA is invalid all would blow up. Some EXEs seem to be # specially nasty and have an invalid RVA. data = self.get_data(rva) except PEFormatError, e: self.__warnings.append( 'Invalid resources directory. Can\'t read ' + 'directory data at RVA: 0x%x' % rva) return None # Get the resource directory structure, that is, the header # of the table preceding the actual entries # resource_dir = self.__unpack_data__( self.__IMAGE_RESOURCE_DIRECTORY_format__, data, file_offset = self.get_offset_from_rva(rva) ) if resource_dir is None: # If can't parse resources directory then silently return. # This directory does not necesarily have to be valid to # still have a valid PE file self.__warnings.append( 'Invalid resources directory. Can\'t parse ' + 'directory data at RVA: 0x%x' % rva) return None dir_entries = [] # Advance the rva to the positon immediately following the directory # table header and pointing to the first entry in the table # rva += resource_dir.sizeof() number_of_entries = ( resource_dir.NumberOfNamedEntries + resource_dir.NumberOfIdEntries ) for idx in range(number_of_entries): res = self.parse_resource_entry(rva) if res is None: self.__warnings.append( 'Error parsing the resources directory, ' + 'Entry %d is invalid, RVA = 0x%x. ' % (idx, rva) ) break | 89f19a90b425f0951e2a3860a3f90834315ae393 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/4177/89f19a90b425f0951e2a3860a3f90834315ae393/pefile.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1109,
67,
4683,
67,
5149,
12,
2890,
16,
5633,
69,
16,
963,
33,
20,
16,
1026,
67,
4962,
69,
273,
599,
16,
1801,
273,
374,
4672,
3536,
3201,
326,
2703,
1867,
18,
225,
16803,
326,
5633,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1109,
67,
4683,
67,
5149,
12,
2890,
16,
5633,
69,
16,
963,
33,
20,
16,
1026,
67,
4962,
69,
273,
599,
16,
1801,
273,
374,
4672,
3536,
3201,
326,
2703,
1867,
18,
225,
16803,
326,
5633,... |
setup_tool.setImportContext('profile-eXtremeManagement:default' | setup_tool.setImportContext('profile-eXtremeManagement:default') | def applyGenericSetupProfile(portal, out): setup_tool = getToolByName(portal, 'portal_setup') setup_tool.setImportContext('profile-eXtremeManagement:default' print >> out, "Applied the generic setup profile for eXtremeManagement" setup_tool.runAllImportSteps(purge_old=False) setup_tool.setImportContext('profile-CMFPlone:plone') | b1a732606c77ae83dc7ca2598e51b8de6f799ee8 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/10431/b1a732606c77ae83dc7ca2598e51b8de6f799ee8/AppInstall.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2230,
7014,
7365,
4029,
12,
24386,
16,
596,
4672,
3875,
67,
6738,
273,
336,
6364,
5911,
12,
24386,
16,
296,
24386,
67,
8401,
6134,
3875,
67,
6738,
18,
542,
5010,
1042,
2668,
5040,
17,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2230,
7014,
7365,
4029,
12,
24386,
16,
596,
4672,
3875,
67,
6738,
273,
336,
6364,
5911,
12,
24386,
16,
296,
24386,
67,
8401,
6134,
3875,
67,
6738,
18,
542,
5010,
1042,
2668,
5040,
17,
... |
self._menubar_menu.set_paths(self._get_document_paths()) self._determine_menu_sensitivity(self._get_document_paths()) | self._menubar_menu.set_paths([self._get_document_path()]) self._determine_menu_sensitivity([self._get_document_path()]) | def update_ui(self): self.update_base_dir() | f536a3b3d5a07da5e3600d24ae1bb1ebfa1cbd71 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/5119/f536a3b3d5a07da5e3600d24ae1bb1ebfa1cbd71/rabbitvcs-plugin.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1089,
67,
4881,
12,
2890,
4672,
365,
18,
2725,
67,
1969,
67,
1214,
1435,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1089,
67,
4881,
12,
2890,
4672,
365,
18,
2725,
67,
1969,
67,
1214,
1435,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-1... |
evtype = struct.unpack("!H", body[:2]) | evtype, = struct.unpack("!H", body[:2]) | def unpack_event(body): if len(body)<2: raise ProtocolError("EVENT body too short.") evtype = struct.unpack("!H", body[:2]) body = body[2:] if evtype == EVENT_TYPE.CIRCUITSTATUS: if len(body)<5: raise ProtocolError("CIRCUITSTATUS event too short.") status,ident = struct.unpack("!BL", body[:5]) path = _unterminate(body[5:]).split(",") args = status, ident, path elif evtype == EVENT_TYPE.STREAMSTATUS: if len(body)<5: raise ProtocolError("CIRCUITSTATUS event too short.") status,ident = struct.unpack("!BL", body[:5]) target = _unterminate(body[5:]) args = status, ident, target elif evtype == EVENT_TYPE.ORCONNSTATUS: if len(body)<2: raise ProtocolError("CIRCUITSTATUS event too short.") status = ord(body[0]) target = _unterminate(body[1:]) args = status, target elif evtype == EVENT_TYPE.BANDWIDTH: if len(body)<8: raise ProtocolError("BANDWIDTH event too short.") read, written = struct.unpack("!LL",body[:8]) args = read, written elif evtype == EVENT_TYPE.WARN: args = (_unterminate(body),) elif evtype == EVENT_TYPE.NEWDESC: args = (_unterminate(body).split(","),) else: args = (body,) return evtype, args | 13bd755ae65c5978243ad6e7028a478dc51b7547 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/7483/13bd755ae65c5978243ad6e7028a478dc51b7547/TorControl.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
6167,
67,
2575,
12,
3432,
4672,
309,
562,
12,
3432,
13,
32,
22,
30,
1002,
4547,
668,
2932,
10454,
1417,
4885,
3025,
1199,
13,
2113,
723,
16,
273,
1958,
18,
17309,
2932,
5,
44,
3113,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
6167,
67,
2575,
12,
3432,
4672,
309,
562,
12,
3432,
13,
32,
22,
30,
1002,
4547,
668,
2932,
10454,
1417,
4885,
3025,
1199,
13,
2113,
723,
16,
273,
1958,
18,
17309,
2932,
5,
44,
3113,
... |
Return the ideal with given gens over `\mathbf{Z}`. | Return the ideal with given gens over `\ZZ`. | def left_ideal(self, gens, check=True): r""" Return the ideal with given gens over `\mathbf{Z}`. | 1ec65dfa0c3aab294b8ecca3f5f5493e154b7172 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/9417/1ec65dfa0c3aab294b8ecca3f5f5493e154b7172/quaternion_algebra.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2002,
67,
831,
287,
12,
2890,
16,
314,
773,
16,
866,
33,
5510,
4672,
436,
8395,
2000,
326,
23349,
598,
864,
314,
773,
1879,
1375,
64,
15949,
17156,
95,
62,
97,
8338,
2,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2002,
67,
831,
287,
12,
2890,
16,
314,
773,
16,
866,
33,
5510,
4672,
436,
8395,
2000,
326,
23349,
598,
864,
314,
773,
1879,
1375,
64,
15949,
17156,
95,
62,
97,
8338,
2,
-100,
-100,
-... |
EXAMPLES:: | EXAMPLES: A generator object for binary sequencences of length 3, listed:: | def search_forest_iterator(roots, childs): """ INPUT: - ``roots``: a list (or iterable) - ``childs``: a function returning a list (or iterable) Returns an iterator on the nodes of the forest having the given roots, and where ``child(x)`` returns the childs of the node ``x`` of the forest. EXAMPLES:: sage: from sage.combinat.backtrack import search_forest_iterator sage: list(search_forest_iterator([[]], lambda l: [l+[0], l+[1]] if len(l) < 3 else [])) [[], [0], [0, 0], [0, 0, 0], [0, 0, 1], [0, 1], [0, 1, 0], [0, 1, 1], [1], [1, 0], [1, 0, 0], [1, 0, 1], [1, 1], [1, 1, 0], [1, 1, 1]] """ #Invariant: stack[i] contains an iterator for the siblings of the i-th node of the current branch stack = [iter(roots)] while len(stack) > 0: # Try to get the next node at this depth try: node = stack[-1].next() except StopIteration: #If there are no more, go back up the tree # We also need to check if we've exhausted all # possibilities stack.pop() continue yield node stack.append( iter(childs(node)) ) | d1bfa2893dda8805bd4c6dd2e7784bcbd0f5a99d /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/9417/d1bfa2893dda8805bd4c6dd2e7784bcbd0f5a99d/backtrack.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1623,
67,
1405,
334,
67,
9838,
12,
22078,
16,
21619,
4672,
3536,
12943,
30,
225,
300,
12176,
22078,
10335,
30,
279,
666,
261,
280,
6087,
13,
225,
300,
12176,
3624,
87,
10335,
30,
279,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1623,
67,
1405,
334,
67,
9838,
12,
22078,
16,
21619,
4672,
3536,
12943,
30,
225,
300,
12176,
22078,
10335,
30,
279,
666,
261,
280,
6087,
13,
225,
300,
12176,
3624,
87,
10335,
30,
279,
... |
print ' rrd.update( %s )' % (u,) | if debug: log( 'rrd.update( %s )' % (u,) ) | def deliver(self, sub, msg, insec, rock): | dbd18abb3b7486c00964cbd1506a90b2bd96974e /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3643/dbd18abb3b7486c00964cbd1506a90b2bd96974e/elvinrrd.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
11795,
12,
2890,
16,
720,
16,
1234,
16,
316,
3321,
16,
23486,
4672,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
11795,
12,
2890,
16,
720,
16,
1234,
16,
316,
3321,
16,
23486,
4672,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-... |
inClass, inPublic, inProtected, inPrivate = True, False, False, False | inClass, inStruct, inPublic, inProtected, inPrivate = True, False, False, False, False m = re.search("^\s*struct\s+(\w+)\s*:?\s+", line.stripped) if m: structName = m.group(1) inClass, inStruct, inPublic, inProtected, inPrivate = False, True, False, False, False | def flagLines(lines): inClass = False inPublic = False inProtected = False inPrivate = False nNested = 0 className = "" for line in lines: # class information m = re.search("^\s*class\s+(\w+)\s*:?\s+", line.stripped) if m: className = m.group(1) inClass, inPublic, inProtected, inPrivate = True, False, False, False if (inClass and re.search("^\s*public:", line.stripped)): inClass, inPublic, inProtected, inPrivate = True, True, False, False if (inClass and re.search("^\s*protected:", line.stripped)): inClass, inPublic, inProtected, inPrivate = True, False, True, False if (inClass and re.search("^\s*private:", line.stripped)): inClass, inPublic, inProtected, inPrivate = True, False, False, True if (inClass and re.search("^};\s*", line.stripped)): inClass, inPublic, inProtected, inPrivate = False, False, False, False line.inClass = inClass line.inPublic = inPublic line.inProtected = inProtected line.inPrivate = inPrivate if inClass: line.className = className if (re.search("\{", line.stripped)): nNested += 1 if (re.search("\}", line.stripped)): nNested -= 1 line.nNested = nNested #char = "" #if inClass: char += "C" #if inPrivate: char += "V" #if inProtected: char += "T" #if inPublic: char += "B" #print char + " " + line.stripped, # variables, functions, templates line.variableNames = getVariableNames(line.stripped) line.functionNames = getFunctionNames(line.stripped) line.templateNames = getTemplateNames(line.stripped) return lines | 00f0224f7ac035382ad33b0bc56b5770db79b0ba /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/6529/00f0224f7ac035382ad33b0bc56b5770db79b0ba/style.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2982,
5763,
12,
3548,
4672,
225,
316,
797,
377,
273,
1083,
316,
4782,
565,
273,
1083,
316,
15933,
273,
1083,
316,
6014,
282,
273,
1083,
225,
290,
8649,
273,
374,
2658,
273,
1408,
225,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2982,
5763,
12,
3548,
4672,
225,
316,
797,
377,
273,
1083,
316,
4782,
565,
273,
1083,
316,
15933,
273,
1083,
316,
6014,
282,
273,
1083,
225,
290,
8649,
273,
374,
2658,
273,
1408,
225,
... |
if 'printdisabled' in collections: | if 'printdisabled' in collections or 'lendinglibrary' in collections: | def hide_books(start): mend = [] fix_works = set() db_iter = db.query("select identifier, collection, updated from metadata where (noindex is not null or curatestate='dark') and mediatype='texts' and scandate is not null and updated > $start order by updated", {'start': start}) for row in db_iter: ia = row.identifier if row.collection: collections = set(i.lower().strip() for i in row.collection.split(';')) if 'printdisabled' in collections: continue print `ia`, row.updated for eq in query({'type': '/type/edition', 'ocaid': ia}): print eq['key'] e = ol.get(eq['key']) if 'ocaid' not in e: continue if 'works' in e: fix_works.update(e['works']) print e['key'], `e.get('title', None)` del e['ocaid'] mend.append(e) print 'removing links from %d editions' % len(mend) print ol.save_many(mend, 'remove link') requests = [] for wkey in fix_works: requests += update_work(withKey(wkey)) if fix_works: solr_update(requests + ['<commit/>'], debug=True) | 0ca2ba9abef7bc4e7a966877772b3d83199b1c48 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3913/0ca2ba9abef7bc4e7a966877772b3d83199b1c48/load_scribe.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
6853,
67,
12567,
12,
1937,
4672,
312,
409,
273,
5378,
2917,
67,
18597,
273,
444,
1435,
1319,
67,
2165,
273,
1319,
18,
2271,
2932,
4025,
2756,
16,
1849,
16,
3526,
628,
1982,
1625,
261,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
6853,
67,
12567,
12,
1937,
4672,
312,
409,
273,
5378,
2917,
67,
18597,
273,
444,
1435,
1319,
67,
2165,
273,
1319,
18,
2271,
2932,
4025,
2756,
16,
1849,
16,
3526,
628,
1982,
1625,
261,
... |
"Daemon process control library and tools for Unix-bases systems", | "Daemon process control library and tools for Unix-based systems", | def read(*rnames): return open(os.path.join(os.path.dirname(__file__), *rnames)).read() | 221cc2da12cffc249d77d8b9106b0906f707daea /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/9782/221cc2da12cffc249d77d8b9106b0906f707daea/setup.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
855,
30857,
86,
1973,
4672,
327,
1696,
12,
538,
18,
803,
18,
5701,
12,
538,
18,
803,
18,
12287,
12,
972,
768,
972,
3631,
380,
86,
1973,
13,
2934,
896,
1435,
225,
2,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
855,
30857,
86,
1973,
4672,
327,
1696,
12,
538,
18,
803,
18,
5701,
12,
538,
18,
803,
18,
12287,
12,
972,
768,
972,
3631,
380,
86,
1973,
13,
2934,
896,
1435,
225,
2,
-100,
-100,
-100,... |
DebugPrint(4, "DEBUG: call readCertinfo") | DebugPrint(4, "DEBUG: call readCertinfo(" + str(localJobId) + ", " + str(probeName) + ")") | def verifyFromCertInfo(xmlDoc, userIdentityNode, namespace, prefix): " Use localJobID and probeName to find cert info file and insert info into XML record" # Collect data needed by certinfo reader DebugPrint(4, "DEBUG: Get JobIdentity") JobIdentityNode = GetNode(xmlDoc.getElementsByTagNameNS(namespace, 'JobIdentity')) if JobIdentityNode == None: return DebugPrint(4, "DEBUG: Get JobIdentity: OK") localJobId = GetNodeData(JobIdentityNode.getElementsByTagNameNS(namespace, 'LocalJobId')) DebugPrint(4, "DEBUG: Get localJobId: ", localJobId) usageRecord = userIdentityNode.parentNode probeName = GetNodeData(usageRecord.getElementsByTagNameNS(namespace, 'ProbeName')) DebugPrint(4, "DEBUG: Get probeName: ", probeName) # Read certinfo DebugPrint(4, "DEBUG: call readCertinfo") certInfo = readCertInfo(localJobId, probeName) DebugPrint(4, "DEBUG: call readCertinfo: OK") DebugPrint(4, "DEBUG: certInfo: " + str(certInfo)) if certInfo == None or (not certInfo.has_key('DN')) or (not certInfo['DN']): DebugPrint(4, "Returning without processing certInfo") return # Use certinfo DebugPrint(4, "DEBUG: fixing DN") certInfo['DN'] = FixDN(certInfo['DN']) # "Standard" slash format # First, find a KeyInfo node if it is there DebugPrint(4, "DEBUG: looking for KeyInfo node") keyInfoNS = 'http://www.w3.org/2000/09/xmldsig#'; keyInfoNode = GetNode(userIdentityNode.getElementsByTagNameNS(keyInfoNS, 'KeyInfo')) DNnode = GetNode(userIdentityNode.getElementsByTagNameNS(namespace, 'DN')) if DNnode and DN.firstChild: # Override DebugPrint(4, "DEBUG: overriding DN from certInfo") DN.firstChild.data = certInfo['DN'] else: DebugPrint(4, "DEBUG: creating fresh DN node") if not DNnode: DNnode = xmlDoc.createElementNS(namespace, 'DN') textNode = xmlDoc.createTextNode(certInfo['DN']) DNnode.appendChild(textNode) if not DNnode.parentNode: userIdentityNode.appendChild(DNnode) DebugPrint(4, "DEBUG: creating fresh DN node: OK") # Return VO information for insertion in a common place. DebugPrint(4, "DEBUG: returning VOName " + certInfo['FQAN'] + " and ReportableVOName " + certInfo['VO']) return { 'VOName': certInfo['FQAN'], 'ReportableVOName': certInfo['VO']} | b5229961d4b4d9734248596436019a238ceb3c4b /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/637/b5229961d4b4d9734248596436019a238ceb3c4b/Gratia.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3929,
1265,
5461,
966,
12,
2902,
1759,
16,
729,
4334,
907,
16,
1981,
16,
1633,
4672,
315,
2672,
1191,
2278,
734,
471,
10695,
461,
358,
1104,
3320,
1123,
585,
471,
2243,
1123,
1368,
3167,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3929,
1265,
5461,
966,
12,
2902,
1759,
16,
729,
4334,
907,
16,
1981,
16,
1633,
4672,
315,
2672,
1191,
2278,
734,
471,
10695,
461,
358,
1104,
3320,
1123,
585,
471,
2243,
1123,
1368,
3167,... |
0.63496391478473613 + 1.2984575814159773*I | 0.634963914784736 + 1.29845758141597*I | def sinh(self): """ EXAMPLES: sage: (1+I).sinh() 0.63496391478473613 + 1.2984575814159773*I """ return self.parent()(self._pari_().sinh()) | 293427b47ee2bb564117092226099ed44acd171f /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9417/293427b47ee2bb564117092226099ed44acd171f/complex_number.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
5367,
76,
12,
2890,
4672,
3536,
5675,
8900,
11386,
30,
272,
410,
30,
261,
21,
15,
45,
2934,
21861,
76,
1435,
374,
18,
4449,
7616,
21607,
29488,
5193,
27,
5718,
397,
404,
18,
5540,
5193... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
5367,
76,
12,
2890,
4672,
3536,
5675,
8900,
11386,
30,
272,
410,
30,
261,
21,
15,
45,
2934,
21861,
76,
1435,
374,
18,
4449,
7616,
21607,
29488,
5193,
27,
5718,
397,
404,
18,
5540,
5193... |
unsafe = 1 break except AttributeError: fixed[len(fixed) - 1] += rebuilt_lines[x] continue except IndexError: unsafe = 1 if unsafe: return rebuilt_lines else: return fixed | previous_line = fixed[len(fixed) - 1].rstrip("\n") movesect = current_line[0:m_next_mark.end()] + "\n" previous_line += movesect fixed[len(fixed) - 1] = previous_line current_line = current_line[m_next_mark.end():] m_next_mark = p_refmarker.search(current_line) if len(current_line.strip()) > 0: previous_line = fixed[len(fixed) - 1].rstrip("\n") movesect = current_line previous_line += movesect fixed[len(fixed) - 1] = previous_line return fixed | def correct_rebuilt_lines(rebuilt_lines, p_refmarker): """Try to correct any cases where a reference line has been incorrectly split based upon a wrong numeration marker. That is to say, given the following situation: [1] Smith, J blah blah [2] Brown, N blah blah see reference [56] for more info [3] Wills, A blah blah ... The first part of the 3rd line clearly belongs with line 2. This function will try to fix this situation, to have the following situation: [1] Smith, J blah blah [2] Brown, N blah blah see reference [56] for more info [3] Wills, A blah blah If it cannot correctly guess the correct break-point in such a line, it will give up and the original list of reference lines will be returned. @param rebuilt_lines: (list) the rebuilt reference lines @param p_refmarker: (compiled regex pattern object) the pattern used to match regex line numeration markers. **MUST HAVE A GROUP 'marknum' to encapsulate the mark number!** (e.g. r'\[(?P<marknum>\d+)\]') @return: (list) of strings. If necessary, the corrected reference lines. Else the orginal 'rebuilt' lines. """ fixed = [] unsafe = 0 try: m = p_refmarker.match(rebuilt_lines[0]) last_marknum = int(m.group("marknum")) if last_marknum != 1: ## Even the first mark isnt 1 - probaby too dangerous to try to repair return rebuilt_lines except (IndexError, AttributeError, ValueError): ## Sometihng went wrong. Either no references, not a numbered line marker (int() failed), or ## no reference line marker (NoneType was passed). In any case, unable to test for correct ## reference line numberring - just return the lines as they were. return rebuilt_lines ## Loop through each line in "rebuilt_lines" and test the mark at the beginning. ## If current-line-mark = previous-line-mark + 1, the line will be taken to be correct and appended ## to the list of fixed-lines. If not, then the loop will attempt to test whether the current line ## marker is actually part of the previous line by looking in the current line for another marker ## that has the numeric value of previous-marker + 1. If found, that marker will be taken as the true ## marker for the line and the leader of the line (up to the point of this marker) will be appended ## to the revious line. E.g.: ## [1] Smith, J blah blah ## [2] Brown, N blah blah see reference ## [56] for more info [3] Wills, A blah blah ## ... ## ## ...will be transformed into: ## [1] Smith, J blah blah ## [2] Brown, N blah blah see reference [56] for more info ## [3] Wills, A blah blah ## ... ## first line is correct, to put it into fixed: fixed.append(rebuilt_lines[0]) try: for x in xrange(1, len(rebuilt_lines)): m = p_refmarker.match(rebuilt_lines[x]) try: if int(m.group("marknum")) == last_marknum + 1: ## The marker number for this reference line is correct. ## Append it to the 'fixed' lines and move on. fixed.append(rebuilt_lines[x]) last_marknum += 1 continue elif len(rebuilt_lines[x][m.end():].strip()) == 0: ## This line consists of a marker-number only - it is not a ## correct marker. Append it to the last line. fixed[len(fixed) - 1] += rebuilt_lines[x] continue else: ## This marker != previous-marker + 1. ## May have taken some of the last line into this line. Can we find the ## next marker in this line? ## Test for this situation: ## [54] for more info [3] Wills, A blah blah m_fix = p_refmarker.search(rebuilt_lines[x][m.end():]) if m_fix is not None and int(m_fix.group("marknum")) == last_marknum + 1: ## found next marker in line. Test to see that marker is followed by ## something meaningful i.e. a letter at least (name). ## I.e. We want to fix this: ## [54] for more info [3] Wills, A blah blah ## ## but we don't want to fix this: ## [54] for more info or even reference [3] ## ## as that would be unsafe. m_test_nxt_mark_not_eol = \ sre.search(sre.escape(m_fix.group()) + '\s*[A-Za-z]', rebuilt_lines[x]) if m_test_nxt_mark_not_eol is not None: ## move this section back to its real line: ## get the segment of this line to be moved to the previous line ## (append a newline to it too): movesect = rebuilt_lines[x][0:m_test_nxt_mark_not_eol.start()] + "\n" ## Now get the previous line into a variable (without its newline at the end): previous_line = fixed[len(fixed) - 1].rstrip("\n") ## Now append the section to be moved to the previous line variable. ## Check the last character of the previous line. If it's a space, then ## just directly append this new section. Else, append a space then this new section. previous_line += "%s%s" % ((previous_line[len(previous_line) - 1] != " " and " ") or (""), movesect) fixed[len(fixed) - 1] = previous_line ## Now append the remainder of the current line to the list of fixed lines, and move on to the ## next line: fixed.append(rebuilt_lines[x][m_test_nxt_mark_not_eol.start():]) last_marknum += 1 continue else: ## The next marker in the line was not followed by text. It is unsafe to move it. ## Give up trying to correct these reference lines - it's not safe to continue. unsafe = 1 break else: ## Unable to find another marker in the line that starts with the incorrect marker. ## It is therefore unsafe to attempt to correct the lines: just return the original lines. unsafe = 1 break except AttributeError: ## This line does not have a line marker at the start! This line shall be added to the end of the previous line. fixed[len(fixed) - 1] += rebuilt_lines[x] continue except IndexError: ## Somewhere, the boundaries of the list of references were over-stepped. Just return the original set of reference lines. unsafe = 1 if unsafe: ## return the original set of references. return rebuilt_lines else: ## return the newly corrected references. return fixed | 11ccb10bc99c9907d6df324bad7c06a4d678afe0 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/12027/11ccb10bc99c9907d6df324bad7c06a4d678afe0/refextract.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3434,
67,
266,
12406,
67,
3548,
12,
266,
12406,
67,
3548,
16,
293,
67,
1734,
11145,
4672,
3536,
7833,
358,
3434,
1281,
6088,
1625,
279,
2114,
980,
711,
2118,
25621,
1416,
2511,
12318,
27... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3434,
67,
266,
12406,
67,
3548,
12,
266,
12406,
67,
3548,
16,
293,
67,
1734,
11145,
4672,
3536,
7833,
358,
3434,
1281,
6088,
1625,
279,
2114,
980,
711,
2118,
25621,
1416,
2511,
12318,
27... |
elif self.Playlist.GetItem(idx + 1).GetText() == song_struct.DisplayFilename: | elif (self.Playlist.GetItem(idx + 1, self.TitleCol).GetText() == song_struct.Title) and (self.Playlist.GetItem(idx + 1, self.SecondColumn).GetText() == song_struct.DisplayFilename): | def _startDrag(self, e): # Wrap the song_struct in a DataObject. song_struct = self.PlaylistSongStructList[e.GetIndex()] songs = [song_struct] data = SongStructDataObject(songs) | 9ebc03dc54c47337c3ba2168f0dee12df16600f4 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/12653/9ebc03dc54c47337c3ba2168f0dee12df16600f4/pykaraoke.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
1937,
11728,
12,
2890,
16,
425,
4672,
468,
4266,
326,
17180,
67,
1697,
316,
279,
18914,
18,
17180,
67,
1697,
273,
365,
18,
30194,
55,
932,
3823,
682,
63,
73,
18,
967,
1016,
1435,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
1937,
11728,
12,
2890,
16,
425,
4672,
468,
4266,
326,
17180,
67,
1697,
316,
279,
18914,
18,
17180,
67,
1697,
273,
365,
18,
30194,
55,
932,
3823,
682,
63,
73,
18,
967,
1016,
1435,
... |
output.write(" <synopsis>%s</synopsis>\n" % synopsis); | output.write(" <synopsis>%s</synopsis>\n" % escape(synopsis)); | def parse(filename, output): global symbols global sections parser = CParser(filename) parser.collect_references() idx = parser.parse() info = parse_top_comment(filename, parser.top_comment) output.write(" <example filename='%s'>\n" % filename) try: synopsis = info['synopsis'] output.write(" <synopsis>%s</synopsis>\n" % synopsis); except: print "Example %s lacks a synopsis description" % (filename) try: purpose = info['purpose'] output.write(" <purpose>%s</purpose>\n" % purpose); except: print "Example %s lacks a purpose description" % (filename) try: usage = info['usage'] output.write(" <usage>%s</usage>\n" % usage); tests.append(usage) except: print "Example %s lacks an usage description" % (filename) try: author = info['author'] output.write(" <author>%s</author>\n" % author); except: print "Example %s lacks an author description" % (filename) try: copy = info['copy'] output.write(" <copy>%s</copy>\n" % copy); except: print "Example %s lacks a copyright description" % (filename) try: section = info['section'] output.write(" <section>%s</section>\n" % section); if sections.has_key(section): sections[section].append(filename) else: sections[section] = [filename] except: print "Example %s lacks a section description" % (filename) for topic in info.keys(): if topic != "purpose" and topic != "usage" and \ topic != "author" and topic != "copy" and \ topic != "section" and topic != "synopsis": str = info[topic] output.write(" <extra topic='%s'>%s</extra>\n" % str) output.write(" <includes>\n") for include in idx.includes.keys(): if include.find("libxml") != -1: output.write(" <include>%s</include>\n" % (escape(include))) output.write(" </includes>\n") output.write(" <uses>\n") for ref in idx.references.keys(): id = idx.references[ref] name = id.get_name() line = id.get_lineno() if symbols.has_key(name): sinfo = symbols[name] refs = sinfo[0] # gather at most 5 references per symbols if refs > 5: continue sinfo[refs] = filename sinfo[0] = refs + 1 else: symbols[name] = [1, filename] info = find_symbol(name) if info != None: type = info[0] file = info[1] output.write(" <%s line='%d' file='%s' name='%s'/>\n" % (type, line, file, name)) else: type = id.get_type() output.write(" <%s line='%d' name='%s'/>\n" % (type, line, name)) output.write(" </uses>\n") output.write(" </example>\n") return idx | 241e19d4afea87d71e761803877ba9371d70d816 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12659/241e19d4afea87d71e761803877ba9371d70d816/index.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1109,
12,
3459,
16,
876,
4672,
2552,
7963,
2552,
7178,
225,
2082,
273,
385,
2678,
12,
3459,
13,
2082,
18,
14676,
67,
14353,
1435,
2067,
273,
2082,
18,
2670,
1435,
1123,
273,
1109,
67,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1109,
12,
3459,
16,
876,
4672,
2552,
7963,
2552,
7178,
225,
2082,
273,
385,
2678,
12,
3459,
13,
2082,
18,
14676,
67,
14353,
1435,
2067,
273,
2082,
18,
2670,
1435,
1123,
273,
1109,
67,
... |
frame = frame.build_frame({'command': command, 'headers': headers, 'body': body}) return frame | return frame.build_frame({'command': command, 'headers': headers, 'body': body}) | def parse_frame(self, nb=False): """Parse data from socket | 0df02ebe57d54978f6ea9957a0c169e31f40c1da /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/2733/0df02ebe57d54978f6ea9957a0c169e31f40c1da/frame.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1109,
67,
3789,
12,
2890,
16,
4264,
33,
8381,
4672,
3536,
3201,
501,
628,
2987,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1109,
67,
3789,
12,
2890,
16,
4264,
33,
8381,
4672,
3536,
3201,
501,
628,
2987,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-... |
return self.allowed_ips | return self._allowed_ips | def get_allowed_ips(self): """ips can contain one '*' as a wildcard (eg 192.168.*).""" return self.allowed_ips | d939c6b1999d3c98f025af9b1be7148bf49a97ec /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2803/d939c6b1999d3c98f025af9b1be7148bf49a97ec/settings.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
67,
8151,
67,
7146,
12,
2890,
4672,
3536,
7146,
848,
912,
1245,
5306,
487,
279,
8531,
261,
1332,
20217,
18,
23329,
4509,
13,
12123,
327,
365,
6315,
8151,
67,
7146,
225,
2,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
67,
8151,
67,
7146,
12,
2890,
4672,
3536,
7146,
848,
912,
1245,
5306,
487,
279,
8531,
261,
1332,
20217,
18,
23329,
4509,
13,
12123,
327,
365,
6315,
8151,
67,
7146,
225,
2,
-100,
-... |
>>> int(bool(it.next())) 1 | >>> it.next() True | def item(self): """Get the iterator value | c2450cd2db47fce45ceeda283ccc45145d72ec81 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9778/c2450cd2db47fce45ceeda283ccc45145d72ec81/tales.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
761,
12,
2890,
4672,
3536,
967,
326,
2775,
460,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
761,
12,
2890,
4672,
3536,
967,
326,
2775,
460,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,... |
portaudio_patch_files = ["patches/portaudio-1420-runtime_device_change_detection.patch"] | portaudio_patch_files = ["patches/portaudio-1420-runtime_device_change_detection.patch", "patches/portaudio-1420-compile_snow_leopard.patch", "patches/portaudio-1420-pa_mac_core_x64_assert_fix.patch"] | def get_svn_revision(svn_dir, max_revision=None): environment = dict((name, value) for name, value in os.environ.iteritems() if name!='LANG' and not name.startswith('LC_')) environment['LC_ALL'] = 'C' if max_revision is None: svn_info = distutils_exec_process(["svn", "info", svn_dir], True, env=environment) else: svn_info = distutils_exec_process(["svn", "-r", str(max_revision), "info", svn_dir], True, env=environment) return int(re.search("Last Changed Rev: (\d+)", svn_info).group(1)) | 2ea682dd8acf0d7517243da3640a8f69154f497e /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/3449/2ea682dd8acf0d7517243da3640a8f69154f497e/setup_pjsip.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
67,
31505,
67,
13057,
12,
31505,
67,
1214,
16,
943,
67,
13057,
33,
7036,
4672,
3330,
273,
2065,
12443,
529,
16,
460,
13,
364,
508,
16,
460,
316,
1140,
18,
28684,
18,
2165,
3319,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
67,
31505,
67,
13057,
12,
31505,
67,
1214,
16,
943,
67,
13057,
33,
7036,
4672,
3330,
273,
2065,
12443,
529,
16,
460,
13,
364,
508,
16,
460,
316,
1140,
18,
28684,
18,
2165,
3319,
... |
if _prelude.prelude_option_recv_set(msg) < 0: | if not _prelude.prelude_option_recv_set(msg): | def commit(self, analyzer_path, instance): msg = self._request(analyzer_path, _prelude.PRELUDE_MSG_OPTION_COMMIT, instance) if _prelude.prelude_option_recv_set(msg) < 0: raise Error("could not read commit answer") | 3bfc82e9c10910b733a22d5234759df6bed00671 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/11628/3bfc82e9c10910b733a22d5234759df6bed00671/prelude.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3294,
12,
2890,
16,
15116,
67,
803,
16,
791,
4672,
1234,
273,
365,
6315,
2293,
12,
13647,
67,
803,
16,
389,
1484,
80,
1317,
18,
3670,
11778,
1639,
67,
11210,
67,
7425,
67,
18658,
16,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3294,
12,
2890,
16,
15116,
67,
803,
16,
791,
4672,
1234,
273,
365,
6315,
2293,
12,
13647,
67,
803,
16,
389,
1484,
80,
1317,
18,
3670,
11778,
1639,
67,
11210,
67,
7425,
67,
18658,
16,
... |
r')(\s*)(/)(?!=)', bygroups(Text, String.Regex), 'multiline-regex'), | r')(\s*)(/)', bygroups(Text, String.Regex), 'multiline-regex'), | def intp_string_callback(self, match, ctx): yield match.start(1), String.Other, match.group(1) nctx = LexerContext(match.group(3), 0, ['interpolated-string']) for i, t, v in self.get_tokens_unprocessed(context=nctx): yield match.start(3)+i, t, v yield match.start(4), String.Other, match.group(4) # end ctx.pos = match.end() | 33513306101af8abef3bde742211fe8556c25fd0 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6148/33513306101af8abef3bde742211fe8556c25fd0/agile.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
509,
84,
67,
1080,
67,
3394,
12,
2890,
16,
845,
16,
1103,
4672,
2824,
845,
18,
1937,
12,
21,
3631,
514,
18,
8290,
16,
845,
18,
1655,
12,
21,
13,
290,
5900,
273,
14234,
1042,
12,
19... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
509,
84,
67,
1080,
67,
3394,
12,
2890,
16,
845,
16,
1103,
4672,
2824,
845,
18,
1937,
12,
21,
3631,
514,
18,
8290,
16,
845,
18,
1655,
12,
21,
13,
290,
5900,
273,
14234,
1042,
12,
19... |
print "Returning %d columns" % len(columns) | log.debug("Returning %d columns" % len(columns)) | def parse(self): columns = None for result_file in self.result_files: print "Parsing: %s" % result_file.name.val image = self.image_from_original_file(result_file) provider = self.original_file_provider data = provider.get_original_file_data(result_file) try: rows = list(csv.reader(data, delimiter='\t')) finally: data.close() rows.reverse() if columns is None: columns = self.get_empty_columns(len(rows[0])) for row in rows: try: for i, value in enumerate(row): value = float(value) columns[i + 2].values.append(value) columns[self.IMAGE_COL].values.append(image.id.val) except ValueError: for i, value in enumerate(row): columns[i + 2].name = value break print "Returning %d columns" % len(columns) return MeasurementParsingResult([columns]) | 2abc05f6ec4c7dd58ee9bb618f839a4be9554279 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12409/2abc05f6ec4c7dd58ee9bb618f839a4be9554279/populate_roi.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1109,
12,
2890,
4672,
2168,
273,
599,
364,
563,
67,
768,
316,
365,
18,
2088,
67,
2354,
30,
1172,
315,
13963,
30,
738,
87,
6,
738,
563,
67,
768,
18,
529,
18,
1125,
1316,
273,
365,
1... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1109,
12,
2890,
4672,
2168,
273,
599,
364,
563,
67,
768,
316,
365,
18,
2088,
67,
2354,
30,
1172,
315,
13963,
30,
738,
87,
6,
738,
563,
67,
768,
18,
529,
18,
1125,
1316,
273,
365,
1... |
hdr = rpmUtils.miscutils.hdrFromPackage(ts, srpm) except rpmUtils.RpmUtilsError, e: error("Specified srpm %s cannot be found/opened" % srpm) sys.exit(50) if hdr[rpm.RPMTAG_SOURCEPACKAGE] != 1: error("Specified srpm isn't a srpm! Can't go on") sys.exit(50) try: my = None | def main(): # before we go on, make sure the user is a member of the 'mock' group. member = False for item in os.getgroups(): try: grptup = grp.getgrgid(item) except KeyError, e: continue if grptup[0] == 'mock': member = True if not member: print "You need to be a member of the mock group for this to work" sys.exit(1) # and make sure they're not root if os.geteuid() == 0: error("Don't try to run mock as root!") sys.exit(1) # config path config_path='/etc/mock' # defaults config_opts = {} config_opts['basedir'] = '/var/lib/mock/' # root name is automatically added to this config_opts['chroot'] = '/usr/sbin/mock-helper chroot' config_opts['mount'] = '/usr/sbin/mock-helper mount' config_opts['umount'] = '/usr/sbin/mock-helper umount' config_opts['rm'] = '/usr/sbin/mock-helper rm' config_opts['mknod'] = '/usr/sbin/mock-helper mknod' config_opts['yum'] = '/usr/sbin/mock-helper yum' config_opts['runuser'] = '/sbin/runuser' config_opts['chroot_dep_package'] = 'buildsys-build' config_opts['chrootuser'] = 'mockbuild' config_opts['chrootgroup'] = 'mockbuild' config_opts['chrootuid'] = 500 config_opts['chrootgid'] = 500 config_opts['chroothome'] = '/builddir' config_opts['clean'] = True config_opts['debug'] = False config_opts['quiet'] = False config_opts['target_arch'] = 'i386' config_opts['files'] = {} config_opts['yum.conf'] = '' config_opts['macros'] = """ | 55ae4aecb11782e193d9e8cbb18d09d3421ad958 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/7328/55ae4aecb11782e193d9e8cbb18d09d3421ad958/mock.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2774,
13332,
468,
1865,
732,
1960,
603,
16,
1221,
3071,
326,
729,
353,
279,
3140,
434,
326,
296,
22851,
11,
1041,
18,
3140,
273,
1083,
364,
761,
316,
1140,
18,
588,
4650,
13332,
775,
3... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2774,
13332,
468,
1865,
732,
1960,
603,
16,
1221,
3071,
326,
729,
353,
279,
3140,
434,
326,
296,
22851,
11,
1041,
18,
3140,
273,
1083,
364,
761,
316,
1140,
18,
588,
4650,
13332,
775,
3... | |
action_str = '<action command="delete" search="glob" path="/tmp/foo" regex="^foo$"/>' | action_str = '<action command="delete" search="glob" path="/tmp/foo" regex="^foo2$"/>' | def test_regex(self): """Unit test for regex option""" _iglob = glob.iglob glob.iglob = lambda x: ['/tmp/foo'] FileUtilities.getsize = lambda x: 1 action_str = '<action command="delete" search="glob" path="/tmp/foo" regex="^foo$"/>' result = self._action_str_to_result(action_str) self.assert_(result['path'], '/tmp/foo') action_str = '<action command="delete" search="glob" path="/tmp/foo" regex="^bar$"/>' self.assertRaises(StopIteration, lambda : self._action_str_to_result(action_str)) glob.iglob = _iglob | a742ee21dfeab4639a5f1eaff3dc6d4855930f20 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7853/a742ee21dfeab4639a5f1eaff3dc6d4855930f20/Action.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
7584,
12,
2890,
4672,
3536,
2802,
1842,
364,
3936,
1456,
8395,
389,
360,
1295,
273,
4715,
18,
360,
1295,
4715,
18,
360,
1295,
273,
3195,
619,
30,
10228,
19,
5645,
19,
11351,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
7584,
12,
2890,
4672,
3536,
2802,
1842,
364,
3936,
1456,
8395,
389,
360,
1295,
273,
4715,
18,
360,
1295,
4715,
18,
360,
1295,
273,
3195,
619,
30,
10228,
19,
5645,
19,
11351,
... |
from MySQLdb.connections import Connection if not hasattr(Connection, 'set_character_set'): encoding = self.dbEncoding def character_set_name(self): return encoding + '_' + encoding Connection.character_set_name = character_set_name | dbEncoding = self.dbEncoding if dbEncoding: from MySQLdb.connections import Connection if not hasattr(Connection, 'set_character_set'): def character_set_name(self): return dbEncoding + '_' + dbEncoding Connection.character_set_name = character_set_name | def makeConnection(self): from MySQLdb.connections import Connection if not hasattr(Connection, 'set_character_set'): # monkeypatch pre MySQLdb 1.2.1 encoding = self.dbEncoding def character_set_name(self): return encoding + '_' + encoding Connection.character_set_name = character_set_name try: conn = self.module.connect(host=self.host, port=self.port, db=self.db, user=self.user, passwd=self.password, **self.kw) if MySQLdb.version_info[:3] >= (1, 2, 2): conn.ping(True) # Attempt to reconnect. This setting is persistent. except self.module.OperationalError, e: raise OperationalError( "%s; used connection string: host=%s, port=%s, db=%s, user=%s, pwd=%s" % ( e, self.host, self.port, self.db, self.user, self.password) ) | 2262c355bafb1ec622c9d839428f94561870fce2 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/8798/2262c355bafb1ec622c9d839428f94561870fce2/mysqlconnection.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1221,
1952,
12,
2890,
4672,
628,
13485,
1966,
18,
13313,
1930,
4050,
309,
486,
3859,
12,
1952,
16,
296,
542,
67,
11560,
67,
542,
11,
4672,
468,
6921,
856,
2272,
675,
13485,
1966,
404,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1221,
1952,
12,
2890,
4672,
628,
13485,
1966,
18,
13313,
1930,
4050,
309,
486,
3859,
12,
1952,
16,
296,
542,
67,
11560,
67,
542,
11,
4672,
468,
6921,
856,
2272,
675,
13485,
1966,
404,
... |
pmap.draw_layout(gc, 0, 0, layout) pmap.draw_layout(gc, 0, height, layout) | if (self.__alignment == pango.ALIGN_RIGHT): pmap.draw_layout(gc, width, 0, layout) pmap.draw_layout(gc, width, height, layout) elif (self.__alignment == pango.ALIGN_CENTER): pmap.draw_layout(gc, int(width / 2.0), 0, layout) pmap.draw_layout(gc, int(width / 2.0), height, layout) else: pmap.draw_layout(gc, 0, 0, layout) pmap.draw_layout(gc, 0, height, layout) | def __render_text(self, layout): | 07ff62a140e92140a8669174a063ccfa6c8c38e4 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/5768/07ff62a140e92140a8669174a063ccfa6c8c38e4/TargetLabel.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
5902,
67,
955,
12,
2890,
16,
3511,
4672,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
5902,
67,
955,
12,
2890,
16,
3511,
4672,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
time.sleep(10) | time.sleep(random.randint(0,20)) | def run (self): debug ("Running Bicho") | a928fde902fac9f13cac26e81e66b0980e638282 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/10888/a928fde902fac9f13cac26e81e66b0980e638282/bg.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
261,
2890,
4672,
225,
1198,
7566,
7051,
605,
1354,
83,
7923,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
261,
2890,
4672,
225,
1198,
7566,
7051,
605,
1354,
83,
7923,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
sql1 = "insert or replace into InvertedIndex values(?,?)" | sql1 = u"insert or replace into InvertedIndex values(?,?)" | def _addTorrentToDB(self, infohash, data, commit=True): torrent_id = self._db.getTorrentID(infohash) #print >> sys.stderr, "-------------- Adding torrent to DB..", bin2str(infohash), str(torrent_id) if torrent_id is None: # not in db infohash_str = bin2str(infohash) self._db.insert('Torrent', commit=True, # must commit to get the torrent id infohash = infohash_str, name = dunno2unicode(data['name']), torrent_file_name = data['torrent_file_name'], length = data['length'], creation_date = data['creation_date'], num_files = data['num_files'], thumbnail = data['thumbnail'], insert_time = data['insert_time'], secret = data['secret'], relevance = data['relevance'], source_id = data['source'], category_id = data['category'], status_id = data['status'], num_seeders = data['num_seeders'], num_leechers = data['num_leechers'], comment = dunno2unicode(data['comment'])) torrent_id = self._db.getTorrentID(infohash) # adding in TorrentFiles table else: # infohash in db where = 'torrent_id = %d'%torrent_id self._db.update('Torrent', where = where, commit=False, name = dunno2unicode(data['name']), torrent_file_name = data['torrent_file_name'], length = data['length'], creation_date = data['creation_date'], num_files = data['num_files'], thumbnail = data['thumbnail'], insert_time = data['insert_time'], secret = data['secret'], relevance = data['relevance'], source_id = data['source'], category_id = data['category'], status_id = data['status'], num_seeders = data['num_seeders'], num_leechers = data['num_leechers'], comment = dunno2unicode(data['comment'])) | dd38a58c867405681df717cbbef78787b6112925 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/9686/dd38a58c867405681df717cbbef78787b6112925/SqliteCacheDBHandler.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
1289,
27266,
547,
774,
2290,
12,
2890,
16,
1123,
2816,
16,
501,
16,
3294,
33,
5510,
4672,
17351,
67,
350,
273,
365,
6315,
1966,
18,
588,
27266,
547,
734,
12,
1376,
2816,
13,
468,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
1289,
27266,
547,
774,
2290,
12,
2890,
16,
1123,
2816,
16,
501,
16,
3294,
33,
5510,
4672,
17351,
67,
350,
273,
365,
6315,
1966,
18,
588,
27266,
547,
734,
12,
1376,
2816,
13,
468,
... |
raise UserInputException("Unable to open MarCCD data file %s." % filename) | raise UserInputException("Unable to open MarCCD data file %s." % self.filename) | def __init__(self,filename): | dd1b10256add48a56e8da953a671093377ebf21d /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/8973/dd1b10256add48a56e8da953a671093377ebf21d/MarCCD.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
3459,
4672,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
3459,
4672,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
dprint("resetting flags to %s" % str(self.resize_flags)) | def resizeColumns(self, flags=[]): """Resize each column according to the flag. | 0ef70c5a19c75ffab56ceb6e4d6bab1286152640 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/11522/0ef70c5a19c75ffab56ceb6e4d6bab1286152640/mpd_major_mode.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
7041,
3380,
12,
2890,
16,
2943,
33,
8526,
4672,
3536,
12182,
1517,
1057,
4888,
358,
326,
2982,
18,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
7041,
3380,
12,
2890,
16,
2943,
33,
8526,
4672,
3536,
12182,
1517,
1057,
4888,
358,
326,
2982,
18,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... | |
print "output file: %s, dimensions: %s, %s, type: %s" %(opts.outF,myOut.RasterXSize,myOut.RasterYSize,gdal.GetDataTypeName(myOutB.DataType)) | print("output file: %s, dimensions: %s, %s, type: %s" %(opts.outF,myOut.RasterXSize,myOut.RasterYSize,gdal.GetDataTypeName(myOutB.DataType))) | def doit(opts, args): if opts.debug: print "gdal_calc.py starting calculation %s" %(opts.calc) ################################################################ # fetch details of input layers ################################################################ # set up some lists to store data for each band myFiles=[] myBands=[] myAlphaList=[] myDataType=[] myDataTypeNum=[] myNDV=[] DimensionsCheck=None # loop through input files - checking dimensions for i,myI in enumerate(AlphaList[0:len(sys.argv)-1]): myF = eval("opts.%s" %(myI)) myBand = eval("opts.%s_band" %(myI)) if myF: myFiles.append(gdal.Open(myF, gdal.GA_ReadOnly)) # check if we have asked for a specific band... if myBand: myBands.append(myBand) else: myBands.append(1) myAlphaList.append(myI) myDataType.append(gdal.GetDataTypeName(myFiles[i].GetRasterBand(myBands[i]).DataType)) myDataTypeNum.append(myFiles[i].GetRasterBand(myBands[i]).DataType) myNDV.append(myFiles[i].GetRasterBand(myBands[i]).GetNoDataValue()) # check that the dimensions of each layer are the same if DimensionsCheck: if DimensionsCheck!=[myFiles[i].RasterXSize, myFiles[i].RasterYSize]: print "Error! Dimensions of file %s (%i, %i) are different from file %s (%i, %i). Cannot proceed" return else: DimensionsCheck=[myFiles[i].RasterXSize, myFiles[i].RasterYSize] if opts.debug: print "file %s: %s, dimensions: %s, %s, type: %s" %(myI,myF,DimensionsCheck[0],DimensionsCheck[1],myDataType[i]) ################################################################ # set up output file ################################################################ # open output file exists if os.path.isfile(opts.outF) and not opts.overwrite: if opts.debug: print "Output file %s exists - filling in results into file" %(opts.outF) myOut=gdal.Open(opts.outF, gdal.GA_Update) if [myOut.RasterXSize,myOut.RasterYSize] != DimensionsCheck: print "Error! Output exists, but is the wrong size. Use the --overwrite option to automatically overwrite the existing file" return myOutB=myOut.GetRasterBand(1) myOutNDV=myOutB.GetNoDataValue() myOutType=gdal.GetDataTypeName(myOutB.DataType) else: # remove existing file and regenerate if os.path.isfile(opts.outF): os.remove(opts.outF) # create a new file if opts.debug: print "Generating output file %s" %(opts.outF) # find data type to use if not opts.type: # use the largest type of the input files myOutType=gdal.GetDataTypeName(max(myDataTypeNum)) else: myOutType=opts.type # create file myOutDrv = gdal.GetDriverByName(opts.format) myOut=myOutDrv.Create(opts.outF, DimensionsCheck[0], DimensionsCheck[1], 1, gdal.GetDataTypeByName(myOutType)) # set output geo info based on first input layer myOut.SetGeoTransform(myFiles[0].GetGeoTransform()) myOut.SetProjection(myFiles[0].GetProjection()) myOutB=myOut.GetRasterBand(1) if opts.NoDataValue!=None: myOutNDV=opts.NoDataValue else: myOutNDV=DefaultNDVLookup[myOutType] myOutB.SetNoDataValue(myOutNDV) # write to band myOutB=None # refetch band myOutB=myOut.GetRasterBand(1) if opts.debug: print "output file: %s, dimensions: %s, %s, type: %s" %(opts.outF,myOut.RasterXSize,myOut.RasterYSize,gdal.GetDataTypeName(myOutB.DataType)) ################################################################ # find block size to chop grids into bite-sized chunks ################################################################ # use the block size of the first layer to read efficiently myBlockSize=myFiles[0].GetRasterBand(myBands[0]).GetBlockSize(); # store these numbers in variables that may change later nXValid = myBlockSize[0] nYValid = myBlockSize[1] # find total x and y blocks to be read nXBlocks = (DimensionsCheck[0] + myBlockSize[0] - 1) / myBlockSize[0]; nYBlocks = (DimensionsCheck[1] + myBlockSize[1] - 1) / myBlockSize[1]; myBufSize = myBlockSize[0]*myBlockSize[1] if opts.debug: print "using blocksize %s x %s" %(myBlockSize[0], myBlockSize[1]) # variables for displaying progress ProgressCt=-1 ProgressMk=-1 ProgressEnd=nXBlocks*nYBlocks ################################################################ # start looping through blocks of data ################################################################ # loop through X-lines for X in range(0,nXBlocks): # in the rare (impossible?) case that the blocks don't fit perfectly # change the block size of the final piece if X==nXBlocks-1: nXValid = DimensionsCheck[0] - X * myBlockSize[0] myBufSize = nXValid*nYValid # find X offset myX=X*myBlockSize[0] # reset buffer size for start of Y loop nYValid = myBlockSize[1] myBufSize = nXValid*nYValid # loop through Y lines for Y in range(0,nYBlocks): ProgressCt+=1 if 10*ProgressCt/ProgressEnd%10!=ProgressMk: ProgressMk=10*ProgressCt/ProgressEnd%10 print 10*ProgressMk, "..", # change the block size of the final piece if Y==nYBlocks-1: nYValid = DimensionsCheck[1] - Y * myBlockSize[1] myBufSize = nXValid*nYValid # find Y offset myY=Y*myBlockSize[1] # create empty buffer to mark where nodata occurs myNDVs=zeros(myBufSize) myNDVs.shape=(nYValid,nXValid) # fetch data for each input layer for i,Alpha in enumerate(myAlphaList): # populate lettered arrays with values myval=BandReadAsArray(myFiles[i].GetRasterBand(myBands[i]), xoff=myX, yoff=myY, win_xsize=nXValid, win_ysize=nYValid) # fill in nodata values myNDVs=1*logical_or(myNDVs==1, myval==myNDV[i]) # create an array of values for this block exec("%s=myval" %Alpha) myval=None # try the calculation on the array blocks try: myResult = eval(opts.calc) except: print "evaluation of calculation %s failed" %(opts.calc) raise # propogate nodata values # (set nodata cells to zero then add nodata value to these cells) myResult = ((1*(myNDVs==0))*myResult) + (myOutNDV*myNDVs) # write data block to the output file BandWriteArray(myOutB, myResult, xoff=myX, yoff=myY) print "100 - Done" #print "Finished - Results written to %s" %opts.outF return | ac160e22187f4c00e013af415ee2cf81dece60a8 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/10289/ac160e22187f4c00e013af415ee2cf81dece60a8/gdal_calc.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
741,
305,
12,
4952,
16,
833,
4672,
225,
309,
1500,
18,
4148,
30,
1172,
315,
19016,
287,
67,
12448,
18,
2074,
5023,
11096,
738,
87,
6,
8975,
4952,
18,
12448,
13,
225,
19709,
5516,
26487... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
741,
305,
12,
4952,
16,
833,
4672,
225,
309,
1500,
18,
4148,
30,
1172,
315,
19016,
287,
67,
12448,
18,
2074,
5023,
11096,
738,
87,
6,
8975,
4952,
18,
12448,
13,
225,
19709,
5516,
26487... |
idstring = "$Id: notation3.py,v 1.86 2001-08-09 21:35:42 timbl Exp $" | idstring = "$Id: notation3.py,v 1.87 2001-08-23 19:56:08 connolly Exp $" | def startDoc(self): | 90b73eb0b26d962c245d671bd25c3dbe490fe899 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3308/90b73eb0b26d962c245d671bd25c3dbe490fe899/notation3.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
787,
1759,
12,
2890,
4672,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
787,
1759,
12,
2890,
4672,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-1... |
comments=self.comments, javascript=self.js_filter, level=self.level+1) | comments=self.comments, javascript=self.javascript, level=self.level+1) | def jsScript (self, script, ver, item): """execute given script with javascript version ver""" self._debug("JS: jsScript %s %s", ver, `script`) assert self.state[0]=='parse', "non-parse state %s" % str(self.state) assert len(self.buf) >= 2, "too small buffer %s" % str(self.buf) self.js_output = 0 self.js_env.attachListener(self) # start recursive html filter (used by jsProcessData) self.js_html = FilterHtmlParser(self.rules, self.pics, self.url, comments=self.comments, javascript=self.js_filter, level=self.level+1) # execute self.js_env.executeScript(unescape_js(script), ver) self.js_env.detachListener(self) # wait for recursive filter to finish self.jsEndScript(item) | 9d308cf36c9925fe55e611168b12175184ce8c08 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3948/9d308cf36c9925fe55e611168b12175184ce8c08/HtmlParser.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3828,
3651,
261,
2890,
16,
2728,
16,
1924,
16,
761,
4672,
3536,
8837,
864,
2728,
598,
11341,
1177,
1924,
8395,
365,
6315,
4148,
2932,
6479,
30,
3828,
3651,
738,
87,
738,
87,
3113,
1924,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3828,
3651,
261,
2890,
16,
2728,
16,
1924,
16,
761,
4672,
3536,
8837,
864,
2728,
598,
11341,
1177,
1924,
8395,
365,
6315,
4148,
2932,
6479,
30,
3828,
3651,
738,
87,
738,
87,
3113,
1924,
... |
numpoints=3, pad=0.1, | numpoints=3, borderpad=0.1, | def _set_legends(self, legends): legends.insert(0, Legend(label='Styles:', linestyle='')) legends.append(Legend(label='', linestyle='')) legends.append(Legend(label='Colors:', linestyle='')) if self._totals: legends.append(Legend(label='total', color=self._total_color)) if self._passed: legends.append(Legend(label='passed', color=self._pass_color)) if self._failed: legends.append(Legend(label='failed', color=self._fail_color)) labels = [ l.get_label() for l in legends ] self._figure.legend(legends, labels, loc='center right', numpoints=3, pad=0.1, prop=FontProperties(size=self._font_size)) | a2cb4aeae174d9f91b538b4b78e89a6351a27a5e /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7408/a2cb4aeae174d9f91b538b4b78e89a6351a27a5e/risto.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
542,
67,
1935,
5839,
12,
2890,
16,
4553,
5839,
4672,
4553,
5839,
18,
6387,
12,
20,
16,
17167,
409,
12,
1925,
2218,
9725,
30,
2187,
26388,
2218,
26112,
4553,
5839,
18,
6923,
12,
16... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
542,
67,
1935,
5839,
12,
2890,
16,
4553,
5839,
4672,
4553,
5839,
18,
6387,
12,
20,
16,
17167,
409,
12,
1925,
2218,
9725,
30,
2187,
26388,
2218,
26112,
4553,
5839,
18,
6923,
12,
16... |
_('[-ac] [-r REV] [REV]...')), | _('[-ac] [-r STARTREV] [REV]...')), | def version_(ui): """output version and copyright information""" ui.write(_("Mercurial Distributed SCM (version %s)\n") % util.version()) ui.status(_( "\nCopyright (C) 2005-2010 Matt Mackall <mpm@selenic.com> and others\n" "This is free software; see the source for copying conditions. " "There is NO\nwarranty; " "not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE.\n" )) | 1c611125cd0c1e5ac854c211a4c757f7110d9af1 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/11312/1c611125cd0c1e5ac854c211a4c757f7110d9af1/commands.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1177,
67,
12,
4881,
4672,
3536,
2844,
1177,
471,
22245,
1779,
8395,
5915,
18,
2626,
24899,
2932,
8478,
1397,
649,
27877,
8795,
49,
261,
1589,
738,
87,
5153,
82,
7923,
738,
1709,
18,
1589... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1177,
67,
12,
4881,
4672,
3536,
2844,
1177,
471,
22245,
1779,
8395,
5915,
18,
2626,
24899,
2932,
8478,
1397,
649,
27877,
8795,
49,
261,
1589,
738,
87,
5153,
82,
7923,
738,
1709,
18,
1589... |
self.max_id += 1 self.wiz_datas[self.max_id] = {} self.wiz_name[self.max_id] = wiz_name self.wiz_uid[self.max_id] = user return self.max_id | self._semaphore.acquire() wiz_id = 0 while True: wiz_id = randint(0, maxint) if wiz_id not in self.wiz_name: break self.wiz_datas[wiz_id] = {} self.wiz_name[wiz_id] = wiz_name self.wiz_uid[wiz_id] = user self._semaphore.release() return wiz_id | def create(self, database, user, passwd, wiz_name, datas=None): security.check(database, user, passwd) # FIXME: this is not thread-safe self.max_id += 1 self.wiz_datas[self.max_id] = {} self.wiz_name[self.max_id] = wiz_name self.wiz_uid[self.max_id] = user return self.max_id | 413d250b0498cd807042d17656564a1848463eee /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9266/413d250b0498cd807042d17656564a1848463eee/wizard.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
752,
12,
2890,
16,
2063,
16,
729,
16,
19093,
16,
341,
452,
67,
529,
16,
5386,
33,
7036,
4672,
4373,
18,
1893,
12,
6231,
16,
729,
16,
19093,
13,
468,
9852,
30,
333,
353,
486,
2650,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
752,
12,
2890,
16,
2063,
16,
729,
16,
19093,
16,
341,
452,
67,
529,
16,
5386,
33,
7036,
4672,
4373,
18,
1893,
12,
6231,
16,
729,
16,
19093,
13,
468,
9852,
30,
333,
353,
486,
2650,
... |
date_trunc('day',pp.create_date)::text as date_create, | to_char(pp.create_date, 'YYYY-MM-DD') as date_create, | def init(self, cr): tools.drop_view_if_exists(cr, 'report_transaction_pos') cr.execute(""" create or replace view report_transaction_pos as ( select min(pp.id) as id, count(pp.id) as no_trans, sum(amount) as amount, pp.journal_id, date_trunc('day',pp.create_date)::text as date_create, ps.user_id, ps.invoice_id from pos_payment pp, pos_order ps WHERE ps.id = pp.order_id group by pp.journal_id, date_trunc('day',pp.create_date), ps.user_id, ps.invoice_id ) """) | 6f5f7dd1c1a87112a4006fc3b5919071ed409066 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7397/6f5f7dd1c1a87112a4006fc3b5919071ed409066/pos.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1208,
12,
2890,
16,
4422,
4672,
8513,
18,
7285,
67,
1945,
67,
430,
67,
1808,
12,
3353,
16,
296,
6006,
67,
7958,
67,
917,
6134,
4422,
18,
8837,
2932,
3660,
752,
578,
1453,
1476,
2605,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1208,
12,
2890,
16,
4422,
4672,
8513,
18,
7285,
67,
1945,
67,
430,
67,
1808,
12,
3353,
16,
296,
6006,
67,
7958,
67,
917,
6134,
4422,
18,
8837,
2932,
3660,
752,
578,
1453,
1476,
2605,
... |
REQ_MASTER = False | def Exec(self, feedback_fn): """Adds the new node to the cluster. | 4be4691d5be0f08780678b18f75ed55b03fdba57 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/7542/4be4691d5be0f08780678b18f75ed55b03fdba57/cmdlib.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3889,
12,
2890,
16,
10762,
67,
4293,
4672,
3536,
3655,
326,
394,
756,
358,
326,
2855,
18,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3889,
12,
2890,
16,
10762,
67,
4293,
4672,
3536,
3655,
326,
394,
756,
358,
326,
2855,
18,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-10... | |
source = open(source_name, 'w') | with open(source_name, 'w') as source: | def test_checkcache(self): getline = linecache.getline try: # Create a source file and cache its contents source_name = support.TESTFN + '.py' with open(source_name, 'w') as source: source.write(SOURCE_1) source.close() getline(source_name, 1) | c2037ed2b9ceebb973d58f88cf4fec8c6435df25 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12029/c2037ed2b9ceebb973d58f88cf4fec8c6435df25/test_linecache.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
1893,
2493,
12,
2890,
4672,
336,
1369,
273,
4739,
557,
807,
18,
588,
1369,
775,
30,
468,
1788,
279,
1084,
585,
471,
1247,
2097,
2939,
1084,
67,
529,
273,
2865,
18,
16961,
197... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
1893,
2493,
12,
2890,
4672,
336,
1369,
273,
4739,
557,
807,
18,
588,
1369,
775,
30,
468,
1788,
279,
1084,
585,
471,
1247,
2097,
2939,
1084,
67,
529,
273,
2865,
18,
16961,
197... |
speech.speakMessage("needs click") | def _activateField(self,docHandle,ID): try: pacc,accChildID=IAccessibleHandler.accessibleObjectFromEvent(docHandle,IAccessibleHandler.OBJID_CLIENT,ID) role=pacc.accRole(accChildID) if role in (IAccessibleHandler.ROLE_SYSTEM_COMBOBOX,IAccessibleHandler.ROLE_SYSTEM_TEXT,IAccessibleHandler.ROLE_SYSTEM_LIST,IAccessibleHandler.ROLE_SYSTEM_SLIDER): self.passThrough=True virtualBufferHandler.reportPassThrough(self) else: #Just try performing the default action of the object, or of one of its ancestors obj=NVDAObjects.IAccessible.IAccessible(IAccessibleObject=pacc,IAccessibleChildID=accChildID) while obj and obj!=self.rootNVDAObject: try: action=obj.IAccessibleObject.accDefaultAction(obj.IAccessibleChildID) if action: try: speech.speakMessage("accDoDefaultAction") obj.IAccessibleObject.accDoDefaultAction(obj.IAccessibleChildID) except: speech.speakMessage("needs click") l=obj.location if l: x=(l[0]+l[2]/2) y=l[1]+(l[3]/2) oldX,oldY=winUser.getCursorPos() winUser.setCursorPos(x,y) speech.speakMessage("%s,%s"%(x,y)) winUser.mouse_event(winUser.MOUSEEVENTF_LEFTDOWN,0,0,None,None) winUser.mouse_event(winUser.MOUSEEVENTF_LEFTUP,0,0,None,None) winuser.setCursorPos(oldX,oldY) break except: pass obj=obj.parent except: pass | ec67d874b8b4de8d22b02a9e55fb0e394f12faf7 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9340/ec67d874b8b4de8d22b02a9e55fb0e394f12faf7/gecko_ia2.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
10014,
974,
12,
2890,
16,
2434,
3259,
16,
734,
4672,
775,
30,
6790,
952,
16,
8981,
1763,
734,
33,
45,
10451,
1503,
18,
26037,
921,
1265,
1133,
12,
2434,
3259,
16,
45,
10451,
1503,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
10014,
974,
12,
2890,
16,
2434,
3259,
16,
734,
4672,
775,
30,
6790,
952,
16,
8981,
1763,
734,
33,
45,
10451,
1503,
18,
26037,
921,
1265,
1133,
12,
2434,
3259,
16,
45,
10451,
1503,... | |
``a_invs``, which is a list of `\leq 5` \emph{integers} `a_1`, `a_2`, `a_3`, `a_4`, and `a_`$. If strictly less than 5 invariants are given, then the first ones are set to 0, so, e.g., ``[3,4] means `a_1=a_2=a_3=0` and `a_4=3`, `a_6=4`. INPUT: - `ainvs` (list or tuple) -- a list of <= 5 integers, the coefficients of a nonsingular Weierstrass equation. - `verbose` (bool, default False) -- verbosity flag. If True, then all Selmer group computations will be verbose. | ``ainvs``, which is a list of 5 or less *integers* `a_1`, `a_2`, `a_3`, `a_4`, and `a_5`. See the docstring of this class for full documentation. | def __init__(self, ainvs, verbose=False): r""" Create the mwrank elliptic curve with invariants ``a_invs``, which is a list of `\leq 5` \emph{integers} `a_1`, `a_2`, `a_3`, `a_4`, and `a_`$. | 9932a677a12413086a59217c4c0d425b0543b6eb /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9417/9932a677a12413086a59217c4c0d425b0543b6eb/interface.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
279,
5768,
87,
16,
3988,
33,
8381,
4672,
436,
8395,
1788,
326,
14721,
11500,
415,
549,
21507,
8882,
598,
316,
15886,
12176,
69,
67,
5768,
87,
68,
9191,
1... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
279,
5768,
87,
16,
3988,
33,
8381,
4672,
436,
8395,
1788,
326,
14721,
11500,
415,
549,
21507,
8882,
598,
316,
15886,
12176,
69,
67,
5768,
87,
68,
9191,
1... |
except IOError,e: | except IOError: | def notifyChannel(self, name, level, msg): warnings.warn("notifyChannel API shouldn't be used anymore, please use " "the standard `logging` module instead", PendingDeprecationWarning, stacklevel=2) from service.web_services import common | 46c754526ca696fe5c70f57e6109140674feb989 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12853/46c754526ca696fe5c70f57e6109140674feb989/netsvc.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
5066,
2909,
12,
2890,
16,
508,
16,
1801,
16,
1234,
4672,
5599,
18,
8935,
2932,
12336,
2909,
1491,
12044,
1404,
506,
1399,
16828,
16,
9582,
999,
315,
315,
5787,
4529,
1375,
11167,
68,
160... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
5066,
2909,
12,
2890,
16,
508,
16,
1801,
16,
1234,
4672,
5599,
18,
8935,
2932,
12336,
2909,
1491,
12044,
1404,
506,
1399,
16828,
16,
9582,
999,
315,
315,
5787,
4529,
1375,
11167,
68,
160... |
r""" Least prime dividing $n$ with $a(1)=1$. INPUT: n -- positive integer OUTPUT: integer -- function value EXAMPLES: sage: a = sloane.A020639;a Least prime dividing n (a(1)=1). sage: a(0) Traceback (most recent call last): ... ValueError: input n (=0) must be a positive integer sage: a(1) 1 sage: a(13) 13 sage: a.list(14) [1, 2, 3, 2, 5, 2, 7, 2, 3, 2, 11, 2, 13, 2] AUTHOR: -- Jaap Spies (2007-01-25) """ def __init__(self): | def __init__(self): r""" Least prime dividing $n$ with $a(1)=1$. INPUT: n -- positive integer OUTPUT: integer -- function value EXAMPLES: sage: a = sloane.A020639;a Least prime dividing n (a(1)=1). sage: a(0) Traceback (most recent call last): ... ValueError: input n (=0) must be a positive integer sage: a(1) 1 sage: a(13) 13 sage: a.list(14) [1, 2, 3, 2, 5, 2, 7, 2, 3, 2, 11, 2, 13, 2] AUTHOR: -- Jaap Spies (2007-01-25) """ | def list(self, n): try: if len(self._b) <= n: raise IndexError else: return self._b[:n] except (AttributeError, IndexError): self._precompute() # try again return self.list(n) | 69a199217d612e1c590af73e16003812c85b93ec /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9890/69a199217d612e1c590af73e16003812c85b93ec/sloane_functions.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
666,
12,
2890,
16,
290,
4672,
775,
30,
309,
562,
12,
2890,
6315,
70,
13,
1648,
290,
30,
1002,
10195,
469,
30,
327,
365,
6315,
70,
10531,
82,
65,
1335,
261,
1499,
668,
16,
10195,
4672... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
666,
12,
2890,
16,
290,
4672,
775,
30,
309,
562,
12,
2890,
6315,
70,
13,
1648,
290,
30,
1002,
10195,
469,
30,
327,
365,
6315,
70,
10531,
82,
65,
1335,
261,
1499,
668,
16,
10195,
4672... |
for link,error in res['Value']['Failed'].items(): gLogger.error("LcgFileCatalogClient.__createDataset: Failed to create link for %s." % link, error) totalError = "%s\n %s : %s" % (totalError,link,error) return S_ERROR(totalError) def __removeDataset(self,datasetName): res = self.__getDirectoryContents(datasetName) | for link, error in res['Value']['Failed'].items(): gLogger.error( "LcgFileCatalogClient.__createDataset: Failed to create link for %s." % link, error ) totalError = "%s\n %s : %s" % ( totalError, link, error ) return S_ERROR( totalError ) def __removeDataset( self, datasetName ): res = self.__getDirectoryContents( datasetName ) | def __createDataset(self,datasetName,lfns): res = self.__makeDirs(datasetName) if not res['OK']: return res links = {} for lfn in lfns: res = self.__getLFNGuid(lfn) if not res['OK']: return res else: link = "%s/%s" % (datasetName, res['Value']) links[link] = lfn res = self.createLink(links) if len(res['Value']['Successful']) == len(links.keys()): return S_OK() totalError = "" for link,error in res['Value']['Failed'].items(): gLogger.error("LcgFileCatalogClient.__createDataset: Failed to create link for %s." % link, error) totalError = "%s\n %s : %s" % (totalError,link,error) return S_ERROR(totalError) | 6280f3782654b93320f684f56a83a6624459bcec /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12864/6280f3782654b93320f684f56a83a6624459bcec/LcgFileCatalogClient.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2640,
10656,
12,
2890,
16,
8682,
461,
16,
20850,
2387,
4672,
400,
273,
365,
16186,
6540,
9872,
12,
8682,
461,
13,
309,
486,
400,
3292,
3141,
3546,
30,
327,
400,
4716,
273,
2618,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2640,
10656,
12,
2890,
16,
8682,
461,
16,
20850,
2387,
4672,
400,
273,
365,
16186,
6540,
9872,
12,
8682,
461,
13,
309,
486,
400,
3292,
3141,
3546,
30,
327,
400,
4716,
273,
2618,
... |
self.logging.info("Trying GIIS %s, %s" % (root['host'], root['base'])) | self.logging.debug("Trying GIIS %s, %s" % (root['host'], root['base'])) | def pick_CEs_from_giis_trees(self, root, tags, vos, seList, blacklist, whitelist, full): """ Recursively traverse the GIIS tree, starting from 'root', return CEs fullfilling requirements. """ | 86dabe12fde57408b0a702eb8d9727f73e5eb734 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/8886/86dabe12fde57408b0a702eb8d9727f73e5eb734/SchedulerARC.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
6002,
67,
1441,
87,
67,
2080,
67,
10052,
291,
67,
17204,
12,
2890,
16,
1365,
16,
2342,
16,
331,
538,
16,
695,
682,
16,
11709,
16,
10734,
16,
1983,
4672,
3536,
31362,
10080,
326,
611,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
6002,
67,
1441,
87,
67,
2080,
67,
10052,
291,
67,
17204,
12,
2890,
16,
1365,
16,
2342,
16,
331,
538,
16,
695,
682,
16,
11709,
16,
10734,
16,
1983,
4672,
3536,
31362,
10080,
326,
611,
... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.