rem stringlengths 1 226k | add stringlengths 0 227k | context stringlengths 6 326k | meta stringlengths 143 403 | input_ids listlengths 256 256 | attention_mask listlengths 256 256 | labels listlengths 128 128 |
|---|---|---|---|---|---|---|
if len(res.int) > context.prec and shouldround: | if res.int >= prec_limit and shouldround: | def _divide(self, other, divmod = 0, context=None): """Return a / b, to context.prec precision. | 636a6b100fe6083388bc5315758326078abe65b4 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8546/636a6b100fe6083388bc5315758326078abe65b4/decimal.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2892,
831,
12,
2890,
16,
1308,
16,
26105,
273,
374,
16,
819,
33,
7036,
4672,
3536,
990,
279,
342,
324,
16,
358,
819,
18,
4036,
6039,
18,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2892,
831,
12,
2890,
16,
1308,
16,
26105,
273,
374,
16,
819,
33,
7036,
4672,
3536,
990,
279,
342,
324,
16,
358,
819,
18,
4036,
6039,
18,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-... |
sreg_data = { } | sreg_data = {} | def approved(self, identity, openidreq, data=False, server_url=None): reply = openidreq.answer(True, identity=identity, server_url=server_url) if data: # TODO sreg_data = { } sreq_req = sreg.SRegRequest.fromOpenIDRequest(openidreq.message) sreg_resp = sreg.SRegResponse.extractResponse(openidreq, sreg_data) sreg_resp.addToOpenIDResponse(reply.fields) return reply | 1bd3c29f6b71d8b251879cc1178ddacbef674ebe /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/888/1bd3c29f6b71d8b251879cc1178ddacbef674ebe/serveopenid.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
20412,
12,
2890,
16,
4215,
16,
21629,
3658,
16,
501,
33,
8381,
16,
1438,
67,
718,
33,
7036,
4672,
4332,
273,
21629,
3658,
18,
13490,
12,
5510,
16,
4215,
33,
10781,
16,
1438,
67,
718,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
20412,
12,
2890,
16,
4215,
16,
21629,
3658,
16,
501,
33,
8381,
16,
1438,
67,
718,
33,
7036,
4672,
4332,
273,
21629,
3658,
18,
13490,
12,
5510,
16,
4215,
33,
10781,
16,
1438,
67,
718,
... |
self.log.info( "[COMPACT] Retrieving records to compact" ) result = self.__selectIndividualForCompactBuckets( typeName, timeLimit, bucketLength, nextBucketLength, querySize, connObj ) | self.log.info( "[COMPACT] Retrieving buckets to compact newer that %s with size %s" % ( Time.fromEpoch( timeLimit ), bucketLength ) ) roundStartTime = time.time() result = self.__selectIndividualForCompactBuckets( typeName, timeLimit, bucketLength, nextBucketLength, querySize, connObj ) | def __slowCompactBucketsForType( self, typeName ): """ Compact all buckets for a given type """ tableName = self.__getTableName( "bucket", typeName ) nowEpoch = Time.toEpoch() retVal = self._getConnection() if not retVal[ 'OK' ]: return retVal connObj = retVal[ 'Value' ] for bPos in range( len( self.dbBucketsLength[ typeName ] ) - 1 ): self.log.info( "[COMPACT] Query %d of %d" % ( bPos, len( self.dbBucketsLength[ typeName ] ) - 1 ) ) secondsLimit = self.dbBucketsLength[ typeName ][ bPos ][0] bucketLength = self.dbBucketsLength[ typeName ][ bPos ][1] timeLimit = ( nowEpoch - nowEpoch % bucketLength ) - secondsLimit nextBucketLength = self.dbBucketsLength[ typeName ][ bPos + 1 ][1] self.log.info( "[COMPACT] Compacting data newer that %s with bucket size %s" % ( Time.fromEpoch( timeLimit ), bucketLength ) ) querySize = 1000 previousRecordsSelected = querySize totalCompacted = 0 while previousRecordsSelected == querySize: #Retrieve the data self.log.info( "[COMPACT] Retrieving records to compact" ) result = self.__selectIndividualForCompactBuckets( typeName, timeLimit, bucketLength, nextBucketLength, querySize, connObj ) if not result[ 'OK' ]: #self.__rollbackTransaction( connObj ) return result bucketsData = result[ 'Value' ] previousRecordsSelected = len( bucketsData ) self.log.info( "[COMPACT] Got %d records to compact (%d done)" % ( previousRecordsSelected, totalCompacted ) ) if len( bucketsData ) == 0: break result = self.__deleteIndividualForCompactBuckets( typeName, bucketsData, connObj ) if not result[ 'OK' ]: #self.__rollbackTransaction( connObj ) return result bucketsData = result[ 'Value' ] self.log.info( "[COMPACT] Compacting %s records %s seconds size for %s" % ( len( bucketsData ), bucketLength, typeName ) ) #Add data for record in bucketsData: startTime = record[-2] endTime = record[-2] + record[-1] valuesList = record[:-2] retVal = self.__splitInBuckets( typeName, startTime, endTime, valuesList, connObj ) if not retVal[ 'OK' ]: self.log.error( "[COMPACT] Error while compacting data for record in %s: %s" % ( typeName, retVal[ 'Value' ] ) ) totalCompacted += len( bucketsData ) self.log.info( "[COMPACT] Finised compaction %d of %d" % ( bPos, len( self.dbBucketsLength[ typeName ] ) - 1 ) ) #return self.__commitTransaction( connObj ) connObj.close() return S_OK() | 20794236a2a7be5aef39b9c2feeef4fc7b6fcab0 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12864/20794236a2a7be5aef39b9c2feeef4fc7b6fcab0/AccountingDB.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
87,
821,
16863,
14517,
22405,
12,
365,
16,
8173,
262,
30,
3536,
23823,
777,
9169,
364,
279,
864,
618,
3536,
4775,
273,
365,
16186,
588,
7147,
12,
315,
7242,
3113,
8173,
262,
2037,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
87,
821,
16863,
14517,
22405,
12,
365,
16,
8173,
262,
30,
3536,
23823,
777,
9169,
364,
279,
864,
618,
3536,
4775,
273,
365,
16186,
588,
7147,
12,
315,
7242,
3113,
8173,
262,
2037,
... |
""" | """pygame._numpysndarray.samples(Sound): return array Reference Sound samples into an array. Creates a new array that directly references the samples in a Sound object. Modifying the array will change the Sound. The array will always be in the format returned from pygame.mixer.get_init(). | def samples (sound): """ """ # Info is a (freq, format, stereo) tuple info = pygame.mixer.get_init () if not info: raise pygame.error, "Mixer not initialized" fmtbytes = (abs (info[1]) & 0xff) >> 3 channels = mixer.get_num_channels () data = sound.get_buffer () shape = (data.length / channels * fmtbytes, ) if channels > 1: shape = (shape[0], 2) typecode = None # Signed or unsigned representation? if format in (pygame.AUDIO_S8, pygame.AUDIO_S16LSB, pygame.AUDIO_S16MSB): typecode = (numpy.uint8, numpy.uint16, None, numpy.uint32)[fmtbytes - 1] else: typecode = (numpy.int8, numpy.int16, None, numpy.int32)[fmtbytes - 1] array = numpy.frombuffer (data, typecode) array.shape = shape return array | da0bcce5e530d0f9eb7bdd9211d2f405671a9945 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/1298/da0bcce5e530d0f9eb7bdd9211d2f405671a9945/_numpysndarray.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
5216,
261,
29671,
4672,
3536,
2074,
13957,
6315,
2107,
84,
1900,
25681,
18,
7319,
12,
24331,
4672,
327,
526,
225,
6268,
348,
772,
5216,
1368,
392,
526,
18,
225,
10210,
279,
394,
526,
716... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
5216,
261,
29671,
4672,
3536,
2074,
13957,
6315,
2107,
84,
1900,
25681,
18,
7319,
12,
24331,
4672,
327,
526,
225,
6268,
348,
772,
5216,
1368,
392,
526,
18,
225,
10210,
279,
394,
526,
716... |
if position != -1 and type == 'downloaded-item': | if position != -1 and typ == 'downloaded-item': | def validate_drop(self, table_view, model, typ, source_actions, parent, position): if position != -1 and type == 'downloaded-item': return widgetset.DRAG_ACTION_MOVE return widgetset.DRAG_ACTION_NONE | 416d488a041306cfa19d3494fb5924aff5b1462a /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12354/416d488a041306cfa19d3494fb5924aff5b1462a/playlist.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1954,
67,
7285,
12,
2890,
16,
1014,
67,
1945,
16,
938,
16,
3815,
16,
1084,
67,
4905,
16,
982,
16,
1754,
4672,
309,
1754,
480,
300,
21,
471,
3815,
422,
296,
2378,
4230,
17,
1726,
4278... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1954,
67,
7285,
12,
2890,
16,
1014,
67,
1945,
16,
938,
16,
3815,
16,
1084,
67,
4905,
16,
982,
16,
1754,
4672,
309,
1754,
480,
300,
21,
471,
3815,
422,
296,
2378,
4230,
17,
1726,
4278... |
return RSA.construct(decodedKey[1: 6]) | n,e,d,p,q=decodedKey[1:6] return RSA.construct((n,e,d,p,q)) | def getPrivateKeyObject_openssh(data, passphrase): kind = data[0][11: 14] if data[1].startswith('Proc-Type: 4,ENCRYPTED'): # encrypted key ivdata = data[2].split(',')[1][:-1] iv = ''.join([chr(int(ivdata[i:i+2],16)) for i in range(0, len(ivdata), 2)]) if not passphrase: raise BadKeyError, 'encrypted key with no passphrase' ba = md5.new(passphrase + iv).digest() bb = md5.new(ba + passphrase + iv).digest() decKey = (ba + bb)[:24] b64Data = base64.decodestring(''.join(data[4:-1])) keyData = des_cbc3_decrypt(b64Data, decKey, iv) removeLen = ord(keyData[-1]) keyData = keyData[:-removeLen] else: keyData = base64.decodestring(''.join(data[1:-1])) try: decodedKey = asn1.parse(keyData) except Exception, e: raise BadKeyError, 'something wrong with decode' if type(decodedKey[0]) == type([]): decodedKey = decodedKey[0] # this happens with encrypted keys if kind == 'RSA': return RSA.construct(decodedKey[1: 6]) elif kind == 'DSA': p, q, g, y, x = decodedKey[1: 6] return DSA.construct((y, g, p, q, x)) | 3cf1ed6058067c1597ae54cb3a1cd95725a1e8d1 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12595/3cf1ed6058067c1597ae54cb3a1cd95725a1e8d1/keys.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
27914,
653,
921,
67,
20346,
674,
12,
892,
16,
15244,
4672,
3846,
273,
501,
63,
20,
6362,
2499,
30,
5045,
65,
309,
501,
63,
21,
8009,
17514,
1918,
2668,
15417,
17,
559,
30,
1059,
16,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
27914,
653,
921,
67,
20346,
674,
12,
892,
16,
15244,
4672,
3846,
273,
501,
63,
20,
6362,
2499,
30,
5045,
65,
309,
501,
63,
21,
8009,
17514,
1918,
2668,
15417,
17,
559,
30,
1059,
16,
... |
platform.python_implementation(), | getattr(platform, 'python_implementation', 'Python'), | def main(self): | 116d78d26f3f8b18baa39d484bf4684e8f09b30f /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/12029/116d78d26f3f8b18baa39d484bf4684e8f09b30f/pybench.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2774,
12,
2890,
4672,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2774,
12,
2890,
4672,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-... |
self.any_issue = False | self.any_issue = False | def write_doc(self, docname, doctree): self.any_issue = False # set when any issue is encountered in this document self.docname = docname visitor = SuspiciousVisitor(doctree, self) doctree.walk(visitor) | 19b3e00627d162f338cd2d1fc0b51ab3f352d38a /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/8546/19b3e00627d162f338cd2d1fc0b51ab3f352d38a/suspicious.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1045,
67,
2434,
12,
2890,
16,
997,
529,
16,
31263,
992,
4672,
225,
365,
18,
2273,
67,
13882,
273,
1083,
468,
444,
1347,
1281,
5672,
353,
9919,
316,
333,
1668,
365,
18,
2434,
529,
273,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1045,
67,
2434,
12,
2890,
16,
997,
529,
16,
31263,
992,
4672,
225,
365,
18,
2273,
67,
13882,
273,
1083,
468,
444,
1347,
1281,
5672,
353,
9919,
316,
333,
1668,
365,
18,
2434,
529,
273,
... |
self.ParseExpectationsLine(line, lineno) | self.ParseExpectationsLine(line, lineno) | def _Read(self, expectations): """For each test in an expectations iterable, generate the expectations for it. """ lineno = 0 for line in expectations: lineno += 1 | 8eb26144f32b9dc19ff34c5291f421ac2bf50daa /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/5060/8eb26144f32b9dc19ff34c5291f421ac2bf50daa/test_expectations.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
1994,
12,
2890,
16,
26305,
4672,
3536,
1290,
1517,
1842,
316,
392,
26305,
6087,
16,
2103,
326,
26305,
364,
518,
18,
3536,
7586,
273,
374,
364,
980,
316,
26305,
30,
7586,
1011,
404,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
1994,
12,
2890,
16,
26305,
4672,
3536,
1290,
1517,
1842,
316,
392,
26305,
6087,
16,
2103,
326,
26305,
364,
518,
18,
3536,
7586,
273,
374,
364,
980,
316,
26305,
30,
7586,
1011,
404,
... |
'platform', 'exit', 'maxint') | 'platform', 'exit', 'maxint') | def reload(self, module, path=None): if path is None and hasattr(module, '__filename__'): head, tail = os.path.split(module.__filename__) path = [os.path.join(head, '')] return ihooks.ModuleImporter.reload(self, module, path) | 3ec38f0ee48b31c80510c60b3df0e6bc4dfb89f9 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8546/3ec38f0ee48b31c80510c60b3df0e6bc4dfb89f9/rexec.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
7749,
12,
2890,
16,
1605,
16,
589,
33,
7036,
4672,
309,
589,
353,
599,
471,
3859,
12,
2978,
16,
4940,
3459,
7250,
4672,
910,
16,
5798,
273,
1140,
18,
803,
18,
4939,
12,
2978,
16186,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
7749,
12,
2890,
16,
1605,
16,
589,
33,
7036,
4672,
309,
589,
353,
599,
471,
3859,
12,
2978,
16,
4940,
3459,
7250,
4672,
910,
16,
5798,
273,
1140,
18,
803,
18,
4939,
12,
2978,
16186,
... |
os.remove(fileToDelete) | try: os.remove(fileToDelete) msg = fileToDelete + ' deleted.' fileWasDeleted = 1 except: msg = fileToDelete + ' NOT deleted (OS error).' | def deleteFile(self, rel_dir, fileToDelete, REQUEST, RESPONSE): """ delete the file """ | c6bc9553ed5360092abb504eda0dae11a0fca1dc /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/1807/c6bc9553ed5360092abb504eda0dae11a0fca1dc/PloneLocalFolderNG.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
29485,
12,
2890,
16,
1279,
67,
1214,
16,
585,
14976,
16,
12492,
16,
20645,
4672,
3536,
1430,
326,
585,
3536,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
29485,
12,
2890,
16,
1279,
67,
1214,
16,
585,
14976,
16,
12492,
16,
20645,
4672,
3536,
1430,
326,
585,
3536,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-1... |
command = 'convert ' + path + ' -resize 300x300 ' + thumb | command = 'convert ' + path + ' -resize 300x300 -antialias ' + thumb | def image_link(self,path,webserver): thumb = "thumb_" + path command = 'convert ' + path + ' -resize 300x300 ' + thumb print command popen = subprocess.Popen(command.split()) popen.communicate() status = popen.returncode s = '[[ImageLink('+webserver+'/'+thumb+','+webserver+'/'+path+',width=300][,alt=none])]]' self.file.write(s) | e994ee1446d54255416ae828b6e19dfbf18f256d /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/5758/e994ee1446d54255416ae828b6e19dfbf18f256d/make_summary_page.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1316,
67,
1232,
12,
2890,
16,
803,
16,
1814,
2266,
502,
4672,
11156,
273,
315,
21639,
9548,
397,
589,
1296,
273,
296,
6283,
296,
397,
589,
397,
296,
300,
15169,
11631,
92,
19249,
300,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1316,
67,
1232,
12,
2890,
16,
803,
16,
1814,
2266,
502,
4672,
11156,
273,
315,
21639,
9548,
397,
589,
1296,
273,
296,
6283,
296,
397,
589,
397,
296,
300,
15169,
11631,
92,
19249,
300,
... |
["debug","Chandler"], | ["debug"], | def distribute(buildenv): _createVersionFile(buildenv) buildVersionShort = \ hardhatutil.RemovePunctuation(buildenv['buildVersion']) # When the build version string is based on one of our CVS tags # (which usually begin with "CHANDLER_") let's remove the "CHANDLER_" # prefix from the string so it doesn't end up in the generated filenames # (so we can avoid getting a distro file named: # "Chandler_linux_CHANDLER_M1.tar.gz", and instead get: # "Chandler_linux_M1.tar.gz") buildVersionShort = buildVersionShort.replace("CHANDLER_", "") if buildenv['version'] == 'debug': if buildenv['os'] == 'posix': distName = 'Chandler_linux_debug_' + buildVersionShort distDir = buildenv['root'] + os.sep + distName buildenv['distdir'] = distDir if os.access(distDir, os.F_OK): hardhatlib.rmdir_recursive(distDir) os.mkdir(distDir) manifestFile = "distrib/linux/manifest.debug.linux" hardhatlib.handleManifest(buildenv, manifestFile) os.chdir(buildenv['root']) compFile1 = hardhatlib.compressDirectory(buildenv, [distName], distName) os.chdir(buildenv['root']) compFile2 = hardhatlib.compressDirectory(buildenv, ["debug","Chandler"], "Chandler_linux_dev_debug_" + buildVersionShort) os.chdir(buildenv['root']) if buildenv['os'] == 'osx': distName = 'Chandler_osx_debug_' + buildVersionShort # when we make an osx distribution, we actually need to put it # in a subdirectory (which has a .app extension). So we set # 'distdir' temporarily to that .app dir so that handleManifest() # puts things in the right place. Then we set 'distdir' to its # parent so that it gets cleaned up further down. distDirParent = buildenv['root'] + os.sep + distName distDir = distDirParent + os.sep + distName + ".app" buildenv['distdir'] = distDir if os.access(distDirParent, os.F_OK): hardhatlib.rmdir_recursive(distDirParent) os.mkdir(distDirParent) os.mkdir(distDir) manifestFile = "distrib/osx/manifest.debug.osx" hardhatlib.handleManifest(buildenv, manifestFile) makeDiskImage = buildenv['hardhatroot'] + os.sep + \ "makediskimage.sh" os.chdir(buildenv['root']) hardhatlib.executeCommand(buildenv, "HardHat", [makeDiskImage, distName], "Creating disk image from " + distName) compFile1 = distName + ".dmg" # reset 'distdir' up a level so that it gets removed below. buildenv['distdir'] = distDirParent distDir = distDirParent os.chdir(buildenv['root']) compFile2 = hardhatlib.compressDirectory(buildenv, ["debug","Chandler"], "Chandler_osx_dev_debug_" + buildVersionShort) if buildenv['os'] == 'win': distName = 'Chandler_win_debug_' + buildVersionShort distDir = buildenv['root'] + os.sep + distName buildenv['distdir'] = distDir if os.access(distDir, os.F_OK): hardhatlib.rmdir_recursive(distDir) os.mkdir(distDir) manifestFile = "distrib" + os.sep + "win" + os.sep + \ "manifest.debug.win" hardhatlib.handleManifest(buildenv, manifestFile) os.chdir(buildenv['root']) compFile1 = hardhatlib.compressDirectory(buildenv, [distName], distName) os.chdir(buildenv['root']) compFile2 = hardhatlib.compressDirectory(buildenv, ["debug","Chandler"], "Chandler_win_dev_debug_" + buildVersionShort) # put the compressed file in the right place if specified 'outputdir' if buildenv['outputdir']: if not os.path.exists(buildenv['outputdir']): os.mkdir(buildenv['outputdir']) if os.path.exists(buildenv['outputdir']+os.sep+compFile1): os.remove(buildenv['outputdir']+os.sep+compFile1) os.rename(compFile1, buildenv['outputdir']+os.sep+compFile1) if os.path.exists(buildenv['outputdir']+os.sep+compFile2): os.remove(buildenv['outputdir']+os.sep+compFile2) os.rename(compFile2, buildenv['outputdir']+os.sep+compFile2) # remove the distribution directory, since we have a tarball/zip if os.access(distDir, os.F_OK): hardhatlib.rmdir_recursive(distDir) if buildenv['version'] == 'release': if buildenv['os'] == 'posix': distName = 'Chandler_linux_' + buildVersionShort distDir = buildenv['root'] + os.sep + distName buildenv['distdir'] = distDir if os.access(distDir, os.F_OK): hardhatlib.rmdir_recursive(distDir) os.mkdir(distDir) manifestFile = "distrib/linux/manifest.linux" hardhatlib.handleManifest(buildenv, manifestFile) os.chdir(buildenv['root']) compFile1 = hardhatlib.compressDirectory(buildenv, [distName], distName) os.chdir(buildenv['root']) compFile2 = hardhatlib.compressDirectory(buildenv, ["release","Chandler"], "Chandler_linux_dev_release_" + buildVersionShort) os.chdir(buildenv['root']) if buildenv['os'] == 'osx': distName = 'Chandler_osx_' + buildVersionShort # when we make an osx distribution, we actually need to put it # in a subdirectory (which has a .app extension). So we set # 'distdir' temporarily to that .app dir so that handleManifest() # puts things in the right place. Then we set 'distdir' to its # parent so that it gets cleaned up further down. distDirParent = buildenv['root'] + os.sep + distName distDir = distDirParent + os.sep + distName + ".app" buildenv['distdir'] = distDir if os.access(distDirParent, os.F_OK): hardhatlib.rmdir_recursive(distDirParent) os.mkdir(distDirParent) os.mkdir(distDir) manifestFile = "distrib/osx/manifest.osx" hardhatlib.handleManifest(buildenv, manifestFile) makeDiskImage = buildenv['hardhatroot'] + os.sep + \ "makediskimage.sh" os.chdir(buildenv['root']) hardhatlib.executeCommand(buildenv, "HardHat", [makeDiskImage, distName], "Creating disk image from " + distName) compFile1 = distName + ".dmg" # reset 'distdir' up a level so that it gets removed below. buildenv['distdir'] = distDirParent distDir = distDirParent os.chdir(buildenv['root']) compFile2 = hardhatlib.compressDirectory(buildenv, ["release","Chandler"], "Chandler_osx_dev_release_" + buildVersionShort) if buildenv['os'] == 'win': distName = 'Chandler_win_' + buildVersionShort distDir = buildenv['root'] + os.sep + distName buildenv['distdir'] = distDir if os.access(distDir, os.F_OK): hardhatlib.rmdir_recursive(distDir) os.mkdir(distDir) manifestFile = "distrib" + os.sep + "win" + os.sep + "manifest.win" hardhatlib.handleManifest(buildenv, manifestFile) os.chdir(buildenv['root']) compFile1 = hardhatlib.compressDirectory(buildenv, [distName], distName) os.chdir(buildenv['root']) compFile2 = hardhatlib.compressDirectory(buildenv, ["release","Chandler"], "Chandler_win_dev_release_" + buildVersionShort) # put the compressed files in the right place if specified 'outputdir' if buildenv['outputdir']: if not os.path.exists(buildenv['outputdir']): os.mkdir(buildenv['outputdir']) if os.path.exists(buildenv['outputdir'] + os.sep + compFile1): os.remove(buildenv['outputdir'] + os.sep + compFile1) os.rename( compFile1, buildenv['outputdir'] + os.sep + compFile1) if os.path.exists(buildenv['outputdir'] + os.sep + compFile2): os.remove(buildenv['outputdir'] + os.sep + compFile2) os.rename( compFile2, buildenv['outputdir'] + os.sep + compFile2) # remove the distribution directory, since we have a tarball/zip if os.access(distDir, os.F_OK): hardhatlib.rmdir_recursive(distDir) | bb746634a0a498285ef59eb777154aa0ece1413b /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9228/bb746634a0a498285ef59eb777154aa0ece1413b/__hardhat__.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
25722,
12,
3510,
3074,
4672,
225,
389,
2640,
1444,
812,
12,
3510,
3074,
13,
225,
1361,
1444,
4897,
273,
521,
7877,
11304,
1367,
18,
3288,
52,
15726,
12,
3510,
3074,
3292,
3510,
1444,
194... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
25722,
12,
3510,
3074,
4672,
225,
389,
2640,
1444,
812,
12,
3510,
3074,
13,
225,
1361,
1444,
4897,
273,
521,
7877,
11304,
1367,
18,
3288,
52,
15726,
12,
3510,
3074,
3292,
3510,
1444,
194... |
args = [None] * 3 * (len(xml) and xml[-1].order) kwargs = {"spacing" : [None] * (len(xml) and xml[-1].order)} | args = [None] * 3 * (len(xml) and (xml[-1].order + 1)) kwargs = {"spacing" : [None] * (len(xml) and (xml[-1].order + 1))} | def bins_from_xml(xml): """ From the XML document tree rooted at xml, retrieve the table describing a binning, and construct and return a rate.Bins object from it. """ xml = table.get_table(xml, BinsTable.tableName) xml.sort(lambda a, b: cmp(a.order, b.order)) args = [None] * 3 * (len(xml) and xml[-1].order) kwargs = {"spacing" : [None] * (len(xml) and xml[-1].order)} for row in xml: args[row.order * 3 : row.order * 3 + 3] = row.min, row.max, row.n kwargs["spacing"][row.order] = row.type if None in args: raise ValueError, "incomplete bin spec: %s" % str(args) return Bins(*args, **kwargs) | bdf096808928d1a8c003f741be1c74ab8afde255 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/5758/bdf096808928d1a8c003f741be1c74ab8afde255/rate.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
8963,
67,
2080,
67,
2902,
12,
2902,
4672,
3536,
6338,
326,
3167,
1668,
2151,
1365,
329,
622,
2025,
16,
4614,
326,
1014,
16868,
279,
4158,
2093,
16,
471,
4872,
471,
327,
279,
4993,
18,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
8963,
67,
2080,
67,
2902,
12,
2902,
4672,
3536,
6338,
326,
3167,
1668,
2151,
1365,
329,
622,
2025,
16,
4614,
326,
1014,
16868,
279,
4158,
2093,
16,
471,
4872,
471,
327,
279,
4993,
18,
... |
(p.post_id == pt.post_id) & ( (p.status == STATUS_PUBLISHED) | (p.pub_date >= datetime.utcnow()))) | (pt.post_id == p.post_id) & (p.status == STATUS_PUBLISHED) & (p.pub_date >= datetime.utcnow())) | def get_cloud(self, max=None, ignore_privileges=False): """Get a categorycloud.""" # XXX: ignore_privileges is currently ignored and no privilege # checking is performed. As a matter of fact only published posts # appear in the cloud. | a19568afaf3324d99c401d3cbd7246fe3025aa69 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/12815/a19568afaf3324d99c401d3cbd7246fe3025aa69/models.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
67,
7277,
12,
2890,
16,
943,
33,
7036,
16,
2305,
67,
28459,
2852,
33,
8381,
4672,
3536,
967,
279,
3150,
7277,
12123,
468,
11329,
30,
2305,
67,
28459,
2852,
353,
4551,
5455,
471,
1... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
67,
7277,
12,
2890,
16,
943,
33,
7036,
16,
2305,
67,
28459,
2852,
33,
8381,
4672,
3536,
967,
279,
3150,
7277,
12123,
468,
11329,
30,
2305,
67,
28459,
2852,
353,
4551,
5455,
471,
1... |
def view_diff_for_path(self, url, latest_revision_number, earliest_revision_number=None, sidebyside=False): from rabbitvcs.ui.diff import diff_factory if earliest_revision_number == None: earliest_revision_number = latest_revision_number self.action = vcs_action_factory( self.vcs, notification=False ) self.action.append( diff_factory, url, earliest_revision_number - 1, url, latest_revision_number, sidebyside=sidebyside ) self.action.start() | def set_loading(self, loading): self.is_loading = loading | 1c802fd2fccd2c3f322ab1831bc02dadb8933cfc /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/5119/1c802fd2fccd2c3f322ab1831bc02dadb8933cfc/log.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
444,
67,
15174,
12,
2890,
16,
7153,
4672,
365,
18,
291,
67,
15174,
273,
7153,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
444,
67,
15174,
12,
2890,
16,
7153,
4672,
365,
18,
291,
67,
15174,
273,
7153,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-10... | |
opath='/'.join(self.getPhysicalPath()) | opath = '/'.join(self.getPhysicalPath()) | def undoable_transactions(self, first_transaction=None, last_transaction=None, PrincipiaUndoBatchSize=None): | 966fde8b9dcbf826b59950a8749914790b8a3c9a /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9658/966fde8b9dcbf826b59950a8749914790b8a3c9a/Undo.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
15436,
429,
67,
20376,
12,
2890,
16,
1122,
67,
7958,
33,
7036,
16,
1142,
67,
7958,
33,
7036,
16,
2301,
4613,
1155,
31224,
23304,
33,
7036,
4672,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
15436,
429,
67,
20376,
12,
2890,
16,
1122,
67,
7958,
33,
7036,
16,
1142,
67,
7958,
33,
7036,
16,
2301,
4613,
1155,
31224,
23304,
33,
7036,
4672,
2,
-100,
-100,
-100,
-100,
-100,
-100,
... |
"""Submit a job for execution. | """Record a job for later submit. | def QueueJob(self, name, *ops): """Submit a job for execution. | 23b4b983afc9b9e81d558f06e4e0cde53703e575 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7542/23b4b983afc9b9e81d558f06e4e0cde53703e575/cli.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
7530,
2278,
12,
2890,
16,
508,
16,
380,
4473,
4672,
3536,
11620,
279,
1719,
364,
4588,
18,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
7530,
2278,
12,
2890,
16,
508,
16,
380,
4473,
4672,
3536,
11620,
279,
1719,
364,
4588,
18,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-1... |
punct = self.sentence_end | pat = self.sentence_end_re | def _fix_sentence_endings (self, chunks): """_fix_sentence_endings(chunks : [string]) | 45bdab6974ba18810507edf93004c2dc34e868ee /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8125/45bdab6974ba18810507edf93004c2dc34e868ee/textwrap.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
904,
67,
19335,
67,
409,
899,
261,
2890,
16,
6055,
4672,
3536,
67,
904,
67,
19335,
67,
409,
899,
12,
14691,
294,
306,
1080,
5717,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
904,
67,
19335,
67,
409,
899,
261,
2890,
16,
6055,
4672,
3536,
67,
904,
67,
19335,
67,
409,
899,
12,
14691,
294,
306,
1080,
5717,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
ctype = TYPES[self.get_type(section, option)] | ctype = TYPES[self.get_type(section, opt)] | def initialize(self): """ Load the initial database options into the config editor, and generate the required widgets. """ pages = [] sects = self.opts.sections() sects.sort() for section in sects: box = gtk.VBox() sectlab = ''.join([section[0].upper(), section[1:]]) tid = self.nb.append_page(box, gtk.Label(sectlab)) pages.append((sectlab, tid)) opts = self.opts.options(section) opts.sort() for opt in opts: ctype = TYPES[self.get_type(section, option)] cw = ctype(self.cb, section, opt) box.pack_start(cw.win, expand=False, padding=4) self.controls[(section, opt)] = cw box.pack_start(gtk.HSeparator(), expand=False, padding=4) self.tree.populate(pages) | 1cdd34f88caa60f74d95c99070d6eba5a6736b84 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2233/1cdd34f88caa60f74d95c99070d6eba5a6736b84/config.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4046,
12,
2890,
4672,
3536,
4444,
326,
2172,
2063,
702,
1368,
326,
642,
4858,
16,
471,
2103,
326,
1931,
10965,
18,
3536,
4689,
273,
5378,
29140,
87,
273,
225,
365,
18,
4952,
18,
11657,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4046,
12,
2890,
4672,
3536,
4444,
326,
2172,
2063,
702,
1368,
326,
642,
4858,
16,
471,
2103,
326,
1931,
10965,
18,
3536,
4689,
273,
5378,
29140,
87,
273,
225,
365,
18,
4952,
18,
11657,
... |
sedcmd+=" -e 'CXXFLAGS=\""+self.settings["CXXFLAGS"]+"\"\\'" | cmds.append('CXXFLAGS="'+self.settings["CXXFLAGS"]+'"') | def chroot_setup(self): cmd("cp /etc/resolv.conf "+self.settings["chroot_path"]+"/etc","Could not copy resolv.conf into place.") #Ugly bunch of sed commands to get /etc/make.conf to hold the correct default values for the stage #we're building. This way, when a user uses a pentium4 stage3, it is set up to compile for pentium4 #using the CFLAGS and USE settings we used. It's possible that this would look nicer written in #python, even without using regexes which aren't really necessary. mycmds=[] mycmds.append("sed -i -e '/# catalyst start/,/# catalyst stop/d'") mycmds.append("sed -i -e 's:^CFLAGS=:#&:' -e 's:^CXXFLAGS=:#&:' -e 's:^CHOST=:#&:' -e 's:^USE=:#&:'") sedcmd="sed -i -e '5i\\' -e '# catalyst start\\' -e '# these settings were added by the catalyst build script" sedcmd+=" that automatically built this stage\\' -e 'CFLAGS=\""+self.settings["CFLAGS"]+"\"\\'" if self.settings.has_key("CXXFLAGS"): sedcmd+=" -e 'CXXFLAGS=\""+self.settings["CXXFLAGS"]+"\"\\'" else: sedcmd+=" -e 'CXXFLAGS=\"${CFLAGS}\"\\'" sedcmd+=" -e 'CHOST=\""+self.settings["CHOST"]+"\"\\'" if self.settings.has_key("HOSTUSE"): sedcmd+=" -e 'USE=\""+string.join(self.settings["HOSTUSE"])+"\"\\'" sedcmd+=" -e '# catalyst end\\' -e ''" mycmds.append(sedcmd) for x in mycmds: mycmd=x+" "+self.settings["chroot_path"]+"/etc/make.conf" cmd(mycmd,"Sed command failed: "+mycmd) | a093e753f843e05532097d5a8f4e595a799934f2 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/7327/a093e753f843e05532097d5a8f4e595a799934f2/targets.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
462,
3085,
67,
8401,
12,
2890,
4672,
1797,
2932,
4057,
342,
14175,
19,
7818,
90,
18,
3923,
13773,
2890,
18,
4272,
9614,
343,
3085,
67,
803,
6,
3737,
6,
19,
14175,
15937,
4445,
486,
161... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
462,
3085,
67,
8401,
12,
2890,
4672,
1797,
2932,
4057,
342,
14175,
19,
7818,
90,
18,
3923,
13773,
2890,
18,
4272,
9614,
343,
3085,
67,
803,
6,
3737,
6,
19,
14175,
15937,
4445,
486,
161... |
return self.call_method(env, target, source + es, **overrides) class WrapLibrary: """\ Wrap the Library builder to add the target as a named construction varible, by which other parts of the build can refer to them. """ def __init__(self, name, env): self.libraryBuilder = env['BUILDERS'][name] self.call_method = self.libraryBuilder.__call__ self.libraryBuilder.__call__ = self.__call__ | return self.method(target, source + es, **overrides) def WrapLibrary(name, env, method): env.AddMethod(LibraryMethod(name, env, method), name) class LibraryMethod: """ Wrap a Library Environment method to add the resulting library target as a named global target. Other parts of the build can refer to these named targets implicitly by using AppendLibrary() or explicitly through GetGlobalTarget(). Typically the environment method being wrapped is actually a SCons.Environment.BuilderWrapper, so this effectively chains that MethodWrapper with this one. Since the existing MethodWrapper is already bound to the Environment instance, the environment does not need to be passed when the chained method is called. """ def __init__(self, name, env, method): self.name = name self.library = method Debug("created library wrapper for %s, library=%s" % (name, self.library)) | def __call__(self, env, target = None, source = _null, **overrides): es = env['EXTRA_SOURCES'] if source == _null and len(es) == 0: return self.call_method(env, target, source, **overrides) else: if type(source) != type([]): source = [source] return self.call_method(env, target, source + es, **overrides) | b9ab93193da8fb8642770b62b603e9a01587d228 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/1466/b9ab93193da8fb8642770b62b603e9a01587d228/__init__.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
1991,
972,
12,
2890,
16,
1550,
16,
1018,
273,
599,
16,
1084,
273,
389,
2011,
16,
2826,
19775,
4672,
5001,
273,
1550,
3292,
22639,
67,
6537,
55,
3546,
309,
1084,
422,
389,
2011,
4... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
1991,
972,
12,
2890,
16,
1550,
16,
1018,
273,
599,
16,
1084,
273,
389,
2011,
16,
2826,
19775,
4672,
5001,
273,
1550,
3292,
22639,
67,
6537,
55,
3546,
309,
1084,
422,
389,
2011,
4... |
do('dmap.sessionid', 101) ]) | do('dmap.sessionid', session_id) ]) | def GET(self): mlog = do('dmap.loginresponse', [ do('dmap.status', 200), do('dmap.sessionid', 101) ]) return self.h(web,mlog.encode()) | 7f1700be343da028c435d27dd80f8aeb85c12223 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/8691/7f1700be343da028c435d27dd80f8aeb85c12223/server.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4978,
12,
2890,
4672,
312,
1330,
273,
741,
2668,
72,
1458,
18,
5819,
2740,
2187,
306,
741,
2668,
72,
1458,
18,
2327,
2187,
4044,
3631,
741,
2668,
72,
1458,
18,
3184,
350,
2187,
1339,
6... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4978,
12,
2890,
4672,
312,
1330,
273,
741,
2668,
72,
1458,
18,
5819,
2740,
2187,
306,
741,
2668,
72,
1458,
18,
2327,
2187,
4044,
3631,
741,
2668,
72,
1458,
18,
3184,
350,
2187,
1339,
6... |
a_x[condition] = 1 + 0.17699*y - 0.50447*y**2 - 0.02427*y**3 + 0.72085*y**4 a_x[condition] = a_x[condition] + 0.01979*y**5 - 0.77530*y**6 + 0.32999*y**7 b_x[condition] = 1.41338*y + 2.28305*y**2 + 1.07233*y**3 - 5.38434*y**4 b_x[condition] = b_x[condition] - 0.62251*y**5 + 5.30260*y**6 - 2.09002*y**7 | a_x[condition] = 1 + 0.104*y - 0.609*y**2 + 0.701*y**3 + 1.137*y**4 a_x[condition] = a_x[condition] - 1.718*y**5 - 0.827*y**6 + 1.647*y**7 - 0.505*y**8 b_x[condition] = 1.952*y + 2.908*y**2 - 3.989*y**3 - 7.985*y**4 b_x[condition] = b_x[condition] + 11.102*y**5 + 5.491*y**6 - 10.805*y**7 + 3.347*y**8 | def setupCCMab(self, wavelen=None): """Calculate a(x) and b(x) for CCM dust model. (x=1/wavelen). Returns a(x) and b(x) can be common to many seds, wavelen is the same. """ # This extinction law taken from Cardelli, Clayton and Mathis ApJ 1989. # The general form is A_l / A(V) = a(x) + b(x)/R_V (where x=1/lambda in microns), # then different values for a(x) and b(x) depending on wavelength regime. # Also, the extinction is parametrized as R_v = A_v / E(B-V). # Magnitudes of extinction (A_l) translates to flux by a_l = -2.5log(f_red / f_nonred). if wavelen == None: wavelen = n.copy(self.wavelen) a_x = n.zeros(len(wavelen), dtype='float') b_x = n.zeros(len(wavelen), dtype='float') # Convert wavelength to x (in inverse microns). x = n.empty(len(wavelen), dtype=float) x = 1.0 / wavelen * 10000.0 # Dust in infrared 0.3 /mu < x < 1.1 /mu (inverse microns). condition = (x>=0.3) & (x<=1.1) if len(a_x[condition]) > 0 : y = x[condition] a_x[condition] = 0.574 * y**1.61 b_x[condition] = -0.527 * y**1.61 # Dust in optical/NIR 1.1 /mu < x < 3.3 /mu region. condition = (x >=1.1) & (x<=3.3) if len(a_x[condition])>0: y = x[condition] - 1.82 a_x[condition] = 1 + 0.17699*y - 0.50447*y**2 - 0.02427*y**3 + 0.72085*y**4 a_x[condition] = a_x[condition] + 0.01979*y**5 - 0.77530*y**6 + 0.32999*y**7 b_x[condition] = 1.41338*y + 2.28305*y**2 + 1.07233*y**3 - 5.38434*y**4 b_x[condition] = b_x[condition] - 0.62251*y**5 + 5.30260*y**6 - 2.09002*y**7 # Dust in ultraviolet and UV (if needed for high-z) 3.3 /mu< x< 8 /mu. condition = (x>=3.3) & (x<5.9) if len(a_x[condition])>0: y = x[condition] a_x[condition] = 1.752 - 0.316*y - 0.104/((y-4.67)**2 + 0.341) b_x[condition] = -3.090 + 1.825*y + 1.206/((y-4.62)**2 + 0.263) condition = (x>5.9) & (x<8) if len(a_x[condition])>0: y = x[condition] Fa_x = n.empty(len(a_x[condition]), dtype=float) Fb_x = n.empty(len(a_x[condition]), dtype=float) Fa_x = -0.04473*(y-5.9)**2 - 0.009779*(y-5.9)**3 Fb_x = 0.2130*(y-5.9)**2 + 0.1207*(y-5.9)**3 a_x[condition] = 1.752 - 0.316*y - 0.104/((y-4.67)**2 + 0.341) + Fa_x b_x[condition] = -3.090 + 1.825*y + 1.206/((y-4.62)**2 + 0.263) + Fb_x # Dust in far UV (if needed for high-z) 8 /mu < x < 10 /mu region. condition = (x >= 8) & (x<= 10.) if len(a_x[condition])>0: y = x[condition]-8.0 a_x[condition] = -1.073 - 0.628*(y) + 0.137*(y)**2 - 0.070*(y)**3 b_x[condition] = 13.670 + 4.257*(y) - 0.420*(y)**2 + 0.374*(y)**3 return a_x, b_x | fe3b601823c6a418c550ad220bad597f52e218ef /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/91/fe3b601823c6a418c550ad220bad597f52e218ef/Sed.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3875,
6743,
49,
378,
12,
2890,
16,
341,
7882,
275,
33,
7036,
4672,
3536,
8695,
279,
12,
92,
13,
471,
324,
12,
92,
13,
364,
385,
9611,
302,
641,
938,
18,
261,
92,
33,
21,
19,
91,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3875,
6743,
49,
378,
12,
2890,
16,
341,
7882,
275,
33,
7036,
4672,
3536,
8695,
279,
12,
92,
13,
471,
324,
12,
92,
13,
364,
385,
9611,
302,
641,
938,
18,
261,
92,
33,
21,
19,
91,
... |
self.reportResults(appConf['appName'], testStartDate, reportFile) | self.reportResults(appConf['appName'], testStartDate, logFile) | def runTests(self, appConf): import time testStartDate = time.strftime(self.timeFormat) | d8d47fb8e9657c82e39f2dad9f3e5d63c17db316 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/5718/d8d47fb8e9657c82e39f2dad9f3e5d63c17db316/qxtest.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
14650,
12,
2890,
16,
595,
3976,
4672,
1930,
813,
1842,
22635,
273,
813,
18,
701,
9982,
12,
2890,
18,
957,
1630,
13,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
14650,
12,
2890,
16,
595,
3976,
4672,
1930,
813,
1842,
22635,
273,
813,
18,
701,
9982,
12,
2890,
18,
957,
1630,
13,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
executable = result['Value'] result = ce.submitJob(executable,'',pilotsToSubmit) | executable = result['Executable'] proxy = result['Proxy'] result = ce.submitJob(executable,proxy,pilotsToSubmit) | def submitJobs(self): """ Go through defined computing elements and submit jobs if necessary """ for queue in self.queueDict: ce = self.queueDict[queue]['CE'] ceName = self.queueDict[queue]['CEName'] queueName = self.queueDict[queue]['QueueName'] result = ce.available() if not result['OK']: self.log.warn('Failed to check the availability of queue %s: %s' (queue,result['message'])) continue totalSlots = result['Value'] ceDict = ce.getParameterDict() result = taskQueueDB.getMatchingTaskQueues( ceDict ) if not result['OK']: self.log.error( 'Could not retrieve TaskQueues from TaskQueueDB', result['Message'] ) return result taskQueueDict = result['Value'] print taskQueueDict if not taskQueueDict: continue totalTQJobs = 0 for tq in taskQueueDict: totalTQJobs += taskQueueDict[tq]['Jobs'] pilotsToSubmit = min(totalSlots,totalTQJobs) pilotsToSubmit = 1 if pilotsToSubmit > 0: self.log.info('Going to submit %d pilots to %s queue' % (pilotsToSubmit,queue) ) result = self.__getExecutable(queue,pilotsToSubmit) if not result['OK']: return result executable = result['Value'] result = ce.submitJob(executable,'',pilotsToSubmit) if not result['OK']: self.log.error('Failed submission to queue %s: %s' (queue,result['Message'])) # Add pilots to the PilotAgentsDB assign pilots to TaskQueue proportionally to the # task queue priorities pilotList = result['Value'] nPilots = len(pilotList) tqPriorityList = [] sumPriority = 0. for tq in taskQueueDict: sumPriority += taskQueueDict[tq]['Priority'] tqPriorityList.append((tq,sumPriority)) rndm = random.random()*sumPriority tqDict = {} for pilotID in pilotList: rndm = random.random()*sumPriority for tq,prio in tqPriorityList: if rndm < prio: tqID = tq break if not tqDict.has_key(tqID): tqDict[tqID] = [] tqDict[tqID].append(pilotID) for tqID,pilotList in tqDict.items(): result = pilotAgentsDB.addPilotTQReference(pilotList, tqID, self.genericPilotDN, self.genericPilotGroup, self.localhost, 'DIRAC', '') if not result['OK']: self.log.error('Failed add pilots to the PilotAgentsDB: %s' % result['Message']) continue for pilot in pilotList: result = pilotAgentsDB.setPilotStatus(pilot,'Submitted',ceName,'Successfuly submitted by the SiteDirector','',queueName) if not result['OK']: self.log.error('Failed to set pilot status: %s' % result['Message']) continue return S_OK() | 87df46f4bfb71617e88bec645fda74d7a411459e /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12864/87df46f4bfb71617e88bec645fda74d7a411459e/SiteDirector.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4879,
7276,
12,
2890,
4672,
3536,
4220,
3059,
2553,
20303,
2186,
471,
4879,
6550,
309,
4573,
3536,
225,
364,
2389,
316,
365,
18,
4000,
5014,
30,
5898,
273,
365,
18,
4000,
5014,
63,
4000,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4879,
7276,
12,
2890,
4672,
3536,
4220,
3059,
2553,
20303,
2186,
471,
4879,
6550,
309,
4573,
3536,
225,
364,
2389,
316,
365,
18,
4000,
5014,
30,
5898,
273,
365,
18,
4000,
5014,
63,
4000,... |
self.targets.add(dbsgmlfile, [sourcefile.file_only], [Command('wt2db -n -s ' + sourcefile.file_only + ' -o ' + dbsgmlfile, output_to='log/wt2db.log', stderr_check=1)]) | self.targets.add(dbsgmlfile, [sourcefile.file_only], [Command('wt2db -n -s ' + sourcefile.file_only + ' -o ' + dbsgmlfile, errors_to='log/wt2db.log', stderr_check=1)]) | def __init__(self, doc_id): self.doc_id = int(doc_id) self.doc = lampadas.docs[self.doc_id] self.workdir = config.cache_dir + str(self.doc_id) + '/work/' self.filename = '' self.targets = Targets() # If the file is not to be published (Archived or Normal status), # or if it has not been mirrored successfully, skip it. if self.doc.pub_status_code<>'N' or self.doc.mirror_time=='': return | 493349fc054afbb2e642c9dbde2e01d225c8adc8 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/6947/493349fc054afbb2e642c9dbde2e01d225c8adc8/Makefile.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
997,
67,
350,
4672,
365,
18,
2434,
67,
350,
273,
509,
12,
2434,
67,
350,
13,
365,
18,
2434,
273,
328,
931,
361,
345,
18,
8532,
63,
2890,
18,
2434,
67... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
997,
67,
350,
4672,
365,
18,
2434,
67,
350,
273,
509,
12,
2434,
67,
350,
13,
365,
18,
2434,
273,
328,
931,
361,
345,
18,
8532,
63,
2890,
18,
2434,
67... |
fname = r"..\..\datasets\microarray\brown\brown-selected.tab" if os.path.exists(fname): table = orange.ExampleTable(fname) attrs = [attr.name for attr in table.domain.attributes] graph.setData(table) graph.updateData(attrs, 1) | data = orange.ExampleTable(r"E:\Development\Orange Datasets\UCI\wine.tab") graph.setData(data) graph.updateData([attr.name for attr in data.domain.attributes]) | def drawCanvas(self, painter): if self.showProbabilities and getattr(self, "potentialsClassifier", None): self.computePotentials() painter.drawImage(self.transform(QwtPlot.xBottom, -1), self.transform(QwtPlot.yLeft, 1), self.potentialsImage) OWGraph.drawCanvas(self, painter) | 7f6cbe615f2b3f8641c476ea41cf9c62a332569f /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/6366/7f6cbe615f2b3f8641c476ea41cf9c62a332569f/OWLinProjGraph.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3724,
12971,
12,
2890,
16,
21295,
4672,
309,
365,
18,
4500,
9152,
5756,
471,
3869,
12,
2890,
16,
315,
26451,
87,
13860,
3113,
599,
4672,
365,
18,
9200,
24947,
87,
1435,
21295,
18,
9446,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3724,
12971,
12,
2890,
16,
21295,
4672,
309,
365,
18,
4500,
9152,
5756,
471,
3869,
12,
2890,
16,
315,
26451,
87,
13860,
3113,
599,
4672,
365,
18,
9200,
24947,
87,
1435,
21295,
18,
9446,
... |
for item in self.conn.do.search('album', root): | for item in self.browse_search_album(root): | def browse(self, widget=None, root='/'): # Populates the library list with entries starting at root if not self.conn: return | ef335bb5758783289bea4a5b77282ae0c87d687a /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/2312/ef335bb5758783289bea4a5b77282ae0c87d687a/sonata.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
21670,
12,
2890,
16,
3604,
33,
7036,
16,
1365,
2218,
2473,
4672,
468,
10264,
17099,
326,
5313,
666,
598,
3222,
5023,
622,
1365,
309,
486,
365,
18,
4646,
30,
327,
2,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
21670,
12,
2890,
16,
3604,
33,
7036,
16,
1365,
2218,
2473,
4672,
468,
10264,
17099,
326,
5313,
666,
598,
3222,
5023,
622,
1365,
309,
486,
365,
18,
4646,
30,
327,
2,
-100,
-100,
-100,
-... |
prefs = preferences.prefs_context() self.default_display_btngrp.setButton(prefs.get(defaultDisplayMode_prefs_key, diVDW)) | self.default_display_btngrp.setButton(self.default_display_mode) | def _setup_atoms_page(self): ''' Setup widgets to initial (default or defined) values on the display page. ''' # Set colors for atom color swatches self.atom_hilite_color_frame.setPaletteBackgroundColor(RGBf_to_QColor(orange)) self.free_valence_color_frame.setPaletteBackgroundColor(RGBf_to_QColor(red)) # Bug 799 fix. Mark 050731 prefs = preferences.prefs_context() self.default_display_btngrp.setButton(prefs.get(defaultDisplayMode_prefs_key, diVDW)) | 3016e55d3518c88ce00a8f44a8c14be4334433de /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/11221/3016e55d3518c88ce00a8f44a8c14be4334433de/UserPrefs.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
8401,
67,
14937,
67,
2433,
12,
2890,
4672,
9163,
10939,
10965,
358,
2172,
261,
1886,
578,
2553,
13,
924,
603,
326,
2562,
1363,
18,
9163,
468,
1000,
5740,
364,
3179,
2036,
1352,
505,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
8401,
67,
14937,
67,
2433,
12,
2890,
4672,
9163,
10939,
10965,
358,
2172,
261,
1886,
578,
2553,
13,
924,
603,
326,
2562,
1363,
18,
9163,
468,
1000,
5740,
364,
3179,
2036,
1352,
505,... |
scatt_rrats = ax.scatter(np.log10(periods_rrat), np.log10(pdots_rrat), \ | scatt_rrats = ax.scatter(periods_rrat, pdots_rrat, \ | def plot_data(pulsars, hightlight=[], binaries=False, rrats=False, \ magnetars=False, snrs=False, axp=False, sgr=False, \ edots=[], ages=[], bsurfs=[]): """Plot P-Pdot diagram using list of pulsars provided. binaries - boolean, initially plot binary markers rrats - boolean, initially plot RRAT markers magnetars - boolean, initially plot magnetar markers snrs - boolean, initially plot SNR markers edots - values for lines of constant edot ages - values for lines of constant ages bsurfs - values for lines of constant surfave b-field """ global scatt_psrs, scatt_psrs_hl periods = np.array([x.p for x in pulsars \ if x.p is not None and x.pdot is not None]) pdots = np.array([x.pdot for x in pulsars \ if x.p is not None and x.pdot is not None]) ax = plt.axes() scatt_psrs = ax.scatter(np.log10(periods), np.log10(pdots), c='k', s=6, \ label='pulsars', picker=PICKER, zorder=0) # Pulsars to highlight numhl = 0 if len(highlight): periods_hl = np.array([h.p for h in highlight \ if h.p is not None and h.pdot is not None]) pdots_hl = np.array([h.pdot for h in highlight \ if h.p is not None and h.pdot is not None]) scatt_psrs_hl = ax.scatter(np.log10(periods_hl), np.log10(pdots_hl), \ c='r', s=50, label='highlight', picker=PICKER, \ marker=(5,1,0), edgecolors='r', zorder=1) numhl = periods_hl.size # Mark binaries periods_bnry = np.array([x.p for x in pulsars \ if x.p is not None and x.pdot is not None \ and x.binary==True]) pdots_bnry = np.array([x.pdot for x in pulsars \ if x.p is not None and x.pdot is not None \ and x.binary==True]) global scatt_binaries if periods_bnry.size: scatter_options = MARKER_OPTIONS.copy() scatter_options.update(BINARY_MARKER) scatt_binaries = ax.scatter(np.log10(periods_bnry), np.log10(pdots_bnry), \ **scatter_options) if not binaries: # Hide binaries for now scatt_binaries.set_visible(False) else: scatt_binaries = None # Mark RRATs periods_rrat = np.array([x.p for x in pulsars+highlight \ if x.p is not None and x.pdot is not None \ and x.rrat==True]) pdots_rrat = np.array([x.pdot for x in pulsars+highlight \ if x.p is not None and x.pdot is not None \ and x.rrat==True]) global scatt_rrats if periods_rrat.size: scatter_options = MARKER_OPTIONS.copy() scatter_options.update(RRAT_MARKER) scatt_rrats = ax.scatter(np.log10(periods_rrat), np.log10(pdots_rrat), \ **scatter_options) if not rrats: # Hide RRATs for now scatt_rrats.set_visible(False) else: scatt_rrats = None # Mark magnetars periods_mag = np.array([x.p for x in pulsars \ if x.p is not None and x.pdot is not None \ and x.magnetar==True]) pdots_mag = np.array([x.pdot for x in pulsars \ if x.p is not None and x.pdot is not None \ and x.magnetar==True]) global scatt_magnetars if periods_mag.size: scatter_options = MARKER_OPTIONS.copy() scatter_options.update(MAGNETAR_MARKER) scatt_magnetars = ax.scatter(np.log10(periods_mag), np.log10(pdots_mag), \ **scatter_options) if not magnetars: # Hide magnetars for now scatt_magnetars.set_visible(False) else: scatt_magnetars = None # Mark sgrs periods_sgr = np.array([x.p for x in pulsars \ if x.p is not None and x.pdot is not None \ and x.sgr==True]) pdots_sgr = np.array([x.pdot for x in pulsars \ if x.p is not None and x.pdot is not None \ and x.sgr==True]) global scatt_sgr if periods_sgr.size: scatter_options = MARKER_OPTIONS.copy() scatter_options.update(SGR_MARKER) scatt_sgr = ax.scatter(np.log10(periods_sgr), np.log10(pdots_sgr), \ **scatter_options) if not sgr: # Hide sgr for now scatt_sgr.set_visible(False) else: scatt_sgr = None # Mark axps periods_axp = np.array([x.p for x in pulsars \ if x.p is not None and x.pdot is not None \ and x.axp==True]) pdots_axp = np.array([x.pdot for x in pulsars \ if x.p is not None and x.pdot is not None \ and x.axp==True]) global scatt_axp if periods_axp.size: scatter_options = MARKER_OPTIONS.copy() scatter_options.update(AXP_MARKER) scatt_axp = ax.scatter(np.log10(periods_axp), np.log10(pdots_axp), \ **scatter_options) if not axp: # Hide axp for now scatt_axp.set_visible(False) else: scatt_axp = None # Mark SNRs periods_snr = np.array([x.p for x in pulsars \ if x.p is not None and x.pdot is not None \ and x.snr==True]) pdots_snr = np.array([x.pdot for x in pulsars \ if x.p is not None and x.pdot is not None \ and x.snr==True]) global scatt_snrs if periods_snr.size: scatter_options = MARKER_OPTIONS.copy() scatter_options.update(SNR_MARKER) scatt_snrs = ax.scatter(np.log10(periods_snr), np.log10(pdots_snr), \ **scatter_options) if not snrs: # Hide SNRs for now scatt_snrs.set_visible(False) else: scatt_snrs = None plt.xlabel("log Period") plt.ylabel(r"log $\mathsf{\dot P}$") plt.title(r"$\mathsf{P-\dot P}$ Diagram") plt.xlim(np.log10((0.001, 100))) plt.ylim(np.log10((10**-22, 10**-8))) draw_lines(bsurfs, edots, ages) print "Plot Inventory:" print "\tNumber of pulsars:", (periods.size + numhl) print "\tNumber highlighted:", numhl print "\tNumber of RRATs:", periods_rrat.size print "\tNumber of magnetars:", periods_mag.size print "\tNumber of binaries:", periods_bnry.size print "\tNumber of SNR associations:", periods_snr.size | eea39b5ffa200e203a272a577114678d00cc2707 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/3485/eea39b5ffa200e203a272a577114678d00cc2707/pyppdot.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3207,
67,
892,
12,
84,
332,
87,
5913,
16,
366,
750,
5099,
22850,
6487,
4158,
5646,
33,
8381,
16,
8354,
2323,
33,
8381,
16,
521,
4447,
2758,
5913,
33,
8381,
16,
4556,
5453,
33,
8381,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3207,
67,
892,
12,
84,
332,
87,
5913,
16,
366,
750,
5099,
22850,
6487,
4158,
5646,
33,
8381,
16,
8354,
2323,
33,
8381,
16,
521,
4447,
2758,
5913,
33,
8381,
16,
4556,
5453,
33,
8381,
... |
if os.path.isfile(source): | print source if os.path.isfile(source): | def run(self): ext = os.path.extsep + 'aar' files = [] | 6fd2ca9da870d10d94cb5bb83f3715dba02073e2 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/14305/6fd2ca9da870d10d94cb5bb83f3715dba02073e2/qtui.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
12,
2890,
4672,
1110,
273,
1140,
18,
803,
18,
408,
10814,
397,
296,
69,
297,
11,
1390,
273,
5378,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
12,
2890,
4672,
1110,
273,
1140,
18,
803,
18,
408,
10814,
397,
296,
69,
297,
11,
1390,
273,
5378,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
def format_url(self, url): link, fragment = url if fragment: return link + " else: return link | def format_url(self, url): link, fragment = url if fragment: return link + " else: return link | def format_url(self, url): | a870cf128cdb4a76dc17c103820d903325fd04fb /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12029/a870cf128cdb4a76dc17c103820d903325fd04fb/wcnew.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
740,
67,
718,
12,
2890,
16,
880,
4672,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
740,
67,
718,
12,
2890,
16,
880,
4672,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
cmd = 'SELECT * FROM Job WHERE JobID=\'%s\'' % jobID | cmd = 'SELECT * FROM Jobs WHERE JobID=\'%s\'' % jobID | def getAllJobAttributes(self, jobID): """ Get all Job Attributes for a given jobID. Return a dictionary with all Job Attributes, return an empty dictionary if matching job found """ | a8a7475e68fb31dba0fc51b7e74cad44aab6c0e4 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/12864/a8a7475e68fb31dba0fc51b7e74cad44aab6c0e4/JobDB.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
5514,
2278,
2498,
12,
2890,
16,
28913,
4672,
3536,
968,
777,
3956,
9055,
364,
279,
864,
28913,
18,
2000,
279,
3880,
598,
777,
3956,
9055,
16,
327,
392,
1008,
3880,
309,
3607,
1719,
1392,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
5514,
2278,
2498,
12,
2890,
16,
28913,
4672,
3536,
968,
777,
3956,
9055,
364,
279,
864,
28913,
18,
2000,
279,
3880,
598,
777,
3956,
9055,
16,
327,
392,
1008,
3880,
309,
3607,
1719,
1392,... |
[{y: w^2 + 2, x: 2*w}, {y: w^2 + 2*w, x: 2*w + 2}, {y: w^2 + w, x: 2*w + 1}] | [{y: w^2 + 2, x: 2*w}, {y: w^2 + w, x: 2*w + 1}, {y: w^2 + 2*w, x: 2*w + 2}] | def variety(self): """ Return the variety of self. | c0c793aac678b1cde891e5571c63539215bdc007 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/9890/c0c793aac678b1cde891e5571c63539215bdc007/multi_polynomial_ideal.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1394,
14369,
12,
2890,
4672,
3536,
2000,
326,
1394,
14369,
434,
365,
18,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1394,
14369,
12,
2890,
4672,
3536,
2000,
326,
1394,
14369,
434,
365,
18,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
self.cmbLicenses.SetValue('Creative Commons') | self.cmbLicenses.SetValue(self.__l_classes['standard']) | def getLicenseClasses(self): """Calls the SOAP API via proxy to get a list of all available license class identifiers.""" | a47b1cd7897157eec77dae6384081722f101c689 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2607/a47b1cd7897157eec77dae6384081722f101c689/license.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
9014,
5142,
4818,
12,
2890,
4672,
3536,
10125,
326,
16434,
1491,
3970,
2889,
358,
336,
279,
666,
434,
777,
2319,
8630,
667,
9863,
12123,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
9014,
5142,
4818,
12,
2890,
4672,
3536,
10125,
326,
16434,
1491,
3970,
2889,
358,
336,
279,
666,
434,
777,
2319,
8630,
667,
9863,
12123,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
def fields_view_get(self, cr, uid, view_id=None, view_type='form', context=None, toolbar=False,submenu=False): print context | def fields_view_get(self, cr, uid, view_id=None, view_type='form', context=None, toolbar=False, submenu=False): height = 2 * 25; | def fields_view_get(self, cr, uid, view_id=None, view_type='form', context=None, toolbar=False,submenu=False): print context message = { 'title' : '_(Deliver Products)', 'info' : '_(Delivery Information)', 'button' : '_Deliver' } if context: if context.get('product_receive', False): message = { 'title' : '_(Receive Products)', 'info' : '_(Receive Information)', 'button' : '_Receive' } result = super(stock_partial_move, self).fields_view_get(cr, uid, view_id, view_type, context, toolbar,submenu) move_obj = self.pool.get('stock.move') move_ids = context.get('active_ids', False) move_ids = move_obj.search(cr, uid, [('id','in',move_ids)]) _moves_arch_lst = """<form string="%(title)s"> <separator colspan="4" string="%(info)s"/> <field name="date" colspan="2"/> <separator colspan="4" string="Move Detail"/> """ % message _moves_fields = result['fields'] if move_ids and view_type in ['form']: for m in move_obj.browse(cr, uid, move_ids, context): if m.state in ('done', 'cancel'): continue _moves_fields.update({ 'move%s_product_id'%(m.id) : { 'string': _('Product'), 'type' : 'many2one', 'relation': 'product.product', 'required' : True, 'readonly' : True, }, 'move%s_product_qty'%(m.id) : { 'string': _('Quantity'), 'type' : 'float', 'required': True, }, 'move%s_product_uom'%(m.id) : { 'string': _('Product UOM'), 'type' : 'many2one', 'relation': 'product.uom', 'required' : True, 'readonly' : True, }, 'move%s_prodlot_id'%(m.id): { 'string': _('Production Lot'), 'type': 'many2one', 'relation': 'stock.production.lot', } }) | 957482358411ba62d97e82c425318e0708be09e2 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/7397/957482358411ba62d97e82c425318e0708be09e2/stock_partial_move.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1466,
67,
1945,
67,
588,
12,
2890,
16,
4422,
16,
4555,
16,
1476,
67,
350,
33,
7036,
16,
1476,
67,
723,
2218,
687,
2187,
819,
33,
7036,
16,
12748,
33,
8381,
16,
27539,
33,
8381,
4672,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1466,
67,
1945,
67,
588,
12,
2890,
16,
4422,
16,
4555,
16,
1476,
67,
350,
33,
7036,
16,
1476,
67,
723,
2218,
687,
2187,
819,
33,
7036,
16,
12748,
33,
8381,
16,
27539,
33,
8381,
4672,... |
bytessent = comminfo[mycommid]['socket'].send(*args) break | comm_socket = comminfo[mycommid]['socket'] (read_sock, write_sock, exception_sock) = select.select([], [comm_socket], [comm_socket], 0.0) if not len(write_sock) == 0: bytessent = write_sock[0].send(*args) break if not len(write_sock) == 0: raise Exception("Socket Error") | def send(self,*args): """ <Purpose> Sends data on a socket. It may send fewer bytes than requested. | b9b372cc8fd497db57be0afc1989160a1935160e /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7995/b9b372cc8fd497db57be0afc1989160a1935160e/emulcomm.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1366,
12,
2890,
16,
14,
1968,
4672,
3536,
411,
10262,
4150,
34,
2479,
87,
501,
603,
279,
2987,
18,
282,
2597,
2026,
1366,
27886,
1731,
2353,
3764,
18,
2,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1366,
12,
2890,
16,
14,
1968,
4672,
3536,
411,
10262,
4150,
34,
2479,
87,
501,
603,
279,
2987,
18,
282,
2597,
2026,
1366,
27886,
1731,
2353,
3764,
18,
2,
-100,
-100,
-100,
-100,
-100,
... |
if not self.info: self.info = 'Session with no streams' | self.info = self.info or 'Session with no streams' self.chat = None for stream in streams: if isinstance(stream, MSRPChat): self.chat = stream break | def update_info(self, streams): self.info = '/'.join([type(stream).__name__ for stream in streams]) if not self.info: self.info = 'Session with no streams' self.manager.update_prompt() | 065a8951372a2fc8bb180e9cd1617916808b15b6 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/5703/065a8951372a2fc8bb180e9cd1617916808b15b6/sip_session2.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1089,
67,
1376,
12,
2890,
16,
8205,
4672,
365,
18,
1376,
273,
2023,
18,
5701,
3816,
723,
12,
3256,
2934,
972,
529,
972,
364,
1407,
316,
8205,
5717,
365,
18,
1376,
273,
365,
18,
1376,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1089,
67,
1376,
12,
2890,
16,
8205,
4672,
365,
18,
1376,
273,
2023,
18,
5701,
3816,
723,
12,
3256,
2934,
972,
529,
972,
364,
1407,
316,
8205,
5717,
365,
18,
1376,
273,
365,
18,
1376,
... |
self.mvc(-self.cpos) | self.end() | def p_key(self): """Process a keypress""" | 2f8a6fc5d2a1d15396f9afa0faefa34f34f8b538 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/6932/2f8a6fc5d2a1d15396f9afa0faefa34f34f8b538/cli.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
293,
67,
856,
12,
2890,
4672,
3536,
2227,
279,
498,
1028,
8395,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
293,
67,
856,
12,
2890,
4672,
3536,
2227,
279,
498,
1028,
8395,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
I = self | I = MPolynomialIdeal(P, self.interreduced_basis()) | def triangular_decomposition(self, algorithm=None, singular=singular_default): r""" Decompose zero-dimensional ideal ``self`` into triangular sets. | 2e77da7bd3536d7d46d876604202e92fd26e21c9 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/9417/2e77da7bd3536d7d46d876604202e92fd26e21c9/multi_polynomial_ideal.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
6882,
13077,
67,
323,
24388,
12,
2890,
16,
4886,
33,
7036,
16,
10048,
33,
17835,
67,
1886,
4672,
436,
8395,
26824,
4150,
3634,
17,
31236,
23349,
12176,
2890,
10335,
1368,
6882,
13077,
1678... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
6882,
13077,
67,
323,
24388,
12,
2890,
16,
4886,
33,
7036,
16,
10048,
33,
17835,
67,
1886,
4672,
436,
8395,
26824,
4150,
3634,
17,
31236,
23349,
12176,
2890,
10335,
1368,
6882,
13077,
1678... |
def WriteHeader(self, file): """Writes header to file""" file.Write( "// This file is auto-generated. DO NOT EDIT!\n" "\n") | def WriteHeader(self, file): """Writes header to file""" file.Write( "// This file is auto-generated. DO NOT EDIT!\n" "\n") | 8a837bb40249dfd9f0a2736abdab4f64592c99db /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9392/8a837bb40249dfd9f0a2736abdab4f64592c99db/build_gles2_cmd_buffer.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
16101,
12,
2890,
16,
585,
4672,
3536,
8368,
1446,
358,
585,
8395,
585,
18,
3067,
12,
16177,
1220,
585,
353,
3656,
17,
11168,
18,
5467,
4269,
21307,
24440,
82,
6,
1548,
82,
7923,
2,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
16101,
12,
2890,
16,
585,
4672,
3536,
8368,
1446,
358,
585,
8395,
585,
18,
3067,
12,
16177,
1220,
585,
353,
3656,
17,
11168,
18,
5467,
4269,
21307,
24440,
82,
6,
1548,
82,
7923,
2,
-10... | |
this = apply(_quickfix.new_NewSeqNo, args) | this = _quickfix.new_NewSeqNo(*args) | def __init__(self, *args): this = apply(_quickfix.new_NewSeqNo, args) try: self.this.append(this) except: self.this = this | 7e632099fd421880c8c65fb0cf610d338d115ee9 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/8819/7e632099fd421880c8c65fb0cf610d338d115ee9/quickfix.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
380,
1968,
4672,
333,
273,
389,
19525,
904,
18,
2704,
67,
1908,
6926,
2279,
30857,
1968,
13,
775,
30,
365,
18,
2211,
18,
6923,
12,
2211,
13,
1335,
30,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
380,
1968,
4672,
333,
273,
389,
19525,
904,
18,
2704,
67,
1908,
6926,
2279,
30857,
1968,
13,
775,
30,
365,
18,
2211,
18,
6923,
12,
2211,
13,
1335,
30,
... |
def handle_command(self, command): | def handle_command(self, command, ignore_dnd = False): | def handle_command(self, command): """Excecutes commands from other components""" if isinstance(command, NotificationCommand): jid = JID(node_or_jid=command.jid) jid_text = jid.bare().as_utf8() text = command.text # Check if contact is DoNotDisturb. If so, queue the message for delayed delivery try: contact = self.contacts[jid_text] if contact.is_dnd(): contact.messages.append(command) return except KeyError: pass self.send_message(jid, text) | a1cc2264545cca0c4816c4c22adeab84d31909fd /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/888/a1cc2264545cca0c4816c4c22adeab84d31909fd/xmppbot.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1640,
67,
3076,
12,
2890,
16,
1296,
16,
2305,
67,
72,
4880,
273,
1083,
4672,
3536,
424,
27832,
993,
4364,
628,
1308,
4085,
8395,
225,
309,
1549,
12,
3076,
16,
8050,
2189,
4672,
14201,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1640,
67,
3076,
12,
2890,
16,
1296,
16,
2305,
67,
72,
4880,
273,
1083,
4672,
3536,
424,
27832,
993,
4364,
628,
1308,
4085,
8395,
225,
309,
1549,
12,
3076,
16,
8050,
2189,
4672,
14201,
... |
if len(ns_record['durations'][0]) >= MIN_RELEVANT_COUNT: | if ns_record == ns_summary[0]: | def CreateReport(self, format='ascii', output_fp=None, csv_path=None): # First generate all of the charts necessary. if format == 'ascii': lowest_latency = self._LowestLatencyAsciiChart() mean_duration = self._MeanRequestAsciiChart() else: lowest_latency = None mean_duration = None | acd7f6e8e8dab899104a9e6fa17649296d41db99 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/4170/acd7f6e8e8dab899104a9e6fa17649296d41db99/reporter.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1788,
4820,
12,
2890,
16,
740,
2218,
9184,
2187,
876,
67,
7944,
33,
7036,
16,
6101,
67,
803,
33,
7036,
4672,
225,
468,
5783,
2103,
777,
434,
326,
23194,
4573,
18,
309,
740,
422,
296,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1788,
4820,
12,
2890,
16,
740,
2218,
9184,
2187,
876,
67,
7944,
33,
7036,
16,
6101,
67,
803,
33,
7036,
4672,
225,
468,
5783,
2103,
777,
434,
326,
23194,
4573,
18,
309,
740,
422,
296,
... |
""" | r""" | def __call__(self, x, check=True): """ Return $x$ coerced into this free monoid. | ca219b0206186a194cdf8e11d2082bec654536f8 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/9417/ca219b0206186a194cdf8e11d2082bec654536f8/string_monoid.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
1991,
972,
12,
2890,
16,
619,
16,
866,
33,
5510,
4672,
436,
8395,
2000,
271,
92,
8,
12270,
72,
1368,
333,
4843,
6921,
839,
18,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
1991,
972,
12,
2890,
16,
619,
16,
866,
33,
5510,
4672,
436,
8395,
2000,
271,
92,
8,
12270,
72,
1368,
333,
4843,
6921,
839,
18,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
Either integers `n,d` such that `d>0`, `gcd(n,d)=1`, `n/d=u (mod m)`, and `T*d*|n|< m`, or ``None``. | Either integers `n,d` such that `d>0`, `\mathop{\mathrm{gcd}}(n,d)=1`, `n/d=u \bmod m`, and `T \* d \* |n| < m`, or ``None``. | def mqrr_rational_reconstruction(u, m, T): r""" Maximal Quotient Rational Reconstruction. FOR research purposes only - this is pure Python, so slow. INPUT: - ``u, m, T`` - integers such that `m > u \ge 0,` `T > 0`. OUTPUT: Either integers `n,d` such that `d>0`, `gcd(n,d)=1`, `n/d=u (mod m)`, and `T*d*|n|< m`, or ``None``. Reference: Monagan, Maximal Quotient Rational Reconstruction: An Almost Optimal Algorithm for Rational Reconstruction (page 11) This algorithm is probabilistic. EXAMPLES:: sage: mqrr_rational_reconstruction(21,3100,13) (21, 1) """ if u == 0: if m > T: return (0,1) else: return None n, d = 0, 0 t0, r0 = 0, m t1, r1 = 1, u while r1 != 0 and r0 > T: q = r0/r1 # C division implicit floor if q > T: n, d, T = r1, t1, q r0, r1 = r1, r0 - q*r1 t0, t1 = t1, t0 - q*t1 if d != 0 and GCD(n,d) == 1: return (n,d) return None | c5cc436e94fabaa9fa38f5732aff2c6f8ca3be27 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/9890/c5cc436e94fabaa9fa38f5732aff2c6f8ca3be27/arith.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
18327,
523,
67,
7034,
287,
67,
266,
591,
4549,
12,
89,
16,
312,
16,
399,
4672,
436,
8395,
4238,
2840,
4783,
352,
1979,
534,
8371,
868,
591,
4549,
18,
225,
12108,
283,
3072,
13694,
1338... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
18327,
523,
67,
7034,
287,
67,
266,
591,
4549,
12,
89,
16,
312,
16,
399,
4672,
436,
8395,
4238,
2840,
4783,
352,
1979,
534,
8371,
868,
591,
4549,
18,
225,
12108,
283,
3072,
13694,
1338... |
if self.mAnchor.y() > deskRect.height() / 2: y = - self.height() - 40 else: y = 40 self.move(deskRect.width() - self.width() - 10, self.mAnchor.y() + y) | def paintEvent(self, evt): mask = QtGui.QRegion() # 10, 10, self.width() - 20, self.height() - 20) | 39bea4b1fe63d275e8bcaf3233e0f7c504feadda /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/1208/39bea4b1fe63d275e8bcaf3233e0f7c504feadda/gui_pyqt.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
12574,
1133,
12,
2890,
16,
6324,
4672,
3066,
273,
23425,
18,
53,
5165,
1435,
468,
1728,
16,
1728,
16,
365,
18,
2819,
1435,
300,
4200,
16,
365,
18,
4210,
1435,
300,
4200,
13,
2,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
12574,
1133,
12,
2890,
16,
6324,
4672,
3066,
273,
23425,
18,
53,
5165,
1435,
468,
1728,
16,
1728,
16,
365,
18,
2819,
1435,
300,
4200,
16,
365,
18,
4210,
1435,
300,
4200,
13,
2,
-100,
... | |
if security_group: params['DBSecurityGroup'] = security_group | if security_groups: l = [] for group in security_groups: if isinstance(group, DBSecurityGroup): l.append(group.name) else: l.append(group) self.build_list_params(params, l, 'DBSecurityGroups.member') | def create_dbinstance(self, id, allocated_storage, instance_class, master_username, master_password, port=3306, engine='MySQL5.1', db_name=None, param_group=None, security_group=None, availability_zone=None, preferred_maintenance_window=None, backup_retention_period=None, preferred_backup_window=None): """ Create a new DBInstance. | 0de9387d433f1588981b85fd6480bf101bda7fab /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/1098/0de9387d433f1588981b85fd6480bf101bda7fab/__init__.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
752,
67,
1966,
1336,
12,
2890,
16,
612,
16,
11977,
67,
5697,
16,
791,
67,
1106,
16,
4171,
67,
5053,
16,
4171,
67,
3664,
16,
1756,
33,
3707,
7677,
16,
4073,
2218,
28723,
25,
18,
21,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
752,
67,
1966,
1336,
12,
2890,
16,
612,
16,
11977,
67,
5697,
16,
791,
67,
1106,
16,
4171,
67,
5053,
16,
4171,
67,
3664,
16,
1756,
33,
3707,
7677,
16,
4073,
2218,
28723,
25,
18,
21,
... |
sql = "SELECT page_code, section_code, sort_order, template_code, data, only_registered, only_admin, only_sysadmin FROM page" | sql = "SELECT page_code, section_code, sort_order, template_code, data, only_dynamic, only_registered, only_admin, only_sysadmin FROM page" | def __init__(self): self.data = {} sql = "SELECT page_code, section_code, sort_order, template_code, data, only_registered, only_admin, only_sysadmin FROM page" cursor = db.select(sql) while (1): row = cursor.fetchone() if row==None: break newPage = Page() newPage.load(row) self.data[newPage.code] = newPage | ed1bc055edb13298c1e3d4f5e0fed00270953ecc /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/1319/ed1bc055edb13298c1e3d4f5e0fed00270953ecc/WebLayer.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
4672,
365,
18,
892,
273,
2618,
1847,
273,
315,
4803,
1363,
67,
710,
16,
2442,
67,
710,
16,
1524,
67,
1019,
16,
1542,
67,
710,
16,
501,
16,
1338,
67,
14507,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
4672,
365,
18,
892,
273,
2618,
1847,
273,
315,
4803,
1363,
67,
710,
16,
2442,
67,
710,
16,
1524,
67,
1019,
16,
1542,
67,
710,
16,
501,
16,
1338,
67,
14507,... |
self.raiseErr("md5s differ for hardlink") | self.printErr("md5s differ for hardlink") | def readPayload(self, func, filenames=None, extract=None): self.__openFd(96 + self.sigdatasize + self.hdrdatasize) devinode = {} # this will contain possibly hardlinked files filenamehash = {} # full filename of all files if filenames == None: filenames = self.getFilenames() if filenames: fileinfo = zip(filenames, self["fileflags"], self["filemodes"], self["filemtimes"], self["filedevices"], self["fileinodes"], self["fileusername"], self["filegroupname"], self["filerdevs"], self["filelinktos"], xrange(len(self["fileinodes"]))) for (fn, flag, mode, mtime, dev, inode, user, group, rdev, linkto, i) in fileinfo: if flag & (RPMFILE_GHOST | RPMFILE_EXCLUDE): continue filenamehash[fn] = fileinfo[i] if S_ISREG(mode): #di = (dev, inode) #if not devinode.has_key(di): # devinode[di] = [] #devinode[di].append(i) devinode.setdefault((dev, inode), []).append(i) for di in devinode.keys(): if len(devinode[di]) <= 1: del devinode[di] # sanity check hardlinks if self.verify: for hardlinks in devinode.values(): j = hardlinks[0] mode = self["filemodes"][j] mtime = self["filemtimes"][j] size = self["filesizes"][j] fmd5 = self["filemd5s"][j] for j in hardlinks[1:]: # dev/inode are already guaranteed to be the same if self["filemodes"][j] != mode: self.raiseErr("modes differ for hardlink") if self["filemtimes"][j] != mtime: self.raiseErr("mtimes differ for hardlink") if self["filesizes"][j] != size: self.raiseErr("sizes differ for hardlink") if self["filemd5s"][j] != fmd5: self.raiseErr("md5s differ for hardlink") cpiosize = self.sig.getOne("payloadsize") archivesize = self.hdr.getOne("archivesize") if archivesize != None: if cpiosize == None: cpiosize = archivesize elif cpiosize != archivesize: self.printErr("wrong archive size") size_in_sig = self.sig.getOne("size_in_sig") if size_in_sig != None: size_in_sig -= self.hdrdatasize if self["payloadcompressor"] in [None, "gzip"]: if size_in_sig != None and size_in_sig >= 8: size_in_sig -= 8 fd = PyGZIP(self.filename, self.fd, cpiosize, size_in_sig) #import gzip #fd = gzip.GzipFile(fileobj=self.fd) elif self["payloadcompressor"] == "bzip2": import bz2, cStringIO if size_in_sig != None: payload = self.fd.read(size_in_sig) else: payload = self.fd.read() fd = cStringIO.StringIO(bz2.decompress(payload)) else: self.printErr("unknown payload compression") return if self["payloadformat"] not in [None, "cpio"]: self.printErr("unknown payload format") return c = CPIO(self.filename, fd, self.issrc, cpiosize) if c.readCpio(func, filenamehash, devinode, filenames, extract) == None: pass # error output is already done else: for filename in filenamehash.iterkeys(): self.printErr("file not in cpio: %s" % filename) if extract and len(devinode.keys()): self.printErr("hardlinked files remain from cpio") self.closeFd() | ff6f3ccd07a935de383f3580dfd65ceb3dc8bc57 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/1143/ff6f3ccd07a935de383f3580dfd65ceb3dc8bc57/oldpyrpm.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
855,
6110,
12,
2890,
16,
1326,
16,
9066,
33,
7036,
16,
2608,
33,
7036,
4672,
365,
16186,
3190,
27263,
12,
10525,
397,
365,
18,
7340,
13178,
554,
397,
365,
18,
16587,
13178,
554,
13,
44... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
855,
6110,
12,
2890,
16,
1326,
16,
9066,
33,
7036,
16,
2608,
33,
7036,
4672,
365,
16186,
3190,
27263,
12,
10525,
397,
365,
18,
7340,
13178,
554,
397,
365,
18,
16587,
13178,
554,
13,
44... |
if self.save_path and self.save_path[-1] != sep: self.save_path += sep | if self.save_path: if isdir(self.save_path) and self.save_path[-1] != sep: self.save_path += sep | def run(self): """ Run this script! """ if self.path: if not exists(self.path): raise IOError, self.path + " does not exist!" | e799495f2b74a4ff3c4f9f262a278b3757981e64 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2399/e799495f2b74a4ff3c4f9f262a278b3757981e64/scripts.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
12,
2890,
4672,
3536,
1939,
333,
2728,
5,
3536,
309,
365,
18,
803,
30,
309,
486,
1704,
12,
2890,
18,
803,
4672,
1002,
8340,
16,
365,
18,
803,
397,
315,
1552,
486,
1005,
4442,
2... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
12,
2890,
4672,
3536,
1939,
333,
2728,
5,
3536,
309,
365,
18,
803,
30,
309,
486,
1704,
12,
2890,
18,
803,
4672,
1002,
8340,
16,
365,
18,
803,
397,
315,
1552,
486,
1005,
4442,
2... |
"""Mapping of language-dependent interpreted text role names to already imported directive functions.""" | """Mapping of local or language-dependent interpreted text role names to role functions.""" | def role_fn(name, rawtext, text, lineno, inliner): code... | edf51e38b3836f1a648b65216a4d06c1f619e4e6 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8194/edf51e38b3836f1a648b65216a4d06c1f619e4e6/roles.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2478,
67,
4293,
12,
529,
16,
1831,
955,
16,
977,
16,
7586,
16,
316,
7511,
264,
4672,
981,
2777,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2478,
67,
4293,
12,
529,
16,
1831,
955,
16,
977,
16,
7586,
16,
316,
7511,
264,
4672,
981,
2777,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-... |
self.console = console self.brokers = [] self.packages = {} self.seqMgr = SequenceManager() self.cv = Condition() self.syncSequenceList = [] self.getResult = [] self.error = None | self.console = console self.brokers = [] self.packages = {} self.seqMgr = SequenceManager() self.cv = Condition() self.syncSequenceList = [] self.getResult = [] self.error = None self.bindingKeyList = self._bindingKeys(rcvObjects, rcvEvents, rcvHeartbeats) self.manageConnections = manageConnections if (manageConnections): raise Exception("manageConnections - not yet implemented") | def __init__(self, console=None): """ Initialize a session. If the console argument is provided, the more advanced asynchronous features are available. If console is defaulted, the session will operate in a simpler, synchronous manner. """ self.console = console self.brokers = [] self.packages = {} self.seqMgr = SequenceManager() self.cv = Condition() self.syncSequenceList = [] self.getResult = [] self.error = None | 37e775a6ad926a53362cdd8bc0a5083eb6cfa230 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/197/37e775a6ad926a53362cdd8bc0a5083eb6cfa230/qmfconsole.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
2983,
33,
7036,
4672,
3536,
9190,
279,
1339,
18,
225,
971,
326,
2983,
1237,
353,
2112,
16,
326,
1898,
16111,
9007,
4467,
854,
2319,
18,
225,
971,
2983,
3... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
2983,
33,
7036,
4672,
3536,
9190,
279,
1339,
18,
225,
971,
326,
2983,
1237,
353,
2112,
16,
326,
1898,
16111,
9007,
4467,
854,
2319,
18,
225,
971,
2983,
3... |
context.log("ADRES bank:%03o, FB:%03o, EB:%03o" % (bank, context.fbank, context.ebank)) | context.log("ADRES: bank:%02o FB:%02o EB:%02o" % (bank, context.fbank, context.ebank)) | def parse_ADRES(self, context, operands): expr = AddressExpression(context, operands) if expr.complete: (bank, offset) = context.memmap.pseudoToSegmented(expr.value) if bank == None or offset == None: context.error("invalid address %06o" % expr.value) else: context.currentRecord.code = [ offset ] context.currentRecord.target = expr.value context.currentRecord.complete = True context.log("ADRES bank:%03o, FB:%03o, EB:%03o" % (bank, context.fbank, context.ebank)) if (context.memmap.isSwitched(expr.value) and bank != context.fbank and bank != context.ebank): context.error("bank (%03o) does not match current F bank (%03o) or E bank (%03o)" % (bank, context.fbank, context.ebank)) | 547fd2bc1e4463728858ab42cd4e9cf927abdd7a /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/8152/547fd2bc1e4463728858ab42cd4e9cf927abdd7a/directive.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1109,
67,
1880,
7031,
12,
2890,
16,
819,
16,
14883,
4672,
3065,
273,
5267,
2300,
12,
2472,
16,
14883,
13,
309,
3065,
18,
6226,
30,
261,
10546,
16,
1384,
13,
273,
819,
18,
3917,
1458,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1109,
67,
1880,
7031,
12,
2890,
16,
819,
16,
14883,
4672,
3065,
273,
5267,
2300,
12,
2472,
16,
14883,
13,
309,
3065,
18,
6226,
30,
261,
10546,
16,
1384,
13,
273,
819,
18,
3917,
1458,
... |
r""" Lucas numbers (beginning with 1): $L(n) = L(n-1) + L(n-2)$ with $L(1) = 1$, $L(2) = 3$. INPUT: n -- positive integer OUTPUT: integer -- function value EXAMPLES: sage: a = sloane.A000204; a Lucas numbers (beginning at 1): L(n) = L(n-1) + L(n-2), L(2) = 3. sage: a(1) 1 sage: a(8) 47 sage: a(200) 627376215338105766356982006981782561278127 sage: a(-4) Traceback (most recent call last): ... ValueError: input n (=-4) must be a positive integer sage: a.list(12) [1, 3, 4, 7, 11, 18, 29, 47, 76, 123, 199, 322] sage: a(0) Traceback (most recent call last): ... ValueError: input n (=0) must be a positive integer AUTHOR: -- Jaap Spies (2007-01-18) """ def __init__(self): | def __init__(self): r""" Lucas numbers (beginning with 1): $L(n) = L(n-1) + L(n-2)$ with $L(1) = 1$, $L(2) = 3$. INPUT: n -- positive integer OUTPUT: integer -- function value EXAMPLES: sage: a = sloane.A000204; a Lucas numbers (beginning at 1): L(n) = L(n-1) + L(n-2), L(2) = 3. sage: a(1) 1 sage: a(8) 47 sage: a(200) 627376215338105766356982006981782561278127 sage: a(-4) Traceback (most recent call last): ... ValueError: input n (=-4) must be a positive integer sage: a.list(12) [1, 3, 4, 7, 11, 18, 29, 47, 76, 123, 199, 322] sage: a(0) Traceback (most recent call last): ... ValueError: input n (=0) must be a positive integer AUTHOR: -- Jaap Spies (2007-01-18) """ | def _eval(self, n): return arith.moebius(n) | d2510d4232e17c9cea8d1a0271c6be68a81d317b /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9417/d2510d4232e17c9cea8d1a0271c6be68a81d317b/sloane_functions.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
8622,
12,
2890,
16,
290,
4672,
327,
225,
419,
483,
18,
8683,
24008,
8384,
12,
82,
13,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
8622,
12,
2890,
16,
290,
4672,
327,
225,
419,
483,
18,
8683,
24008,
8384,
12,
82,
13,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
for id in targetIDs: | for id in targetIds: | def getTargetStates(targetIDs): states = [] for id in targetIDs: states.append(doc.getState(id)) return states | fa72944ebe9c1a503306771dfea5ec44a5989133 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/10998/fa72944ebe9c1a503306771dfea5ec44a5989133/interpreter.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
8571,
7629,
12,
3299,
5103,
4672,
5493,
273,
5378,
364,
612,
316,
1018,
2673,
30,
5493,
18,
6923,
12,
2434,
18,
588,
1119,
12,
350,
3719,
327,
5493,
282,
2,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
8571,
7629,
12,
3299,
5103,
4672,
5493,
273,
5378,
364,
612,
316,
1018,
2673,
30,
5493,
18,
6923,
12,
2434,
18,
588,
1119,
12,
350,
3719,
327,
5493,
282,
2,
-100,
-100,
-100,
-100,
-10... |
print ("Saving bayes database with %d spam and %d good messages" % (bayes.nspam, bayes.nham)) | print "Saving bayes database with %d spam and %d good messages" %\ (bayes.nspam, bayes.nham) | def SaveBayes(self): bayes = self.bayes if self.verbose: print ("Saving bayes database with %d spam and %d good messages" % (bayes.nspam, bayes.nham)) print " ->", self.bayes_filename cPickle.dump(bayes, open(self.bayes_filename,"wb"), 1) if self.verbose: print " ->", self.message_db_filename cPickle.dump(self.message_db, open(self.message_db_filename,"wb"), 1) self.bayes_dirty = False | 2007caa2ba23b5c7ad8ac9e2d48d9cdad6205944 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/6126/2007caa2ba23b5c7ad8ac9e2d48d9cdad6205944/manager.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
7074,
38,
528,
281,
12,
2890,
4672,
324,
528,
281,
273,
365,
18,
70,
528,
281,
309,
365,
18,
11369,
30,
1172,
315,
24660,
324,
528,
281,
2063,
598,
738,
72,
24824,
471,
738,
72,
7494... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
7074,
38,
528,
281,
12,
2890,
4672,
324,
528,
281,
273,
365,
18,
70,
528,
281,
309,
365,
18,
11369,
30,
1172,
315,
24660,
324,
528,
281,
2063,
598,
738,
72,
24824,
471,
738,
72,
7494... |
def reversed (ip): if ip == None: loginfo.log ('unresolved') else: loginfo.log (ip) reverse_lookup (question[0], reversed, servers) else: def resolved (request): model = list (request.dns_question) if request.dns_resources == None: model.append ('') elif len (request.dns_resources) > 1: model.append (netstring.encode ( request.dns_resources )) elif request.dns_resources: model.append (request.dns_resources[0]) else: model.append ('') model.append (request.dns_peer[0]) loginfo.log (netstring.encode (model)) lookup (question, resolved, servers) | question = (ip_peer.in_addr_arpa (question[0]), question[1]) lookup (question, resolved, servers) | def reversed (ip): if ip == None: loginfo.log ('unresolved') else: loginfo.log (ip) | 79906933ee4150b4a6dc21c74f2849f7ab831a23 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2577/79906933ee4150b4a6dc21c74f2849f7ab831a23/dns_client.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
9553,
261,
625,
4672,
309,
2359,
422,
599,
30,
613,
1376,
18,
1330,
7707,
318,
11792,
6134,
469,
30,
613,
1376,
18,
1330,
261,
625,
13,
225,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
9553,
261,
625,
4672,
309,
2359,
422,
599,
30,
613,
1376,
18,
1330,
7707,
318,
11792,
6134,
469,
30,
613,
1376,
18,
1330,
261,
625,
13,
225,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
selStartLine = self.GetLineFromPos(selStartPos) startPos = self.GetLineStartPos(selStartLine) selEndLine = self.GetLineFromPos(selEndPos) if selEndPos != self.GetLineStartPos(selEndLine): | selStartLine = self.LineFromPosition(selStartPos) startPos = self.PositionFromLine(selStartLine) selEndLine = self.LineFromPosition(selEndPos) if selEndPos != self.PositionFromLine(selEndLine): | def reselectSelectionAsBlock(self): selStartPos, selEndPos = self.GetSelection() selStartLine = self.GetLineFromPos(selStartPos) startPos = self.GetLineStartPos(selStartLine) selEndLine = self.GetLineFromPos(selEndPos) if selEndPos != self.GetLineStartPos(selEndLine): selEndLine = selEndLine + 1 endPos = self.GetLineStartPos(selEndLine) | f2d9de9b91fb77d3ba05e87f8e4569617622e102 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/4325/f2d9de9b91fb77d3ba05e87f8e4569617622e102/SourceViews.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
400,
13759,
6233,
1463,
1768,
12,
2890,
4672,
357,
1685,
1616,
16,
357,
1638,
1616,
273,
365,
18,
967,
6233,
1435,
357,
1685,
1670,
273,
365,
18,
1670,
1265,
2555,
12,
1786,
1685,
1616,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
400,
13759,
6233,
1463,
1768,
12,
2890,
4672,
357,
1685,
1616,
16,
357,
1638,
1616,
273,
365,
18,
967,
6233,
1435,
357,
1685,
1670,
273,
365,
18,
1670,
1265,
2555,
12,
1786,
1685,
1616,
... |
if hasattr(ncvar,'ncattrs'): propnames = ncvar.ncattrs() else: propnames = [p for p in dir(ncvar) if p not in ('assignValue','getValue','typecode')] | propnames = getNcAttributes(ncvar) | def getProperties(self): nc = self.store.getcdf() ncvar = nc.variables[self.ncvarname] if hasattr(ncvar,'ncattrs'): propnames = ncvar.ncattrs() else: propnames = [p for p in dir(ncvar) if p not in ('assignValue','getValue','typecode')] return dict([(key,getattr(ncvar,key)) for key in propnames]) | d595d0017e779a5b06214ee30e710e5635deab41 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/381/d595d0017e779a5b06214ee30e710e5635deab41/data.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
9392,
12,
2890,
4672,
8194,
273,
365,
18,
2233,
18,
588,
24799,
1435,
8194,
1401,
273,
8194,
18,
7528,
63,
2890,
18,
14202,
1401,
529,
65,
2270,
1973,
273,
11069,
71,
2498,
12,
14202,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
9392,
12,
2890,
4672,
8194,
273,
365,
18,
2233,
18,
588,
24799,
1435,
8194,
1401,
273,
8194,
18,
7528,
63,
2890,
18,
14202,
1401,
529,
65,
2270,
1973,
273,
11069,
71,
2498,
12,
14202,
... |
self.titlebar = itemlistwidgets.ChannelTitlebar(feed_info.name, icon, add_icon_box=add_icon_box) | if feed_info.is_directory_feed: self.titlebar = itemlistwidgets.ItemListTitlebar(feed_info.name, icon, add_icon_box=add_icon_box) else: self.titlebar = itemlistwidgets.ChannelTitlebar(feed_info.name, icon, add_icon_box=add_icon_box) self.titlebar.connect('save-search', self._on_save_search) | def build_widget(self): feed_info = widgetutil.get_feed_info(self.id) icon = imagepool.get(feed_info.thumbnail, size=(41, 41)) self._make_item_views() | 787676c657fe0d42ed94024aea1f79399d87799e /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/12354/787676c657fe0d42ed94024aea1f79399d87799e/feedcontroller.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1361,
67,
6587,
12,
2890,
4672,
4746,
67,
1376,
273,
3604,
1367,
18,
588,
67,
7848,
67,
1376,
12,
2890,
18,
350,
13,
4126,
273,
1316,
6011,
18,
588,
12,
7848,
67,
1376,
18,
14904,
16... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1361,
67,
6587,
12,
2890,
4672,
4746,
67,
1376,
273,
3604,
1367,
18,
588,
67,
7848,
67,
1376,
12,
2890,
18,
350,
13,
4126,
273,
1316,
6011,
18,
588,
12,
7848,
67,
1376,
18,
14904,
16... |
(r"('| | (r"('| | def get_tokens_unprocessed(self, text): for index, token, value in \ RegexLexer.get_tokens_unprocessed(self, text): if token is Name: if value in self._functions: yield index, Name.Function, value continue elif '.' in value: a, b = value.split('.') yield index, Name, a yield index + len(a), Text, u'.' yield index + len(a) + 1, Name, b continue yield index, token, value | a6790ba1b41192e40167ae16bd32dfb7ef1aeab8 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2885/a6790ba1b41192e40167ae16bd32dfb7ef1aeab8/agile.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
67,
7860,
67,
318,
11005,
12,
2890,
16,
977,
4672,
364,
770,
16,
1147,
16,
460,
316,
521,
15078,
13356,
18,
588,
67,
7860,
67,
318,
11005,
12,
2890,
16,
977,
4672,
309,
1147,
35... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
67,
7860,
67,
318,
11005,
12,
2890,
16,
977,
4672,
364,
770,
16,
1147,
16,
460,
316,
521,
15078,
13356,
18,
588,
67,
7860,
67,
318,
11005,
12,
2890,
16,
977,
4672,
309,
1147,
35... |
j = self.synchro_state.join(self.menu).join(self.target, self.menu.c.id == self.target.c.fk_menu).join(self.entity, self.entity.c.id == self.target.c.fk_entity) | j = self.synchro_state.join(self.menu).join(self.target).join(self.entity) | def getLocationSynchroState(self, uuid): """ Attempt to see if a location uuid needs to be synced, or not | 2880be534df2a5d2d6d366502de120ea4c8c1819 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/5988/2880be534df2a5d2d6d366502de120ea4c8c1819/__init__.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
13312,
55,
2515,
303,
1119,
12,
2890,
16,
3822,
4672,
3536,
12864,
358,
2621,
309,
279,
2117,
3822,
4260,
358,
506,
21048,
16,
578,
486,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
13312,
55,
2515,
303,
1119,
12,
2890,
16,
3822,
4672,
3536,
12864,
358,
2621,
309,
279,
2117,
3822,
4260,
358,
506,
21048,
16,
578,
486,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
""" fl_set_canvas_attributes(pObject, mask, pXSetWindowAttributes) | """ fl_set_canvas_attributes(pObject, mask, pXSetWindowAttributes) @param pObject : pointer to canvas object @param mask : mask num. @param pXSetWindowAttributes : pointer to XSetWindowAttributes | def fl_set_canvas_attributes(pObject, mask, pXSetWindowAttributes): """ fl_set_canvas_attributes(pObject, mask, pXSetWindowAttributes) """ _fl_set_canvas_attributes = cfuncproto( load_so_libforms(), "fl_set_canvas_attributes", None, [cty.POINTER(FL_OBJECT), cty.c_uint, cty.POINTER(XSetWindowAttributes)], """void fl_set_canvas_attributes(FL_OBJECT * ob, unsigned int mask, XSetWindowAttributes * xswa) """) uimask = convert_to_uint(mask) keep_elem_refs(pObject, mask, pXSetWindowAttributes, uimask) _fl_set_canvas_attributes(pObject, uimask, pXSetWindowAttributes) | 197b7293f22d67e3c0811c99134c95d0d70c0239 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/2429/197b7293f22d67e3c0811c99134c95d0d70c0239/library.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1183,
67,
542,
67,
15424,
67,
4350,
12,
84,
921,
16,
3066,
16,
293,
60,
694,
3829,
2498,
4672,
3536,
1183,
67,
542,
67,
15424,
67,
4350,
12,
84,
921,
16,
3066,
16,
293,
60,
694,
38... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1183,
67,
542,
67,
15424,
67,
4350,
12,
84,
921,
16,
3066,
16,
293,
60,
694,
3829,
2498,
4672,
3536,
1183,
67,
542,
67,
15424,
67,
4350,
12,
84,
921,
16,
3066,
16,
293,
60,
694,
38... |
if (rsp != "[]"): | if (rsp != "[('PythonOptionTest2', 'new_value2')]"): self.fail(`rsp`) def test_PythonOption_remove2_conf(self): c = VirtualHost("*", ServerName("test_PythonOption_remove2"), DocumentRoot(DOCUMENT_ROOT), Directory(DOCUMENT_ROOT, SetHandler("mod_python"), PythonHandler("tests::PythonOption_items"), PythonOption('PythonOptionTest'), PythonOption('PythonOptionTest2 "new_value2"'), PythonOption('PythonOptionTest3 new_value3'), PythonDebug("On"))) return str(c) def test_PythonOption_remove2(self): print "\n * Testing PythonOption remove2" rsp = self.vhost_get("test_PythonOption_remove2") if (rsp != "[('PythonOptionTest2', 'new_value2'), ('PythonOptionTest2', 'new_value2')]"): | def test_PythonOption_remove(self): | 90e1203063c35b4b1e9ab0d7e3dc019f3b2f9273 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/10002/90e1203063c35b4b1e9ab0d7e3dc019f3b2f9273/test.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
15774,
1895,
67,
4479,
12,
2890,
4672,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
15774,
1895,
67,
4479,
12,
2890,
4672,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
self.config.setting["id3v2_encoding"] == "iso-8859-1" | self.config.setting["id3v2_encoding"] = "iso-8859-1" | def save_options(self): self.config.setting["clear_existing_tags"] = \ self.ui.clear_existing_tags.isChecked() self.config.setting["write_id3v1"] = \ self.ui.write_id3v1.isChecked() self.config.setting["write_id3v23"] = \ self.ui.write_id3v23.isChecked() if self.ui.enc_iso88591.isChecked(): self.config.setting["id3v2_encoding"] == "iso-8859-1" elif self.ui.enc_utf16.isChecked(): self.config.setting["id3v2_encoding"] == "utf-16" else: self.config.setting["id3v2_encoding"] == "utf-8" self.config.setting["strip_ape_tags"] = \ self.ui.strip_ape_tags.isChecked() | d68d7863d1b9df374786d5ea775de5ff1a821e01 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/12415/d68d7863d1b9df374786d5ea775de5ff1a821e01/tags.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1923,
67,
2116,
12,
2890,
4672,
365,
18,
1425,
18,
8920,
9614,
8507,
67,
11711,
67,
4156,
11929,
273,
521,
365,
18,
4881,
18,
8507,
67,
11711,
67,
4156,
18,
291,
11454,
1435,
365,
18,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1923,
67,
2116,
12,
2890,
4672,
365,
18,
1425,
18,
8920,
9614,
8507,
67,
11711,
67,
4156,
11929,
273,
521,
365,
18,
4881,
18,
8507,
67,
11711,
67,
4156,
18,
291,
11454,
1435,
365,
18,
... |
AutoGenC.Append(LibraryString[0], Dict) else: AutoGenC.Append(LibraryString[3], Dict) | self.AutoGenC.Append(LibraryString[0], Dict) else: self.AutoGenC.Append(LibraryString[3], Dict) | def CreateLibraryConstructorCode(self): # # Library Constructors # ConstructorList = [] for lib in self.LibraryList: if len(lib.ConstructorList) <= 0: continue ConstructorList.extend(lib.ConstructorList) Dict = {'Type':'Constructor', 'Function':ConstructorList} if self.Module.ModuleType == 'BASE': if len(ConstructorList) == 0: AutoGenC.Append(LibraryString[0], Dict) else: AutoGenC.Append(LibraryString[3], Dict) elif self.Module.ModuleType in ['PEI_CORE','PEIM']: if len(ConstructorList) == 0: AutoGenC.Append(LibraryString[1], Dict) else: AutoGenC.Append(LibraryString[4], Dict) elif self.Module.ModuleType in ['DXE_CORE','DXE_DRIVER','DXE_SMM_DRIVER','DXE_RUNTIME_DRIVER','DXE_SAL_DRIVER','UEFI_DRIVER','UEFI_APPLICATION']: if len(ConstructorList) == 0: AutoGenC.Append(LibraryString[2], Dict) else: AutoGenC.Append(LibraryString[5], Dict) | 502c7f8e852c690f3cb00ab852f2902d8eb0db4f /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/914/502c7f8e852c690f3cb00ab852f2902d8eb0db4f/AutoGen.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1788,
9313,
6293,
1085,
12,
2890,
4672,
468,
468,
18694,
14291,
1383,
468,
11417,
682,
273,
5378,
364,
2561,
316,
365,
18,
9313,
682,
30,
309,
562,
12,
2941,
18,
6293,
682,
13,
1648,
3... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1788,
9313,
6293,
1085,
12,
2890,
4672,
468,
468,
18694,
14291,
1383,
468,
11417,
682,
273,
5378,
364,
2561,
316,
365,
18,
9313,
682,
30,
309,
562,
12,
2941,
18,
6293,
682,
13,
1648,
3... |
def lastLocations(JobSpecId,dbCur = None): | except: if dbCur1==None: dbCur.execute("ROLLBACK") dbCur.close() raise def lastLocations(JobSpecId,dbCur1 = None): try: | def isRegistered(JobSpecId,dbCur = None): if(dbCur==None): conn=connect() dbCur=conn.cursor() dbCur.execute("START TRANSACTION") sqlStr='SELECT JobType FROM js_JobSpec WHERE '+ \ 'JobSpecID="'+JobSpecId+'";' dbCur.execute(sqlStr) #due to the schema we either get 0 or 1 row back. rows=dbCur.fetchall() dbCur.execute("COMMIT") if(dbCur==None): dbCur.close() if len(rows)==0: return False return True | 9f18c1304f6cef721d60405dc38d789bd4e6cc08 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8887/9f18c1304f6cef721d60405dc38d789bd4e6cc08/JobStateInfoAPIMySQL.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
353,
10868,
12,
2278,
1990,
548,
16,
1966,
2408,
273,
599,
4672,
309,
12,
1966,
2408,
631,
7036,
4672,
1487,
33,
3612,
1435,
1319,
2408,
33,
4646,
18,
9216,
1435,
1319,
2408,
18,
8837,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
353,
10868,
12,
2278,
1990,
548,
16,
1966,
2408,
273,
599,
4672,
309,
12,
1966,
2408,
631,
7036,
4672,
1487,
33,
3612,
1435,
1319,
2408,
33,
4646,
18,
9216,
1435,
1319,
2408,
18,
8837,
... |
modules[key] = other_modules[key] | thiscoverdir = os.path.dirname(os.path.abspath(filename)) tx = None try: os.makedirs(thiscoverdir) except OSError, x: tx = x if not os.path.isdir(thiscoverdir): if tx: raise tx raise exceptions.IOError, "unknown error prevented creation of directory: %s" % thiscoverdir if "." in modulename: finalname = modulename.split(".")[-1] listfilename = os.path.join(thiscoverdir, finalname + ".cover") else: listfilename = os.path.join(thiscoverdir, modulename + ".cover") try: lines = open(filename, 'r').readlines() except IOError, err: sys.stderr.write("trace: Could not open %s for reading because: %s - skipping\n" % (`filename`, err)) continue try: outfile = open(listfilename, 'w') except IOError, err: sys.stderr.write( '%s: Could not open %s for writing because: %s" \ "- skipping\n' % ("trace", `listfilename`, err)) continue if show_missing: executable_linenos = find_executable_linenos(filename) else: executable_linenos = {} n_lines = 0 n_hits = 0 lines_hit = per_file[key] for i in range(len(lines)): line = lines[i] if lines_hit.has_key(i+1): outfile.write('%5d: ' % lines_hit[i+1]) n_hits = n_hits + 1 n_lines = n_lines + 1 elif blank.match(line): outfile.write(' . ') else: if executable_linenos.has_key(i+1) and \ string.find(lines[i], string.join([' outfile.write('>>>>>> ') else: outfile.write(' '*7) n_lines = n_lines + 1 outfile.write(string.expandtabs(lines[i], 8)) outfile.close() if summary and n_lines: percent = int(100 * n_hits / n_lines) sums[modulename] = n_lines, percent, modulename, filename if summary and sums: mods = sums.keys() mods.sort() print "lines cov% module (path)" for m in mods: n_lines, percent, modulename, filename = sums[m] print "%5d %3d%% %s (%s)" % sums[m] if self.outfile: try: pickle.dump((self.counts, self.calledfuncs,), open(self.outfile, 'w'), 1) except IOError, err: sys.stderr.write("cannot save counts files because %s" % err) | def update(self, other): """Merge in the data from another CoverageResults""" counts = self.counts other_counts = other.counts modules = self.modules other_modules = other.modules | a30eacf45815c14b480972c5097677fac83ed4c0 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8546/a30eacf45815c14b480972c5097677fac83ed4c0/trace.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1089,
12,
2890,
16,
1308,
4672,
3536,
6786,
316,
326,
501,
628,
4042,
30856,
3447,
8395,
6880,
273,
365,
18,
8008,
1308,
67,
8008,
273,
1308,
18,
8008,
4381,
273,
365,
18,
6400,
1308,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1089,
12,
2890,
16,
1308,
4672,
3536,
6786,
316,
326,
501,
628,
4042,
30856,
3447,
8395,
6880,
273,
365,
18,
8008,
1308,
67,
8008,
273,
1308,
18,
8008,
4381,
273,
365,
18,
6400,
1308,
... |
5, 7, 13\}`, or the congruence subgroup Gamma0(p) where p is one of these | 5, 7, 13\}`, or the congruence subgroup Gamma0(p) where `p` is one of these | def OverconvergentModularForms(prime, weight, radius, base_ring=QQ, prec = 20, char = None): r""" Create a space of overconvergent `p`-adic modular forms of level `\Gamma_0(p)`, over the given base ring. The base ring need not be a `p`-adic ring (the spaces we compute with typically have bases over `\QQ`). INPUT: - ``prime`` - a prime number `p`, which must be one of the primes `\{2, 3, 5, 7, 13\}`, or the congruence subgroup Gamma0(p) where p is one of these primes. - ``weight`` - an integer (which at present must be 0 or `\ge 2`), the weight. - ``radius`` - a rational number in the interval `\left( 0, \frac{p}{p+1} \right)`, the radius of overconvergence. - ``base_ring`` (default: `\QQ`), a ring over which to compute. This need not be a `p`-adic ring. - ``prec`` - an integer (default: 20), the number of `q`-expansion terms to compute. - ``char`` - a Dirichlet character modulo `p` or ``None`` (the default). Here ``None`` is interpreted as the trivial character modulo `p`. The character `\chi` and weight `k` must satisfy `(-1)^k = \chi(-1)`, and the base ring must contain an element `v` such that `{\rm ord}_p(v) = \frac{12 r}{p-1}` where `r` is the radius of overconvergence (and `{\rm ord}_p` is normalised so `{\rm ord}_p(p) = 1`). EXAMPLES:: sage: OverconvergentModularForms(3, 0, 1/2) Space of 3-adic 1/2-overconvergent modular forms of weight-character 0 over Rational Field sage: OverconvergentModularForms(3, 16, 1/2) Space of 3-adic 1/2-overconvergent modular forms of weight-character 16 over Rational Field sage: OverconvergentModularForms(3, 3, 1/2, char = DirichletGroup(3,QQ).0) Space of 3-adic 1/2-overconvergent modular forms of weight-character (3, 3, [-1]) over Rational Field """ if is_Gamma0(prime) or is_Gamma1(prime): prime = prime.level() else: prime = ZZ(prime) if char is None: char = trivial_character(prime, base_ring=QQ) if int(prime) not in [2, 3, 5, 7, 13]: raise ValueError, "p must be one of {2, 3, 5, 7, 13}" key = (prime, weight, radius, base_ring, prec, char) if __ocmfdict.has_key(key): w = __ocmfdict[key] M = w() if not (M is None): return M M = OverconvergentModularFormsSpace(*key) __ocmfdict[key] = weakref.ref(M) return M | b526c3d210b712e2659f9ba0fab589c5ff5e9564 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/9890/b526c3d210b712e2659f9ba0fab589c5ff5e9564/genus0.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
10752,
591,
502,
75,
319,
1739,
2490,
18529,
12,
16382,
16,
3119,
16,
5725,
16,
1026,
67,
8022,
33,
53,
53,
16,
13382,
273,
4200,
16,
1149,
273,
599,
4672,
436,
8395,
1788,
279,
3476,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
10752,
591,
502,
75,
319,
1739,
2490,
18529,
12,
16382,
16,
3119,
16,
5725,
16,
1026,
67,
8022,
33,
53,
53,
16,
13382,
273,
4200,
16,
1149,
273,
599,
4672,
436,
8395,
1788,
279,
3476,
... |
return http.Response(stream=message(msg, '/forgotpass')) | return HTMLResponse(stream=message(msg, '/forgotpass')) | def error(msg): return http.Response(stream=message(msg, '/forgotpass')) | 5da3b1cd72cd4a94421b8bc914d7957cdbe6b56d /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9890/5da3b1cd72cd4a94421b8bc914d7957cdbe6b56d/twist.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
555,
12,
3576,
4672,
327,
3982,
1064,
12,
3256,
33,
2150,
12,
3576,
16,
1173,
1884,
13212,
5466,
26112,
225,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
555,
12,
3576,
4672,
327,
3982,
1064,
12,
3256,
33,
2150,
12,
3576,
16,
1173,
1884,
13212,
5466,
26112,
225,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-1... |
if text.has_key('preferencestoken') and len(text['preferencestoken']) > 2: | if text.has_key('preferencestoken'): | def _getUserData(self, text, sysop = False, force = True): """ Get the user data from a wiki page data. | a9c13738b1b5c85b2f392d56a95826baae0cb816 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/4404/a9c13738b1b5c85b2f392d56a95826baae0cb816/wikipedia.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
588,
19265,
12,
2890,
16,
977,
16,
2589,
556,
273,
1083,
16,
2944,
273,
1053,
4672,
3536,
968,
326,
729,
501,
628,
279,
9050,
1363,
501,
18,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
588,
19265,
12,
2890,
16,
977,
16,
2589,
556,
273,
1083,
16,
2944,
273,
1053,
4672,
3536,
968,
326,
729,
501,
628,
279,
9050,
1363,
501,
18,
2,
-100,
-100,
-100,
-100,
-100,
-100,... |
if len(args) != 1: | if len(args) < 1: | def throwerror(): raise Exception("Problem") | 438dfcf0dfa4cc5dec9f90b3c50bb32edb070c2f /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/10201/438dfcf0dfa4cc5dec9f90b3c50bb32edb070c2f/arff2fann.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
604,
1636,
13332,
1002,
1185,
2932,
13719,
7923,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
604,
1636,
13332,
1002,
1185,
2932,
13719,
7923,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
self.bank_account = bank; self.cash_account = cash; | self.bank_account = bank self.cash_account = cash | def save(self): super(Group, self).save() # Create default accounts if not self.account_set.count(): bank = Account(name=ugettext('Bank'), slug='bank', type=Account.ASSET_ACCOUNT, group=self) bank.save() cash = Account(name=ugettext('Cash'), slug='cash', type=Account.ASSET_ACCOUNT, group=self) cash.save() | dd21905328d5180b4ef777fea89801f8d126a803 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/12178/dd21905328d5180b4ef777fea89801f8d126a803/models.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1923,
12,
2890,
4672,
2240,
12,
1114,
16,
365,
2934,
5688,
1435,
468,
1788,
805,
9484,
309,
486,
365,
18,
4631,
67,
542,
18,
1883,
13332,
11218,
273,
6590,
12,
529,
33,
89,
29967,
2668... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1923,
12,
2890,
4672,
2240,
12,
1114,
16,
365,
2934,
5688,
1435,
468,
1788,
805,
9484,
309,
486,
365,
18,
4631,
67,
542,
18,
1883,
13332,
11218,
273,
6590,
12,
529,
33,
89,
29967,
2668... |
(prefix + in_line + '\n') for in_line in text.split('\n')) | [(prefix + in_line + '\n') for in_line in text.split('\n')]) | def comment_begin(self, text): """Return commented version of the passed text WITHOUT end of line/comment.""" prefix = '.\\" ' out_text = ''.join( (prefix + in_line + '\n') for in_line in text.split('\n')) return out_text | 049dd2230cab12b2b7e756a7d09c4104947f8492 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/1278/049dd2230cab12b2b7e756a7d09c4104947f8492/manpage.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2879,
67,
10086,
12,
2890,
16,
977,
4672,
3536,
990,
31813,
1177,
434,
326,
2275,
977,
13601,
5069,
679,
434,
980,
19,
3469,
12123,
1633,
273,
2418,
27576,
296,
596,
67,
955,
273,
875,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2879,
67,
10086,
12,
2890,
16,
977,
4672,
3536,
990,
31813,
1177,
434,
326,
2275,
977,
13601,
5069,
679,
434,
980,
19,
3469,
12123,
1633,
273,
2418,
27576,
296,
596,
67,
955,
273,
875,
... |
m.make('bstall',type='bstall',xyz=(-35,-5,village_height)) | m.make('bstall',type='bstall',xyz=(-41,-5,village_height)) | def default(mapeditor): | c976c23fb381d95c582f4df84eb5bffe8b9263c9 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12932/c976c23fb381d95c582f4df84eb5bffe8b9263c9/define_world.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
805,
12,
1458,
9177,
4672,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
805,
12,
1458,
9177,
4672,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-1... |
pickle.dump (version, fo) pickle.dump (self.Counter, fo) pickle.dump (self.Jobs, fo) pickle.dump (self.Workers, fo) | cPickle.dump (version, fo) cPickle.dump (self.Counter, fo) cPickle.dump (self.Jobs, fo) cPickle.dump (self.Workers, fo) | def write (self, fo): version = DBVersion pickle.dump (version, fo) pickle.dump (self.Counter, fo) pickle.dump (self.Jobs, fo) pickle.dump (self.Workers, fo) self._UpdatedDb = False #self.dump () | 20fa391431a1b90b1c3b6ef9899a92a7762119ad /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/1233/20fa391431a1b90b1c3b6ef9899a92a7762119ad/server.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1045,
261,
2890,
16,
18261,
4672,
1177,
273,
2383,
1444,
276,
17968,
298,
18,
8481,
261,
1589,
16,
18261,
13,
276,
17968,
298,
18,
8481,
261,
2890,
18,
4789,
16,
18261,
13,
276,
17968,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1045,
261,
2890,
16,
18261,
4672,
1177,
273,
2383,
1444,
276,
17968,
298,
18,
8481,
261,
1589,
16,
18261,
13,
276,
17968,
298,
18,
8481,
261,
2890,
18,
4789,
16,
18261,
13,
276,
17968,
... |
,"<a href=\"https://%s/%s.phtml\">Supported Commands</a>" % \ | ,"<a href=\"https://%s/%s.php\">Supported Commands</a>" % \ | def process_groupchat_cmd(self, room, res, cmd): """ I actually process the groupchat commands and do stuff """ if (not ROSTER[room].has_key(res)): self.send_groupchat(room, "%s: I don't know who are!"%(res,)) return aff = ROSTER[room][res]['affiliation'] | c69054290b49f3e8064c13375e17f608b5d298d3 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/11180/c69054290b49f3e8064c13375e17f608b5d298d3/iemchatbot.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1207,
67,
1655,
10880,
67,
4172,
12,
2890,
16,
7725,
16,
400,
16,
1797,
4672,
3536,
467,
6013,
1207,
326,
1041,
10880,
4364,
471,
741,
10769,
3536,
309,
261,
902,
534,
4005,
654,
63,
1... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1207,
67,
1655,
10880,
67,
4172,
12,
2890,
16,
7725,
16,
400,
16,
1797,
4672,
3536,
467,
6013,
1207,
326,
1041,
10880,
4364,
471,
741,
10769,
3536,
309,
261,
902,
534,
4005,
654,
63,
1... |
else return | else: return self | def __float__(self): if len(self._varlist) != 1: raise TypeError, "Cannot coerce %s to a float" % self | dcdf1015fa1f010f55df22355abc64aa2b83c8f4 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/9417/dcdf1015fa1f010f55df22355abc64aa2b83c8f4/calculus.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
5659,
972,
12,
2890,
4672,
309,
562,
12,
2890,
6315,
1401,
1098,
13,
480,
404,
30,
1002,
3580,
16,
315,
4515,
12270,
738,
87,
358,
279,
1431,
6,
738,
365,
2,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
5659,
972,
12,
2890,
4672,
309,
562,
12,
2890,
6315,
1401,
1098,
13,
480,
404,
30,
1002,
3580,
16,
315,
4515,
12270,
738,
87,
358,
279,
1431,
6,
738,
365,
2,
-100,
-100,
-100,
... |
widget=field_map[2] | used_widget=field_map[2] | def formfield_function(field): """ Custom formfield function, so we can inject our own form fields. The mapping of model fields to form fields is defined in 'MODEL_TO_FORM_FIELD_MAP'. It uses the default django mapping as fallback, if there is no match in our custom map. field -- a model field """ for field_map in MODEL_TO_FORM_FIELD_MAP: if isinstance(field, field_map[0]): used_widget = None if field.choices: # the normal django field forms.TypedChoiceField is wired hard # within the original db/models/fields.py. # If we use our custom Select widget, we also have to pass in # some additional validation field attributes. used_widget = Select(attrs={ 'extra_field_attrs':{ 'required':not field.blank, 'help_text':field.help_text, } }) elif len(field_map) == 3: widget=field_map[2] if used_widget: return field.formfield(form_class=field_map[1], widget=used_widget) return field.formfield(form_class=field_map[1]) # return the default formfield, if there is no equivalent return field.formfield() | 2c75a4b6c675c16af6cecfe8fe103e65b4e7d856 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/12008/2c75a4b6c675c16af6cecfe8fe103e65b4e7d856/models.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
646,
1518,
67,
915,
12,
1518,
4672,
3536,
6082,
646,
1518,
445,
16,
1427,
732,
848,
4590,
3134,
4953,
646,
1466,
18,
1021,
2874,
434,
938,
1466,
358,
646,
1466,
353,
2553,
316,
296,
17... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
646,
1518,
67,
915,
12,
1518,
4672,
3536,
6082,
646,
1518,
445,
16,
1427,
732,
848,
4590,
3134,
4953,
646,
1466,
18,
1021,
2874,
434,
938,
1466,
358,
646,
1466,
353,
2553,
316,
296,
17... |
attrs.append('domain="[(\'journal_id\', \'=\', '+str(journal.id)+'),(\'type\',\'<>\',\'view\'), (\'type\',\'<>\',\'closed\')]" on_change="onchange_account_id(account_id)"') | attrs.append('domain="[(\'journal_id\', \'=\', '+str(journal.id)+'),(\'type\',\'<>\',\'view\'), (\'type\',\'<>\',\'closed\')]" on_change="onchange_account_id(account_id, partner_id)"') elif field.field == 'partner_id': attrs.append('on_change="onchange_partner_id(move_id,partner_id,account_id,debit,credit,date,((\'journal_id\' in context) and context[\'journal_id\']) or {})"') | def fields_view_get(self, cr, uid, view_id=None, view_type='form', context={}, toolbar=False): result = super(osv.osv, self).fields_view_get(cr, uid, view_id,view_type,context,toolbar=toolbar) if view_type=='tree' and 'journal_id' in context: title = self.view_header_get(cr, uid, view_id, view_type, context) journal = self.pool.get('account.journal').browse(cr, uid, context['journal_id']) | d96f05d8de12cfa447d79576c3d8b17cbcdda975 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7397/d96f05d8de12cfa447d79576c3d8b17cbcdda975/account_move_line.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1466,
67,
1945,
67,
588,
12,
2890,
16,
4422,
16,
4555,
16,
1476,
67,
350,
33,
7036,
16,
1476,
67,
723,
2218,
687,
2187,
819,
28793,
12748,
33,
8381,
4672,
563,
273,
2240,
12,
538,
90... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1466,
67,
1945,
67,
588,
12,
2890,
16,
4422,
16,
4555,
16,
1476,
67,
350,
33,
7036,
16,
1476,
67,
723,
2218,
687,
2187,
819,
28793,
12748,
33,
8381,
4672,
563,
273,
2240,
12,
538,
90... |
self.rfile = self.connection.makefile('rb', 0) self.wfile = self.connection.makefile('wb', 0) | self.rfile = self.connection.makefile('rb', self.rbufsize) self.wfile = self.connection.makefile('wb', self.wbufsize) | def setup(self): self.connection = self.request self.rfile = self.connection.makefile('rb', 0) self.wfile = self.connection.makefile('wb', 0) | b3d8e61600214b8139107c78c5a5193b383f169a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8125/b3d8e61600214b8139107c78c5a5193b383f169a/SocketServer.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3875,
12,
2890,
4672,
365,
18,
4071,
273,
365,
18,
2293,
365,
18,
86,
768,
273,
365,
18,
4071,
18,
6540,
768,
2668,
6731,
2187,
374,
13,
365,
18,
91,
768,
273,
365,
18,
4071,
18,
6... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3875,
12,
2890,
4672,
365,
18,
4071,
273,
365,
18,
2293,
365,
18,
86,
768,
273,
365,
18,
4071,
18,
6540,
768,
2668,
6731,
2187,
374,
13,
365,
18,
91,
768,
273,
365,
18,
4071,
18,
6... |
svn_info = subprocess.Popen(command, shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE).stdout.readlines() map = dict() for line in svn_info: | svn_info = subprocess.Popen(command, shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE).stdout.readlines() map = {} for line in svn_info: | def getBestMergePaths(url, revision): """ Takes an svn url and gets the associated revision. """ command = 'svn log ' + url + " -r "+str(revision) + " -v" svn_info = subprocess.Popen(command, shell=True, stdout=subprocess.PIPE, stderr=subprocess.PIPE).stdout.readlines() map = dict() for line in svn_info: #match = re.search(r"[\n\r ]+[MADUC][\n\r ]+/.*/src(.*)/.+", line) #match = re.search(r"[\n\r ]+[MADUC][\n\r ]+/(?:trunk|branches/\d+)/src(.*)/.+", line) match = re.search(r"[\n\r ]+[MADUC][\n\r ]+/(?:trunk|branches/\d+)/src([^ ]*)/[^ ]+", line) if match: map[match.group(1)] = match.group(1) return map.keys() | 44364270024fa1f0c4c0875635e210722b0d4b26 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/6076/44364270024fa1f0c4c0875635e210722b0d4b26/drover.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
28782,
6786,
4466,
12,
718,
16,
6350,
4672,
3536,
23004,
392,
5893,
82,
880,
471,
5571,
326,
3627,
6350,
18,
3536,
1296,
273,
296,
31505,
613,
296,
397,
880,
397,
315,
300,
86,
13773,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
28782,
6786,
4466,
12,
718,
16,
6350,
4672,
3536,
23004,
392,
5893,
82,
880,
471,
5571,
326,
3627,
6350,
18,
3536,
1296,
273,
296,
31505,
613,
296,
397,
880,
397,
315,
300,
86,
13773,
... |
return None path=_to_unicode(path) | return None | def ftp2fs(self, path_orig, data): path = self.ftpnorm(path_orig) if path and path=='/': return None path=_to_unicode(path) path2 = filter(None,path.split('/'))[1:] (cr, uid, pool) = data res = pool.get('document.directory').get_object(cr, uid, path2[:]) if not res: raise OSError(2, 'Not such file or directory.') return res | c882bf7db62b8e88398de75c35611c7330aaf528 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7397/c882bf7db62b8e88398de75c35611c7330aaf528/abstracted_fs.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
13487,
22,
2556,
12,
2890,
16,
589,
67,
4949,
16,
501,
4672,
589,
273,
365,
18,
1222,
7449,
535,
12,
803,
67,
4949,
13,
309,
589,
471,
589,
18920,
2473,
30,
327,
599,
589,
22,
273,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
13487,
22,
2556,
12,
2890,
16,
589,
67,
4949,
16,
501,
4672,
589,
273,
365,
18,
1222,
7449,
535,
12,
803,
67,
4949,
13,
309,
589,
471,
589,
18920,
2473,
30,
327,
599,
589,
22,
273,
... |
_spr_rect = Rect(spr.rect.topleft, spr.source_rect.size) | _spr_rect = Rect(spr.rect.topleft, spr.source_rect.size) | def draw(self, surface, bgd=None): """draw all sprites in the right order onto the passed surface. LayeredDirty.draw(surface, bgd=None): return Rect_list | d9760f3e4782abb02dd98080337626eefdad67ee /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/1298/d9760f3e4782abb02dd98080337626eefdad67ee/sprite.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3724,
12,
2890,
16,
9034,
16,
7611,
72,
33,
7036,
4672,
3536,
9446,
777,
1694,
24047,
316,
326,
2145,
1353,
10170,
326,
2275,
9034,
18,
12112,
329,
10785,
18,
9446,
12,
18266,
16,
7611,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3724,
12,
2890,
16,
9034,
16,
7611,
72,
33,
7036,
4672,
3536,
9446,
777,
1694,
24047,
316,
326,
2145,
1353,
10170,
326,
2275,
9034,
18,
12112,
329,
10785,
18,
9446,
12,
18266,
16,
7611,
... |
charset = properties.site_properties.getProperty('default_charset', | charset = properties.site_properties.getProperty('default_charset', | def __call__(self): properties = component.getUtility(IPropertiesTool) charset = properties.site_properties.getProperty('default_charset', 'utf-8').upper() self.request.response.setHeader('Content-Type', 'text/csv; charset=%s' % charset) self.request.response.setHeader('Content-disposition', 'attachment; filename=subscribers_%s_%s.csv' % (self.context.id, datetime.date.today().strftime("%Y%m%d"))) res = cStringIO.StringIO() writer = csv.writer(res, dialect=csv.excel, delimiter=csv_delimiter) for format in self.context.composers.keys(): for subscription in tuple(self.context.subscriptions.query(format=format)): row = [] for item in field.Fields(self.context.composers[format].schema).keys(): v = subscription.composer_data.get(item) or '' row.append(v.encode(charset)) | 6a225131d5814f4fcaf8f720f43c7e7816ee8a6b /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9843/6a225131d5814f4fcaf8f720f43c7e7816ee8a6b/channel.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
1991,
972,
12,
2890,
4672,
1790,
273,
1794,
18,
588,
6497,
12,
45,
2297,
6364,
13,
4856,
273,
1790,
18,
4256,
67,
4738,
18,
588,
1396,
2668,
1886,
67,
9999,
2187,
296,
3158,
17,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
1991,
972,
12,
2890,
4672,
1790,
273,
1794,
18,
588,
6497,
12,
45,
2297,
6364,
13,
4856,
273,
1790,
18,
4256,
67,
4738,
18,
588,
1396,
2668,
1886,
67,
9999,
2187,
296,
3158,
17,
... |
print os.environ['FREPPLE_HOME'] | def runTestSuite(): global debug # Executable to be used for the tests. Exported as an environment variable. # This default executable is the one valid for GCC cygwin and GCC *nux builds. os.environ['EXECUTABLE'] = "../../libtool --mode=execute ../../src/frepple" platform = 'GCC' # Directory names for tests and frepple_home testdir = os.path.dirname(os.path.abspath(sys.argv[0])) os.environ['FREPPLE_HOME'] = os.path.normpath(os.path.join(testdir, '..', 'bin')) # Parse the command line opts = [] tests = [] try: opts, tests = getopt.getopt(sys.argv[1:], "dvbh", ["debug", "vcc", "bcc", "help"]) except getopt.GetoptError: usage() for o, a in opts: if o in ("-v" "--vcc"): # Use executables generated by the Visual C++ compiler platform = 'VCC' os.environ['EXECUTABLE'] = "../../bin/frepple_vcc.exe"; elif o in ("-b" "--bcc"): # Use executables generated by the Borland C++ compiler platform = 'BCC' os.environ['EXECUTABLE'] = "../../bin/frepple_bcc.exe"; elif o in ("-d", "--debug"): debug = True elif o in ("-h", "--help"): # Print help information and exit usage() os.exit(1) # Argh... Special cases for that special platform again... if sys.platform == 'cygwin' and platform == 'VCC': # Test running with cygwin python but testing the vcc executable os.environ['FREPPLE_HOME'] = Popen( "cygpath --windows " + os.environ['FREPPLE_HOME'], stdout=PIPE, shell=True).communicate()[0].strip() if sys.platform == 'win32' and platform == 'GCC': # Test running with windows python but testing the cygwin executable os.environ['FREPPLE_HOME'] = Popen( "cygpath --unix " + os.environ['FREPPLE_HOME'], stdout=PIPE, shell=True).communicate()[0].strip() print os.environ['FREPPLE_HOME'] # Update the search path for shared libraries, such that the modules # can be picked up. # LD_LIBRARY_PATH variable for Linux, Solaris # LIBPATH for AIX # SHLIB_PATH for HPUX # PATH for windows, cygwin for var in ('LD_LIBRARY_PATH','LIBPATH','SHLIB_PATH','PATH'): if var in os.environ: os.environ[var] += os.pathsep + os.environ['FREPPLE_HOME'] else: os.environ[var] = os.environ['FREPPLE_HOME'] # Define a list with tests to run if len(tests) == 0: # No tests specified, so run them all subdirs = os.listdir(testdir) subdirs.sort() for i in subdirs: if i == '.svn' or os.path.isfile(i): continue tests.append(i) else: # A list of tests has been specified, and we now validate it for i in tests: if not os.path.isdir(os.path.join(testdir, i)): print "Warning: Test directory " + i + " doesn't exist" tests.remove(i) # Now define the test suite AllTests = unittest.TestSuite() for i in tests: i = os.path.normpath(i) tmp = os.path.join(testdir, i, i) if os.path.isfile(tmp) or os.path.isfile(tmp + '.exe'): # Type 1: (compiled) executable if platform == "GCC": AllTests.addTest(freppleTest(i, 'runExecutable')) elif os.path.isfile(os.path.join(testdir, i, 'runtest.py')): # Type 2: perl script runtest.pl available AllTests.addTest(freppleTest(i, 'runScript')) elif os.path.isfile(tmp + '.xml'): # Type 3: input xml file specified AllTests.addTest(freppleTest(i, 'runXML')) else: # Undetermined - not a test directory print "Warning: Unrecognized test in directory " + i # Finally, run the test suite now print "Running", AllTests.countTestCases(), \ "tests from directory", testdir, \ "with FREPPLE_HOME", os.environ['FREPPLE_HOME'] unittest.TextTestRunner(verbosity=2).run(AllTests) | 1b972629738dcf63482973584814dc6784bd8218 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/8604/1b972629738dcf63482973584814dc6784bd8218/runtest.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
4709,
13587,
13332,
2552,
1198,
468,
18342,
358,
506,
1399,
364,
326,
7434,
18,
1312,
1798,
487,
392,
3330,
2190,
18,
468,
1220,
805,
9070,
353,
326,
1245,
923,
225,
364,
611,
6743... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
4709,
13587,
13332,
2552,
1198,
468,
18342,
358,
506,
1399,
364,
326,
7434,
18,
1312,
1798,
487,
392,
3330,
2190,
18,
468,
1220,
805,
9070,
353,
326,
1245,
923,
225,
364,
611,
6743... | |
import App.config cfg = App.config.getConfiguration() sw=os.path.dirname(os.path.dirname(cfg.softwarehome)) for home in (cfg.instancehome, sw): r=_getPath(home, prefix, name, suffixes) if r is not None: return r | if result is None: import App.config cfg = App.config.getConfiguration() sw=os.path.dirname(os.path.dirname(cfg.softwarehome)) for home in (cfg.instancehome, sw): r=_getPath(home, prefix, name, suffixes) if r is not None: result = r if result is None: try: l = name.find('.') if l > 0: realName = name[l + 1:] toplevel = name[:l] m = __import__(toplevel) d = os.path.join(m.__path__[0], prefix, realName) for s in suffixes: if s: s="%s.%s" % (d, s) else: s=d if os.path.exists(s): result = s break except: pass return result | def getPath(prefix, name, checkProduct=1, suffixes=('',)): """Find a file in one of several relative locations Arguments: prefix -- The location, relative to some home, to look for the file name -- The name of the file. This must not be a path. checkProduct -- a flag indicating whether product directories should be used as additional hope ares to be searched. This defaults to a true value. If this is true and the name contains a dot, then the text before the dot is treated as a product name and the product package directory is used as anothe rhome. suffixes -- a sequences of file suffixes to check. By default, the name is used without a suffix. The search takes on multiple homes which are the instance home, the directory containing the directory containing the software home, and possibly product areas. """ d,n = path_split(name) if d: raise ValueError, ( 'The file name, %s, should be a simple file name' % name) if checkProduct: l = name.find('.') if l > 0: p = name[:l] n = name[l + 1:] for product_dir in Products.__path__: r = _getPath(product_dir, os.path.join(p, prefix), n, suffixes) if r is not None: return r import App.config cfg = App.config.getConfiguration() sw=os.path.dirname(os.path.dirname(cfg.softwarehome)) for home in (cfg.instancehome, sw): r=_getPath(home, prefix, name, suffixes) if r is not None: return r | 309d6f33d7c27eaa225f2babd6413398266fb74a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9658/309d6f33d7c27eaa225f2babd6413398266fb74a/Extensions.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4339,
12,
3239,
16,
508,
16,
866,
4133,
33,
21,
16,
18333,
33,
2668,
2187,
3719,
30,
3536,
3125,
279,
585,
316,
1245,
434,
11392,
3632,
7838,
225,
13599,
30,
225,
1633,
1493,
1021,
211... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4339,
12,
3239,
16,
508,
16,
866,
4133,
33,
21,
16,
18333,
33,
2668,
2187,
3719,
30,
3536,
3125,
279,
585,
316,
1245,
434,
11392,
3632,
7838,
225,
13599,
30,
225,
1633,
1493,
1021,
211... |
""" Generator that yield the images without license. It's based on a tool of the toolserver. """ | """ Generator that yield the files without license. It's based on a tool of the toolserver. """ | def untaggedGenerator(self, untaggedProject, limit): """ Generator that yield the images without license. It's based on a tool of the toolserver. """ lang = untaggedProject.split('.', 1)[0] project = '.%s' % untaggedProject.split('.', 1)[1] if lang == 'commons': link = 'http://toolserver.org/~daniel/WikiSense/UntaggedImages.php?wikifam=commons.wikimedia.org&since=-100d&until=&img_user_text=&order=img_timestamp&max=100&order=img_timestamp&format=html' else: link = 'http://toolserver.org/~daniel/WikiSense/UntaggedImages.php?wikilang=%s&wikifam=%s&order=img_timestamp&max=%s&ofs=0&max=%s' % (lang, project, limit, limit) text = self.site.getUrl(link, no_hostname = True) regexp = r"""<td valign='top' title='Name'><a href='http://.*?\.org/w/index\.php\?title=(.*?)'>.*?</a></td>""" results = re.findall(regexp, text) if results == []: wikipedia.output(link) raise NothingFound(u'Nothing found! Try to use the tool by yourself to be sure that it works!') else: for result in results: wikiPage = wikipedia.Page(self.site, result) yield wikiPage | 8913e1740cb89be9fcacfa047c0f66333f53d469 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/4404/8913e1740cb89be9fcacfa047c0f66333f53d469/checkimages.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
640,
23023,
3908,
12,
2890,
16,
640,
23023,
4109,
16,
1800,
4672,
3536,
10159,
716,
2824,
326,
1390,
2887,
8630,
18,
2597,
1807,
2511,
603,
279,
5226,
434,
326,
5226,
3567,
18,
3536,
330... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
640,
23023,
3908,
12,
2890,
16,
640,
23023,
4109,
16,
1800,
4672,
3536,
10159,
716,
2824,
326,
1390,
2887,
8630,
18,
2597,
1807,
2511,
603,
279,
5226,
434,
326,
5226,
3567,
18,
3536,
330... |
def runtest(test, generate, verbose, quiet, testdir=None, huntrleaks=False): | def runtest(test, generate, verbose, quiet, testdir=None, huntrleaks=False, debug=False): | def runtest(test, generate, verbose, quiet, testdir=None, huntrleaks=False): """Run a single test. test -- the name of the test generate -- if true, generate output, instead of running the test and comparing it to a previously created output file verbose -- if true, print more messages quiet -- if true, don't print 'skipped' messages (probably redundant) testdir -- test directory huntrleaks -- run multiple times to test for leaks; requires a debug build; a triple corresponding to -R's three arguments Return: -2 test skipped because resource denied -1 test skipped for some other reason 0 test failed 1 test passed """ try: return runtest_inner(test, generate, verbose, quiet, testdir, huntrleaks) finally: cleanup_test_droppings(test, verbose) | 1faed11e7c8a032086c03a6248a1cab9f2725938 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/8125/1faed11e7c8a032086c03a6248a1cab9f2725938/regrtest.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
3813,
12,
3813,
16,
2103,
16,
3988,
16,
10902,
16,
1842,
1214,
33,
7036,
16,
366,
318,
313,
298,
581,
87,
33,
8381,
16,
1198,
33,
8381,
4672,
3536,
1997,
279,
2202,
1842,
18,
2... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
3813,
12,
3813,
16,
2103,
16,
3988,
16,
10902,
16,
1842,
1214,
33,
7036,
16,
366,
318,
313,
298,
581,
87,
33,
8381,
16,
1198,
33,
8381,
4672,
3536,
1997,
279,
2202,
1842,
18,
2... |
"confirmed_forecasts_only" : fields.boolean("Validated Forecasts", help = "Check to take validated forecasts only. " \ | "confirmed_forecasts_only": fields.boolean("Validated Forecasts", help = "Check to take validated forecasts only. " \ | coeff = coeff_uom2def * coeff_def2uom | 6a21b5d661b74602c870fac41f3c8c42e8bc6c33 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/7397/6a21b5d661b74602c870fac41f3c8c42e8bc6c33/stock_planning.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
11943,
273,
11943,
67,
89,
362,
22,
536,
380,
11943,
67,
536,
22,
89,
362,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
11943,
273,
11943,
67,
89,
362,
22,
536,
380,
11943,
67,
536,
22,
89,
362,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
namecolumn = gtk.TreeViewColumn( _('Podcast')) | iconcolumn = gtk.TreeViewColumn('') | def new(self): if gpodder.interface == gpodder.MAEMO: # Maemo-specific changes to the UI global scalable_dir scalable_dir = scalable_dir.replace('.svg', '.png') self.app = hildon.Program() gtk.set_application_name('gPodder') self.window = hildon.Window() self.window.connect('delete-event', self.on_gPodder_delete_event) self.window.connect('window-state-event', self.window_state_event) self.itemUpdateChannel.show() self.UpdateChannelSeparator.show() # Give toolbar to the hildon window self.toolbar.parent.remove(self.toolbar) self.toolbar.set_style(gtk.TOOLBAR_ICONS) self.window.add_toolbar(self.toolbar) self.app.add_window(self.window) self.vMain.reparent(self.window) self.gPodder = self.window # Reparent the main menu menu = gtk.Menu() for child in self.mainMenu.get_children(): child.reparent(menu) self.itemQuit.reparent(menu) self.window.set_menu(menu) self.mainMenu.destroy() self.window.show() # do some widget hiding self.toolbar.remove(self.toolTransfer) self.itemTransferSelected.hide_all() self.item_email_subscriptions.hide_all() | e14200a9ba1c18883fbd91a008db394fec8f52c6 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/12778/e14200a9ba1c18883fbd91a008db394fec8f52c6/gui.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
394,
12,
2890,
4672,
309,
4178,
369,
765,
18,
5831,
422,
4178,
369,
765,
18,
5535,
3375,
51,
30,
468,
490,
69,
351,
83,
17,
12524,
3478,
358,
326,
6484,
2552,
8292,
429,
67,
1214,
82... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
394,
12,
2890,
4672,
309,
4178,
369,
765,
18,
5831,
422,
4178,
369,
765,
18,
5535,
3375,
51,
30,
468,
490,
69,
351,
83,
17,
12524,
3478,
358,
326,
6484,
2552,
8292,
429,
67,
1214,
82... |
for e in self._backend.iterator_out_edges(vertices, labels): yield e for e in self._backend.iterator_in_edges(vertices, labels): yield e | from itertools import chain return chain(self._backend.iterator_out_edges(vertices, labels), self._backend.iterator_in_edges(vertices, labels)) | def edge_iterator(self, vertices=None, labels=True, ignore_direction=False): """ Returns an iterator over the edges incident with any vertex given. If the graph is directed, iterates over edges going out only. If vertices is None, then returns an iterator over all edges. If self is directed, returns outgoing edges only. INPUT: - ``labels`` - if False, each edge is a tuple (u,v) of vertices. - ``ignore_direction`` - (default False) only applies to directed graphs. If True, searches across edges in either direction. EXAMPLES:: sage: for i in graphs.PetersenGraph().edge_iterator([0]): ... print i (0, 1, None) (0, 4, None) (0, 5, None) sage: D = DiGraph( { 0 : [1,2], 1: [0] } ) sage: for i in D.edge_iterator([0]): ... print i (0, 1, None) (0, 2, None) :: sage: G = graphs.TetrahedralGraph() sage: list(G.edge_iterator(labels=False)) [(0, 1), (0, 2), (0, 3), (1, 2), (1, 3), (2, 3)] :: sage: D = DiGraph({1:[0], 2:[0]}) sage: list(D.edge_iterator(0)) [] sage: list(D.edge_iterator(0, ignore_direction=True)) [(1, 0, None), (2, 0, None)] """ if vertices is None: vertices = self elif vertices in self: vertices = [vertices] else: vertices = [v for v in vertices if v in self] if ignore_direction and self._directed: for e in self._backend.iterator_out_edges(vertices, labels): yield e for e in self._backend.iterator_in_edges(vertices, labels): yield e elif self._directed: for e in self._backend.iterator_out_edges(vertices, labels): yield e else: for e in self._backend.iterator_edges(vertices, labels): yield e | 808b5a15ac3fb6dc169df27472da46433ab17eae /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9890/808b5a15ac3fb6dc169df27472da46433ab17eae/generic_graph.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3591,
67,
9838,
12,
2890,
16,
6928,
33,
7036,
16,
3249,
33,
5510,
16,
2305,
67,
9855,
33,
8381,
4672,
3536,
2860,
392,
2775,
1879,
326,
5231,
22933,
598,
1281,
5253,
864,
18,
971,
326,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3591,
67,
9838,
12,
2890,
16,
6928,
33,
7036,
16,
3249,
33,
5510,
16,
2305,
67,
9855,
33,
8381,
4672,
3536,
2860,
392,
2775,
1879,
326,
5231,
22933,
598,
1281,
5253,
864,
18,
971,
326,... |
if userhome.endswith('/'): i += 1 | userhome = userhome.rstrip('/') | def expanduser(path): """Expand ~ and ~user constructions. If user or $HOME is unknown, do nothing.""" if not path.startswith('~'): return path i = path.find('/', 1) if i < 0: i = len(path) if i == 1: if 'HOME' not in os.environ: import pwd userhome = pwd.getpwuid(os.getuid()).pw_dir else: userhome = os.environ['HOME'] else: import pwd try: pwent = pwd.getpwnam(path[1:i]) except KeyError: return path userhome = pwent.pw_dir if userhome.endswith('/'): i += 1 return userhome + path[i:] | 9c9a9ab6340eb2225f57fbef6d6cf9e3da547142 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8546/9c9a9ab6340eb2225f57fbef6d6cf9e3da547142/posixpath.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
14717,
12,
803,
4672,
3536,
12271,
4871,
471,
4871,
1355,
16171,
87,
18,
225,
971,
729,
578,
271,
14209,
353,
5917,
16,
741,
5083,
12123,
309,
486,
589,
18,
17514,
1918,
2668,
20206,
467... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
14717,
12,
803,
4672,
3536,
12271,
4871,
471,
4871,
1355,
16171,
87,
18,
225,
971,
729,
578,
271,
14209,
353,
5917,
16,
741,
5083,
12123,
309,
486,
589,
18,
17514,
1918,
2668,
20206,
467... |
nodePath = loader.loadModel(Filename(mopathFilename)) | nodePath = loader.loadModel( Filename.fromOsSpecific(mopathFilename)) | def loadCurveFromFile(self): # Use first directory in model path mPath = getModelPath() if mPath.getNumDirectories() > 0: if `mPath.getDirectory(0)` == '.': path = '.' else: path = mPath.getDirectory(0).toOsSpecific() else: path = '.' if not os.path.isdir(path): print 'MopathRecorder Info: Empty Model Path!' print 'Using current directory' path = '.' mopathFilename = askopenfilename( defaultextension = '.egg', filetypes = (('Egg Files', '*.egg'), ('Bam Files', '*.bam'), ('All files', '*')), initialdir = path, title = 'Load Nurbs Curve', parent = self.parent) if mopathFilename: self.reset() nodePath = loader.loadModel(Filename(mopathFilename)) self.curveCollection = ParametricCurveCollection() # MRM: Add error check self.curveCollection.addCurves(nodePath.node()) nodePath.removeNode() if self.curveCollection: # Draw the curve self.nurbsCurveDrawer.setCurves(self.curveCollection) self.nurbsCurveDrawer.draw() # Save a pointset for this curve self.extractPointSetFromCurveCollection() else: self.reset() | 5d65191a898ef157f4cc26ac0a27ebf4b737e2c3 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8543/5d65191a898ef157f4cc26ac0a27ebf4b737e2c3/MopathRecorder.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1262,
9423,
13087,
12,
2890,
4672,
468,
2672,
1122,
1867,
316,
938,
589,
312,
743,
273,
7454,
743,
1435,
309,
312,
743,
18,
588,
2578,
13071,
1435,
405,
374,
30,
309,
1375,
81,
743,
18... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1262,
9423,
13087,
12,
2890,
4672,
468,
2672,
1122,
1867,
316,
938,
589,
312,
743,
273,
7454,
743,
1435,
309,
312,
743,
18,
588,
2578,
13071,
1435,
405,
374,
30,
309,
1375,
81,
743,
18... |
sage: g.dominating_set(value_only=True) | sage: g.dominating_set(value_only=True) | def dominating_set(self, independent=False, value_only=False, solver=None, verbose=0): r""" Returns a minimum dominating set of the graph represented by the list of its vertices. For more information, see the `Wikipedia article on dominating sets <http://en.wikipedia.org/wiki/Dominating_set>`_. | 08b3ef8c500132f63d2dfd7371451a85a1a6987d /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9417/08b3ef8c500132f63d2dfd7371451a85a1a6987d/generic_graph.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4092,
267,
1776,
67,
542,
12,
2890,
16,
14807,
33,
8381,
16,
460,
67,
3700,
33,
8381,
16,
12776,
33,
7036,
16,
3988,
33,
20,
4672,
436,
8395,
2860,
279,
5224,
4092,
267,
1776,
444,
4... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4092,
267,
1776,
67,
542,
12,
2890,
16,
14807,
33,
8381,
16,
460,
67,
3700,
33,
8381,
16,
12776,
33,
7036,
16,
3988,
33,
20,
4672,
436,
8395,
2860,
279,
5224,
4092,
267,
1776,
444,
4... |
cur_line.append(chunks[0][0:space_left]) chunks[0] = chunks[0][space_left:] | cur_line.append(reversed_chunks[-1][:space_left]) reversed_chunks[-1] = reversed_chunks[-1][space_left:] | def _handle_long_word(self, chunks, cur_line, cur_len, width): """_handle_long_word(chunks : [string], cur_line : [string], cur_len : int, width : int) | 8bfa8935ea2fc645e5f166f180c2ba18cf11ade3 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8546/8bfa8935ea2fc645e5f166f180c2ba18cf11ade3/textwrap.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
4110,
67,
5748,
67,
1095,
12,
2890,
16,
6055,
16,
662,
67,
1369,
16,
662,
67,
1897,
16,
1835,
4672,
3536,
67,
4110,
67,
5748,
67,
1095,
12,
14691,
294,
306,
1080,
6487,
662,
67,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
4110,
67,
5748,
67,
1095,
12,
2890,
16,
6055,
16,
662,
67,
1369,
16,
662,
67,
1897,
16,
1835,
4672,
3536,
67,
4110,
67,
5748,
67,
1095,
12,
14691,
294,
306,
1080,
6487,
662,
67,... |
stream = GzipFile(arg, 'rt') | stream = GzipFile(arg, 'rb') | def main(): parser = optparse.OptionParser( usage="\n\t%prog [options] [file] ...") parser.add_option( '-s', '--stats', action="store_true", dest="stats", default=False, help="generate statistics instead") parser.add_option( '--color', '--colour', type="choice", choices=('never', 'always', 'auto'), metavar='WHEN', dest="color", default="always", help="coloring: never, always, or auto [default: %default]") (options, args) = parser.parse_args(sys.argv[1:]) if options.color == 'always' or options.color == 'auto' and sys.stdout.isatty(): formatter = AnsiFormatter() else: formatter = Formatter() if options.stats: factory = StatsTraceParser else: factory = TraceParser if args: for arg in args: if arg.endswith('.gz'): stream = GzipFile(arg, 'rt') elif arg.endswith('.bz2'): from bz2 import BZ2File stream = BZ2File(arg, 'rt') else: stream = open(arg, 'rt') parser = factory(stream, formatter) parser.parse() else: parser = factory(sys.stdin, formatter) parser.parse() | 16dfadb3c55ab6ff7a36f905e6e6d4faa75d9e16 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/12828/16dfadb3c55ab6ff7a36f905e6e6d4faa75d9e16/xml2txt.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2774,
13332,
2082,
273,
2153,
2670,
18,
1895,
2678,
12,
4084,
1546,
64,
82,
64,
88,
9,
14654,
306,
2116,
65,
306,
768,
65,
1372,
7923,
2082,
18,
1289,
67,
3482,
12,
2400,
87,
2187,
3... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2774,
13332,
2082,
273,
2153,
2670,
18,
1895,
2678,
12,
4084,
1546,
64,
82,
64,
88,
9,
14654,
306,
2116,
65,
306,
768,
65,
1372,
7923,
2082,
18,
1289,
67,
3482,
12,
2400,
87,
2187,
3... |
axisLineActor = self._tdObjectsDict[tdObject]['axisLineActor'] axisPoints = self._tdObjectsDict[tdObject]['axisPoints'] | axisLineActor = objectDict['axisLineActor'] axisPoints = objectDict['axisPoints'] | def _axisToSlice(self, tdObject, sliceDirection): """If tdObject has an axis, make the axis lie in the plane defined by sliceDirection. """ | fc0788b84d2d17e23cfbf2eba51bf9dd35371f2d /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/4494/fc0788b84d2d17e23cfbf2eba51bf9dd35371f2d/tdObjects.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
4890,
774,
5959,
12,
2890,
16,
6655,
921,
16,
2788,
8212,
4672,
3536,
2047,
6655,
921,
711,
392,
2654,
16,
1221,
326,
2654,
328,
1385,
316,
326,
11017,
2553,
635,
2788,
8212,
18,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
4890,
774,
5959,
12,
2890,
16,
6655,
921,
16,
2788,
8212,
4672,
3536,
2047,
6655,
921,
711,
392,
2654,
16,
1221,
326,
2654,
328,
1385,
316,
326,
11017,
2553,
635,
2788,
8212,
18,
... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.