rem stringlengths 1 226k | add stringlengths 0 227k | context stringlengths 6 326k | meta stringlengths 143 403 | input_ids listlengths 256 256 | attention_mask listlengths 256 256 | labels listlengths 128 128 |
|---|---|---|---|---|---|---|
i += 1 | i += 1 | def _get_categories(self, products,form): cat_ids=[] res=[] self.pricelist = form['price_list'] self._set_quantity(form) pool = pooler.get_pool(self.cr.dbname) pro_ids=[] for product in products: pro_ids.append(product.id) if product.categ_id.id not in cat_ids: cat_ids.append(product.categ_id.id) cats = pool.get('product.category').read(self.cr, self.uid, cat_ids, ['name']) for cat in cats: product_ids=pool.get('product.product').search(self.cr, self.uid, [('id','in',pro_ids),('categ_id','=',cat['id'])]) products = [] for product in pool.get('product.product').read(self.cr, self.uid, product_ids, ['name','code']): val={ 'id':product['id'], 'name':product['name'], 'code':product['code'] } i = 1 for qty in self.quantity: if qty == 0: val['qty'+str(i)] = "" else: val['qty'+str(i)]=self._get_price(self.pricelist, product['id'], qty) i += 1 products.append(val) res.append({'name':cat['name'],'products':products}) return res | aa7b817b4222f234e8e037c630dd89e3baba3594 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/7397/aa7b817b4222f234e8e037c630dd89e3baba3594/product_pricelist.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
588,
67,
8995,
12,
2890,
16,
10406,
16,
687,
4672,
6573,
67,
2232,
33,
8526,
400,
33,
8526,
365,
18,
683,
335,
5449,
273,
646,
3292,
8694,
67,
1098,
3546,
365,
6315,
542,
67,
16... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
588,
67,
8995,
12,
2890,
16,
10406,
16,
687,
4672,
6573,
67,
2232,
33,
8526,
400,
33,
8526,
365,
18,
683,
335,
5449,
273,
646,
3292,
8694,
67,
1098,
3546,
365,
6315,
542,
67,
16... |
scanhdlr.set_exit_node("$"+current_exit.idhex) | scanhdlr.set_exit_node("$"+current_exit_idhex) | def main(argv): # make sure we have something to test for if len(argv) < 2: print '' print 'Please provide at least one test option:' print '--pernode <n>' print '--resume [<n>]' print '--rescan [<n>]' print '--ssl' print '--http' print '--html' | 31ef876320f86b9a4a86e18c29a48abb1b3d4aac /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/3762/31ef876320f86b9a4a86e18c29a48abb1b3d4aac/soat.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2774,
12,
19485,
4672,
468,
1221,
3071,
732,
1240,
5943,
358,
1842,
364,
309,
562,
12,
19485,
13,
411,
576,
30,
1172,
875,
1172,
296,
8496,
5615,
622,
4520,
1245,
1842,
1456,
2497,
1172,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2774,
12,
19485,
4672,
468,
1221,
3071,
732,
1240,
5943,
358,
1842,
364,
309,
562,
12,
19485,
13,
411,
576,
30,
1172,
875,
1172,
296,
8496,
5615,
622,
4520,
1245,
1842,
1456,
2497,
1172,... |
self.assertEqual(dir(None), dir(Ellipsis)) | self.assertEqual(dir(NotImplemented), dir(Ellipsis)) if test_support.check_impl_detail(): self.assertEqual(dir(None), dir(Ellipsis)) | def getdict(self): return "Not a dict!" | 4a0d8c36fabcd955e24a465c9276c005529fa702 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/8125/4a0d8c36fabcd955e24a465c9276c005529fa702/test_descr.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
1576,
12,
2890,
4672,
327,
315,
1248,
279,
2065,
4442,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
1576,
12,
2890,
4672,
327,
315,
1248,
279,
2065,
4442,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,... |
year_start = period_break(dates,'year') | year_start = period_break(dates_,'year') | def first_label(label_flags): if label_flags[0] == 0: return label_flags[1] else: return label_flags[0] | aabea5e0836589442e382fbe7be9145dfcf8a139 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/12971/aabea5e0836589442e382fbe7be9145dfcf8a139/plotlib.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1122,
67,
1925,
12,
1925,
67,
7133,
4672,
309,
1433,
67,
7133,
63,
20,
65,
422,
374,
30,
327,
1433,
67,
7133,
63,
21,
65,
469,
30,
327,
1433,
67,
7133,
63,
20,
65,
2,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1122,
67,
1925,
12,
1925,
67,
7133,
4672,
309,
1433,
67,
7133,
63,
20,
65,
422,
374,
30,
327,
1433,
67,
7133,
63,
21,
65,
469,
30,
327,
1433,
67,
7133,
63,
20,
65,
2,
-100,
-100,
... |
sendCodec = Codec(self._broker.conn.spec) | sendCodec = Codec() | def _sendMethodRequest(self, name, args, kwargs, synchronous=False, timeWait=None): for method in self._schema.getMethods(): if name == method.name: aIdx = 0 sendCodec = Codec(self._broker.conn.spec) seq = self._session.seqMgr._reserve((method, synchronous)) self._broker._setHeader(sendCodec, 'M', seq) self._objectId.encode(sendCodec) self._schema.getKey().encode(sendCodec) sendCodec.write_str8(name) | 72fcb9ecf5a97a16c1b91f83696f84b81cf029f5 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/197/72fcb9ecf5a97a16c1b91f83696f84b81cf029f5/console.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
4661,
1305,
691,
12,
2890,
16,
508,
16,
833,
16,
1205,
16,
12209,
33,
8381,
16,
813,
5480,
33,
7036,
4672,
364,
707,
316,
365,
6315,
4821,
18,
588,
4712,
13332,
309,
508,
422,
7... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
4661,
1305,
691,
12,
2890,
16,
508,
16,
833,
16,
1205,
16,
12209,
33,
8381,
16,
813,
5480,
33,
7036,
4672,
364,
707,
316,
365,
6315,
4821,
18,
588,
4712,
13332,
309,
508,
422,
7... |
if applied == []: self.__set_current(None) else: self.__set_current(applied[-1]) | def pop_patch(self, name, keep = False): """Pops the top patch from the stack """ applied = self.get_applied() applied.reverse() assert(name in applied) | 532cdf94f80be6fb95682d9673c12a1d1aecf9a0 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/12366/532cdf94f80be6fb95682d9673c12a1d1aecf9a0/stack.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1843,
67,
2272,
12,
2890,
16,
508,
16,
3455,
273,
1083,
4672,
3536,
52,
4473,
326,
1760,
4729,
628,
326,
2110,
3536,
6754,
273,
365,
18,
588,
67,
438,
3110,
1435,
6754,
18,
9845,
1435,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1843,
67,
2272,
12,
2890,
16,
508,
16,
3455,
273,
1083,
4672,
3536,
52,
4473,
326,
1760,
4729,
628,
326,
2110,
3536,
6754,
273,
365,
18,
588,
67,
438,
3110,
1435,
6754,
18,
9845,
1435,... | |
encoding='utf-8', mimeType='text/plain', compression='bz2'): | encoding='utf-8', mimetype='text/plain', compression='bz2'): | def makeValue(self, data, encoding='utf-8', mimeType='text/plain', compression='bz2'): | e7a849e8a8ae37385f9a390a1a584eaa150b512f /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9228/e7a849e8a8ae37385f9a390a1a584eaa150b512f/Types.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1221,
620,
12,
2890,
16,
501,
16,
2688,
2218,
3158,
17,
28,
2187,
10707,
2218,
955,
19,
7446,
2187,
9154,
2218,
25292,
22,
11,
4672,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1221,
620,
12,
2890,
16,
501,
16,
2688,
2218,
3158,
17,
28,
2187,
10707,
2218,
955,
19,
7446,
2187,
9154,
2218,
25292,
22,
11,
4672,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
doc = Request(self, 'user.getLovedTracks', self.api_key, params).execute() | doc = _Request(self, 'user.getLovedTracks', self.api_key, params).execute() | def getLovedTracks(self): """Returns the last 50 tracks loved by this user. """ params = self._getParams() doc = Request(self, 'user.getLovedTracks', self.api_key, params).execute() if not doc: return None list = [] for track in doc.getElementsByTagName('track'): title = self._extract(track, 'name', 0) artist = self._extract(track, 'name', 1) list.append(Track(artist, title, *self.auth_data)) return list | 903c9b1622fe56617e5099f20abdc69f2090f8e0 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9926/903c9b1622fe56617e5099f20abdc69f2090f8e0/pylast.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
1504,
2155,
22138,
12,
2890,
4672,
3536,
1356,
326,
1142,
6437,
13933,
437,
2155,
635,
333,
729,
18,
3536,
225,
859,
273,
365,
6315,
588,
1370,
1435,
997,
273,
389,
691,
12,
2890,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
1504,
2155,
22138,
12,
2890,
4672,
3536,
1356,
326,
1142,
6437,
13933,
437,
2155,
635,
333,
729,
18,
3536,
225,
859,
273,
365,
6315,
588,
1370,
1435,
997,
273,
389,
691,
12,
2890,
... |
else: return | else: return | def open(self, version='', transaction=None, temporary=0, force=None, waitflag=1, mvcc=True, txn_mgr=None, synch=True): """Return a database Connection for use by application code. | 3f5dc8adcf31c234935a3c71325d44bf2f319df0 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/10048/3f5dc8adcf31c234935a3c71325d44bf2f319df0/DB.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1696,
12,
2890,
16,
1177,
2218,
2187,
2492,
33,
7036,
16,
6269,
33,
20,
16,
2944,
33,
7036,
16,
2529,
6420,
33,
21,
16,
7701,
952,
33,
5510,
16,
7827,
67,
13552,
33,
7036,
16,
272,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1696,
12,
2890,
16,
1177,
2218,
2187,
2492,
33,
7036,
16,
6269,
33,
20,
16,
2944,
33,
7036,
16,
2529,
6420,
33,
21,
16,
7701,
952,
33,
5510,
16,
7827,
67,
13552,
33,
7036,
16,
272,
... |
class SubversionTests(unittest.TestCase): | class SubversionTests(DjangoTestCase): | def testBadRoot(self): """Testing a bad CVSROOT""" file = 'test/testfile' rev = Revision('1.1') badrepo = Repository(name='CVS', path=self.cvs_repo_path + '2', tool=Tool.objects.get(name='CVS')) badtool = badrepo.get_scmtool() | 464956886f1e828de795677bd4bb7f27108f35f5 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/1600/464956886f1e828de795677bd4bb7f27108f35f5/tests.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
6434,
2375,
12,
2890,
4672,
3536,
22218,
279,
5570,
385,
14640,
9185,
8395,
585,
273,
296,
3813,
19,
3813,
768,
11,
5588,
273,
13132,
2668,
21,
18,
21,
6134,
5570,
7422,
273,
6281,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
6434,
2375,
12,
2890,
4672,
3536,
22218,
279,
5570,
385,
14640,
9185,
8395,
585,
273,
296,
3813,
19,
3813,
768,
11,
5588,
273,
13132,
2668,
21,
18,
21,
6134,
5570,
7422,
273,
6281,... |
count = self.SendMessage(win32defines.TTM_GETTOOLCOUNT, 0, 0) | count = self.SendMessage(win32defines.TTM_GETTOOLCOUNT) | def _PlayWithToolTipControls(self): "Just playing for now!" # get the number of tooltips associated with the control count = self.SendMessage(win32defines.TTM_GETTOOLCOUNT, 0, 0) | d79966405191de9504cf064467476aa5b42db4be /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/6953/d79966405191de9504cf064467476aa5b42db4be/common_controls.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
11765,
1190,
29250,
16795,
12,
2890,
4672,
315,
19642,
23982,
364,
2037,
4442,
468,
336,
326,
1300,
434,
358,
4397,
7146,
3627,
598,
326,
3325,
1056,
273,
365,
18,
3826,
1079,
12,
8... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
11765,
1190,
29250,
16795,
12,
2890,
4672,
315,
19642,
23982,
364,
2037,
4442,
468,
336,
326,
1300,
434,
358,
4397,
7146,
3627,
598,
326,
3325,
1056,
273,
365,
18,
3826,
1079,
12,
8... |
u = urllib2.urlopen("ftp://ftp.mirror.nl/pub/mirror/gnu/", timeout=None) self.assertEqual(u.fp.fp._sock.gettimeout(), 60) | u = urllib2.urlopen("ftp://ftp.mirror.nl/pub/mirror/gnu/", timeout=None) self.assertEqual(u.fp.fp.raw._sock.gettimeout(), 60) | def test_ftp_NoneWithdefault(self): prev = socket.getdefaulttimeout() socket.setdefaulttimeout(60) try: u = urllib2.urlopen("ftp://ftp.mirror.nl/pub/mirror/gnu/", timeout=None) self.assertEqual(u.fp.fp._sock.gettimeout(), 60) finally: socket.setdefaulttimeout(prev) | 51b2a8d774e7420ac15a785d34fb4344035675fd /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/8125/51b2a8d774e7420ac15a785d34fb4344035675fd/test_urllib2net.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
11727,
67,
7036,
1190,
1886,
12,
2890,
4672,
2807,
273,
2987,
18,
588,
1886,
4538,
1435,
2987,
18,
542,
1886,
4538,
12,
4848,
13,
775,
30,
582,
273,
11527,
22,
18,
295,
18589... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
11727,
67,
7036,
1190,
1886,
12,
2890,
4672,
2807,
273,
2987,
18,
588,
1886,
4538,
1435,
2987,
18,
542,
1886,
4538,
12,
4848,
13,
775,
30,
582,
273,
11527,
22,
18,
295,
18589... |
cur=self.conn.cursor() cur.execute('select max(rowid) from %s' % name) return int(cur.fetchone()[0]) | return self.conn.db.sqlite_last_insert_rowid() | def getAutoIncrement(self, name): cur=self.conn.cursor() cur.execute('select max(rowid) from %s' % name) return int(cur.fetchone()[0]) | 691b93bb7906a934fb49f635dbc273f56a1fb8fb /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2304/691b93bb7906a934fb49f635dbc273f56a1fb8fb/sqliteconn.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
26707,
10798,
12,
2890,
16,
508,
4672,
327,
365,
18,
4646,
18,
1966,
18,
19460,
67,
2722,
67,
6387,
67,
492,
350,
1435,
225,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
26707,
10798,
12,
2890,
16,
508,
4672,
327,
365,
18,
4646,
18,
1966,
18,
19460,
67,
2722,
67,
6387,
67,
492,
350,
1435,
225,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-1... |
code = re.sub(r'U\+([0-9A-F]{2})([0-9A-F]{2}\b)',r'\1\t\2',items[0]).split('\t') | code = r1.sub(r'\1\t\2',items[0]).split('\t') | def process_readings(self, source, fout): oucode = 0 | 9145b8e82b3e4a975e637a96d80862a186515a18 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/10700/9145b8e82b3e4a975e637a96d80862a186515a18/unihan_conv.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1207,
67,
896,
899,
12,
2890,
16,
1084,
16,
17382,
4672,
16694,
710,
273,
374,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1207,
67,
896,
899,
12,
2890,
16,
1084,
16,
17382,
4672,
16694,
710,
273,
374,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-1... |
django.core.management.syncdb (verbosity = 1, interactive = False) | try: django.core.management.syncdb (verbosity = 1, interactive = False) except TypeError: django.core.management.syncdb () | def syncdb (db): from django.conf import settings (old_ENGINE, old_NAME, old_USER, old_PASSWORD, old_HOST, old_PORT) = (settings.DATABASE_ENGINE, settings.DATABASE_NAME, settings.DATABASE_USER, settings.DATABASE_PASSWORD, settings.DATABASE_HOST, settings.DATABASE_PORT) settings.DATABASE_ENGINE = 'postgresql' settings.DATABASE_NAME = 'postgres' settings.DATABASE_USER = pgembed._postgres_user settings.DATABASE_PASSWORD = '' settings.DATABASE_HOST = db settings.DATABASE_PORT = '5432' import django.core.management django.core.management.syncdb (verbosity = 1, interactive = False) (settings.DATABASE_ENGINE, settings.DATABASE_NAME, settings.DATABASE_USER, settings.DATABASE_PASSWORD, settings.DATABASE_HOST, settings.DATABASE_PORT) = (old_ENGINE, old_NAME, old_USER, old_PASSWORD, old_HOST, old_PORT) | fb7ea3774dc6679ccd05f2af1884e825373a2afc /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/7253/fb7ea3774dc6679ccd05f2af1884e825373a2afc/djschemadiff.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3792,
1966,
261,
1966,
4672,
628,
13710,
18,
3923,
1930,
1947,
261,
1673,
67,
28980,
16,
1592,
67,
1985,
16,
1592,
67,
4714,
16,
1592,
67,
13784,
16,
1592,
67,
8908,
16,
1592,
67,
6354... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3792,
1966,
261,
1966,
4672,
628,
13710,
18,
3923,
1930,
1947,
261,
1673,
67,
28980,
16,
1592,
67,
1985,
16,
1592,
67,
4714,
16,
1592,
67,
13784,
16,
1592,
67,
8908,
16,
1592,
67,
6354... |
included = os.path.basename(fname)[:-2] + '_INCLUDED' pattern = ' if re.search(pattern, contents) == None: | if not found: | """All .h files must have #ifdef <file>_INCLUDED protection.""" | bce6219e2ec841edbfb80ef3b8a94d78c446a2c2 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/675/bce6219e2ec841edbfb80ef3b8a94d78c446a2c2/check_source.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
3536,
1595,
263,
76,
1390,
1297,
1240,
468,
430,
536,
411,
768,
34,
67,
706,
11686,
7660,
17862,
12123,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
3536,
1595,
263,
76,
1390,
1297,
1240,
468,
430,
536,
411,
768,
34,
67,
706,
11686,
7660,
17862,
12123,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-... |
PYTHONINCSEARCHPATH = [ distutils.sysconfig.get_python_inc() ] PYTHONLIBSEARCHPATH = [ distutils.sysconfig.get_config_vars()["DESTSHARED"] + os.path.sep + "libpython*" ] | PYTHONINCSEARCHPATH = [ distutils.sysconfig.get_python_inc() + os.path.sep + "Python.h" ] if distutils.sysconfig.get_config_vars().has_key("DESTSHARED"): PYTHONLIBSEARCHPATH = [ distutils.sysconfig.get_config_vars()["DESTSHARED"] + os.path.sep + "libpython*" ] else: PYTHONLIBSEARCHPATH = [] | def checkPython(options): PYTHONINCSEARCHPATH = [ distutils.sysconfig.get_python_inc() ] PYTHONLIBSEARCHPATH = [ distutils.sysconfig.get_config_vars()["DESTSHARED"] + os.path.sep + "libpython*" ] PYTHONLIBSTATICSEARCHPATH = [] if sys.platform == "win32": PYTHONLIBSEARCHPATH += [ sys.prefix + "\Libs\python*.lib" ] PYTHONINCSEARCHPATH += [ sys.prefix + "\include\Python.h" ] elif sys.platform == "linux2": PYTHONLIBSEARCHPATH += [ "/usr/local/lib/libpython2.3*.so", \ "/usr/local/lib/[Pp]ython*/libpython2.3*.so", \ "/usr/lib/libpython2.3*.so", \ "/usr/lib/[Pp]ython*/libpython2.3*.so", \ "/usr/lib/[Pp]ython*/config/libpython2.3*.so", \ "/usr/local/lib/[Pp]ython*/config/libpython2.3*.so"] PYTHONLIBSTATICSEARCHPATH += [ "/usr/local/lib/libpython2.3*.a", \ "/usr/local/lib/[Pp]ython2.3*/libpython2.3*.a", \ "/usr/lib/libpython2.3*.a", \ "/usr/lib/[Pp]ython2.3*/libpython2.3*.a", \ "/usr/lib/[Pp]ython2.3*/config/libpython2.3*.a", \ "/usr/local/lib/[Pp]ython2.3*/config/libpython2.3*.a"] PYTHONINCSEARCHPATH += [ "/usr/local/include/[Pp]ython2.3*/Python.h", \ "/usr/include/[Pp]ython2.3*/Python.h"] elif sys.platform == "freebsd4" or sys.platform == "freebsd5": PYTHONLIBSEARCHPATH += [ "/usr/local/lib/libpython2.3*.so", \ "/usr/local/lib/[Pp]ython2.3*/libpython2.3*.so", \ "/usr/lib/libpython2.3*.so", \ "/usr/lib/[Pp]ython2.3*/libpython2.3*.so", \ "/usr/lib/[Pp]ython2.3*/config/libpython2.3*.so", \ "/usr/local/lib/[Pp]ython2.3*/config/libpython2.3*.so"] PYTHONLIBSTATICSEARCHPATH += [ "/usr/local/lib/libpython2.3*.a", \ "/usr/local/lib/[Pp]ython2.3*/libpython2.3*.a", \ "/usr/lib/libpython2.3*.a", \ "/usr/lib/[Pp]ython2.3*/libpython2.3*.a", \ "/usr/lib/[Pp]ython2.3*/config/libpython2.3*.a", \ "/usr/local/lib/[Pp]ython2.3*/config/libpython2.3*.a"] PYTHONINCSEARCHPATH += [ "/usr/local/include/[Pp]ython2.3*/Python.h", \ "/usr/include/[Pp]ython2.3*/Python.h"] elif sys.platform == "darwin": PYTHONINCSEARCHPATH += [ "/System/Library/Frameworks/Python.framework/Versions/Current/Headers/Python.h" ] PYTHONLIBSEARCHPATH += [ ] PYTHONLIBSTATICSEARCHPATH += ["/System/Library/Frameworks/Python.framework/Versions/Current/Python", \ "/System/Library/Frameworks/Python.framework/Versions/Current/lib/[Pp]ython*/config/libpython*.a", \ "/usr/local/lib/[Pp]ython*/config/libpython*.a"] else: sys.stdout.write(red("ERROR")+": Unknown platform %s to checkPython()\n" % sys.platform ) sys.exit() # if --static if options.staticlink: PYTHONLIBSEARCHPATH = PYTHONLIBSTATICSEARCHPATH # if it was overiden... if options.py_incpath: PYTHONINCSEARCHPATH = [ options.py_incpath ] if options.py_libpath: PYTHONLIBSEARCHPATH = [ options.py_libpath ] sys.stdout.write( "Checking Python version... " ) if sys.hexversion >= 0x020300F0: sys.stdout.write(green("ok\n")) else: sys.stdout.write( red("failed") + "\n" ) sys.stdout.write( bold("Wolfpack requires Python version greater than 2.3.0 ") ) sys.exit(); sys.stdout.write( "Checking CPU byte order... %s\n" % sys.byteorder ) if sys.byteorder != 'little': sys.stdout.write(yellow("Warning:") + " Wolfpack support for big endian systems is completely experimental and unlikey to work\n" ) sys.stdout.write( "Searching for Python library... " ) global py_libpath global py_libfile py_libfile, py_libpath = findFile( PYTHONLIBSEARCHPATH ) if ( py_libfile ): sys.stdout.write("%s\n" % os.path.join( py_libpath, py_libfile ) ) else: sys.stdout.write(red("Not Found!") + "\n") sys.exit() global py_incpath py_incfile = None sys.stdout.write( "Searching for Python includes... " ) py_incfile, py_incpath = findFile( PYTHONINCSEARCHPATH ) if ( py_incfile ): sys.stdout.write( "%s\n" % py_incpath ) else: sys.stdout.write(red("Not Found!") + "\n") sys.exit() return True | 6106b53503aa0c7e3a80d7e5eabafcff4a569afd /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2534/6106b53503aa0c7e3a80d7e5eabafcff4a569afd/configure.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
866,
15774,
12,
2116,
4672,
12191,
20131,
23213,
17251,
4211,
273,
306,
2411,
5471,
18,
9499,
1425,
18,
588,
67,
8103,
67,
9523,
1435,
397,
1140,
18,
803,
18,
10814,
397,
315,
15774,
18,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
866,
15774,
12,
2116,
4672,
12191,
20131,
23213,
17251,
4211,
273,
306,
2411,
5471,
18,
9499,
1425,
18,
588,
67,
8103,
67,
9523,
1435,
397,
1140,
18,
803,
18,
10814,
397,
315,
15774,
18,... |
op1.int.reverse() op2.int.reverse() | def __add__(self, other, context=None): """Returns self + other. | 55342e783760513207a56d5a31fa49e579a96cd3 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3187/55342e783760513207a56d5a31fa49e579a96cd3/decimal.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
1289,
972,
12,
2890,
16,
1308,
16,
819,
33,
7036,
4672,
3536,
1356,
365,
397,
1308,
18,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
1289,
972,
12,
2890,
16,
1308,
16,
819,
33,
7036,
4672,
3536,
1356,
365,
397,
1308,
18,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,... | |
retcode = 0 | retcode = subprocess.call(["svn", "update", tmpdir]) | def scrape(): new = True try: os.mkdir(tmpdir) except OSError, what: if what.errno != errno.EEXIST: raise new = False if os.stat(tmpdir).st_uid != os.getuid(): raise MetasploitScrapingError("suspicious permissions for " + tmpdir) if new: retcode = subprocess.call(["svn", "co", exploits_svn_url, tmpdir]) else: retcode = 0 #subprocess.call(["svn", "update", tmpdir]) if retcode: raise MetasploitScrapingError("svn command returned failure") for root, dirs, files in os.walk(tmpdir): for fn in files: if not fn.endswith('.rb'): continue zdict = findcve(os.path.join(root, fn)) if zdict: #print 'hit', root, fn, zdict yield zdict | 8ce56d42c506b2bf394c584ac50dae84c31f1bb5 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/888/8ce56d42c506b2bf394c584ac50dae84c31f1bb5/metasploitsvn.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
888,
25360,
13332,
394,
273,
1053,
775,
30,
1140,
18,
26686,
12,
5645,
1214,
13,
1335,
10002,
16,
4121,
30,
309,
4121,
18,
19088,
480,
8402,
18,
41,
11838,
30,
1002,
394,
273,
1083,
30... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
888,
25360,
13332,
394,
273,
1053,
775,
30,
1140,
18,
26686,
12,
5645,
1214,
13,
1335,
10002,
16,
4121,
30,
309,
4121,
18,
19088,
480,
8402,
18,
41,
11838,
30,
1002,
394,
273,
1083,
30... |
raise MunkiDownloadError("No installer_item_location in item info.") | raise MunkiDownloadError("No %s in item info." % download_item_key) | def download_installeritem(item_pl): """Downloads a installer item. Raises an error if there are issues...""" location = item_pl.get('installer_item_location') if not location: raise MunkiDownloadError("No installer_item_location in item info.") ManagedInstallDir = munkicommon.pref('ManagedInstallDir') downloadbaseurl = munkicommon.pref('PackageURL') or \ munkicommon.pref('SoftwareRepoURL') + '/pkgs/' if not downloadbaseurl.endswith('/'): downloadbaseurl = downloadbaseurl + '/' munkicommon.display_debug2('Download base URL is: %s' % downloadbaseurl) mycachedir = os.path.join(ManagedInstallDir, 'Cache') # build a URL, quoting the the location to encode reserved characters pkgurl = downloadbaseurl + urllib2.quote(location) # grab last path component of location to derive package name. pkgname = os.path.basename(location) destinationpath = os.path.join(mycachedir, pkgname) munkicommon.display_detail('Downloading %s from %s' % (pkgname, location)) # bump up verboseness so we get download percentage done feedback. # this is kind of a hack... oldverbose = munkicommon.verbose munkicommon.verbose = oldverbose + 1 dl_message = 'Downloading %s...' % pkgname try: changed = getHTTPfileIfChangedAtomically(pkgurl, destinationpath, resume=True, message=dl_message) except CurlDownloadError: munkicommon.verbose = oldverbose raise # set verboseness back. munkicommon.verbose = oldverbose if changed: if not verifySoftwarePackageIntegrity(destinationpath, item_pl, 'installer_item_hash'): raise PackageVerificationError() | 7233ea384c0d8f73c40e1623556e9fc24f98637e /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6241/7233ea384c0d8f73c40e1623556e9fc24f98637e/updatecheck.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4224,
67,
20163,
1726,
12,
1726,
67,
412,
4672,
3536,
22141,
279,
16050,
761,
18,
20497,
392,
555,
309,
1915,
854,
8296,
7070,
3660,
2117,
273,
761,
67,
412,
18,
588,
2668,
20163,
67,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4224,
67,
20163,
1726,
12,
1726,
67,
412,
4672,
3536,
22141,
279,
16050,
761,
18,
20497,
392,
555,
309,
1915,
854,
8296,
7070,
3660,
2117,
273,
761,
67,
412,
18,
588,
2668,
20163,
67,
... |
if hasattr(ob, '__call__'): return 1 else: return isinstance(ob, types.ClassType) | return hasattr(ob, '__call__') or isinstance(ob, types.ClassType) | def safe_callable(ob): # Works with ExtensionClasses and Acquisition. if hasattr(ob, '__class__'): if hasattr(ob, '__call__'): return 1 else: return isinstance(ob, types.ClassType) else: return callable(ob) | 8430ae6c5d023444299a01fa4d815c7986f5ea7b /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9658/8430ae6c5d023444299a01fa4d815c7986f5ea7b/Catalog.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4183,
67,
7293,
12,
947,
4672,
468,
4147,
87,
598,
10021,
4818,
471,
12848,
22094,
18,
309,
3859,
12,
947,
16,
4940,
1106,
7250,
4672,
327,
3859,
12,
947,
16,
4940,
1991,
972,
6134,
57... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4183,
67,
7293,
12,
947,
4672,
468,
4147,
87,
598,
10021,
4818,
471,
12848,
22094,
18,
309,
3859,
12,
947,
16,
4940,
1106,
7250,
4672,
327,
3859,
12,
947,
16,
4940,
1991,
972,
6134,
57... |
print "NOTE: Ignoring %s, using %s"%(new[zh].asasciilink(),pl.asasciilink()) | print "NOTE: Ignoring %s, using %s"%(new['zh'].asasciilink(),pl.asasciilink()) | def treesearch(pl): arr = {pl:None} # First make one step based on the language itself try: n = treestep(arr, pl, abort_on_redirect = 1) except wikipedia.IsRedirectPage: print "Is redirect page" return if n == 0 and not arr[pl]: print "Mother doesn't exist" return if untranslated: if len(arr) > 1: print "Already has translations" else: if bell: sys.stdout.write('\07') newhint = raw_input("Hint:") if not newhint: return hints.append(newhint) # Then add translations if we survived. autotranslate(pl, arr, same = same) modifications = 1 while modifications: modifications = 0 for newpl in arr.keys(): if arr[newpl] is None: modifications += treestep(arr, newpl) return arr | 6ac2f9e153e8dbe360021e0a957614ba7b896fe4 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/4404/6ac2f9e153e8dbe360021e0a957614ba7b896fe4/treelang.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
9787,
3392,
991,
12,
412,
4672,
2454,
273,
288,
412,
30,
7036,
97,
468,
5783,
1221,
1245,
2235,
2511,
603,
326,
2653,
6174,
775,
30,
290,
273,
9787,
395,
881,
12,
5399,
16,
886,
16,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
9787,
3392,
991,
12,
412,
4672,
2454,
273,
288,
412,
30,
7036,
97,
468,
5783,
1221,
1245,
2235,
2511,
603,
326,
2653,
6174,
775,
30,
290,
273,
9787,
395,
881,
12,
5399,
16,
886,
16,
... |
def find_library_file (self, dirs, lib): | def find_library_file (self, dirs, lib, debug=0): if debug: try_names = [lib + "_d", lib] else: try_names = [lib] | def find_library_file (self, dirs, lib): | e9adda1548c31c130ad38e398f08e61b00a3094b /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3187/e9adda1548c31c130ad38e398f08e61b00a3094b/msvccompiler.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1104,
67,
12083,
67,
768,
261,
2890,
16,
7717,
16,
2561,
4672,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1104,
67,
12083,
67,
768,
261,
2890,
16,
7717,
16,
2561,
4672,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
for a in pkg_resources.iter_entry_points( 'paste.global_paster_command', 'create') | for a in pkg_resources.iter_entry_points( 'paste.global_paster_command', 'create') | def webbuilder_process(context, request): valid_actions = ['submit_cgwbDownload', 'submit_cgwbGenerate'] errors = [] output_dir_prefix = None params = dict(request.params) actions = [action for action in request.params if re.match('^submit_', action)] noecho = [params.pop(action) for action in actions] configuration = params.pop('configuration', None) download_path = None if not configuration: errors.append('Choose a configuration') if not configuration in root.configurations: errors.append('Invalid configurations') if not errors: for action in actions: if action in valid_actions: try: paster = get_paster(configuration) createcmd = [a for a in pkg_resources.iter_entry_points( 'paste.global_paster_command', 'create') ][0].load() noecho = [params.update({param: True}) for param in params if params[param] in [u'on', u'checkbox_enabled']] project = params.pop('project', None) if not project: project = '' project = project.strip() if not project: raise Exception('Project does not exist or is empty') if action == 'submit_cgwbDownload': output_dir_prefix = tempfile.mkdtemp() else: output_dir_prefix = os.path.join(get_generation_path(), project) if os.path.exists(output_dir_prefix) and len(os.listdir(output_dir_prefix))>0: if not get_settings().get('debug', '').lower() == 'true' : raise Exception('Directory %s exists and is not empty' % output_dir_prefix) # keep track of boolean options for descendant templates boolean_consumed_options = [] for template in paster.templates_data: cparams = params.copy() output_dir = output_dir_prefix if template['self'].output: output_dir = os.path.join(output_dir, template['self'].output) top = ['-q', '-t', template['name'], '--no-interactive', '-o', output_dir, project ] if template['aliases']: for var, alias, default in template['aliases']: # we have a value if alias in cparams: cparams[var.name] = cparams[alias] #blank value, resetting else: if var.name in cparams: del cparams[var.name] # set to false unticked options for g in template['groups']: for myo in g['options']: if myo[1] == 'boolean': boolean_consumed_options.append(myo[0].name) for option in boolean_consumed_options: if not option in cparams: cparams[option] = False # cleanup variables names = dict([(var.name,var) for var in template['template'].vars]) for iv in cparams.keys(): if not iv in names: del cparams[iv] continue else: if not cparams[iv]: if not names[iv].default: del cparams[iv] continue # be sure not to have unicode params there because paster will swallow them up top.extend( ["%s=%s" % i for i in cparams.items()] ) cmd = createcmd('create') try: ret = cmd.run(top) except Exception, e: remove_path(output_dir) raise # do post genration stuff postprocess(paster, output_dir_prefix, project, params) if action == 'submit_cgwbDownload': ts = '%s'%datetime.datetime.now() ts = ts.replace(':', '-').replace(' ', '_')[:-4] filename = '%s-%s.tar.gz' % (project, ts) file_path = os.path.join(get_generation_path(), filename) fic = tarfile.open(file_path, 'w:gz') fic.add(output_dir_prefix, './') fic.close() download_path = os.path.basename(file_path) resp = Response( body = open(file_path).read(), content_type = 'application/x-tar-gz', headerlist = [ ('Content-Disposition', 'attachment; filename="%s"' % os.path.basename(file_path) ), ('Content-Transfer-Encoding', 'binary') ], request = request, ) os.remove(file_path) return resp except NoSuchConfigurationError: errors.append( '%s/ The required configuration ' 'does not exists : %s' % ( action, configuration) ) except MinibuildNotFoundException, e: #raise errors.append( '<div class="error">' '<p>%s/ Error while reading paster variables:</p>' '<p class="pythonerror">%r</p>' '<br/><p>Are you inside a minitage ? </p>' '<br/><ul><li> If no, please untick minitage support in the minitage ' 'section (press your browser\'s back button)</p>' '<br/><li> If yes, you must set the "MT" environment variable to ' 'the root installation of minitage or use the minitage.instance.env template ' 'and source the resulting environment file before launching cgwb<br/>' '<p>You can generate one with:</p><br/>' '<pre>\n' '\t$minitage/bin/easy_install -U minitage.paste\n' '\t$minitage/bin/paster create -t minitage.instances.env cgwb-dev (or cgwb)' '</pre>' '<br/><p>And before launching cgwb, if you do not have done that previously in your running shell:</p><br/>' '<pre>\n' '\tsource $minitage/bfg/cgwb(-dev)/sys/share/minitage/minitage.env' '</pre>' '</li></ul><br/>' '' % (action, e) ) except Parser.ParseError, e: errors.append( '<div class="error">' '<p>%s/ Error while reading paster variables:</p>' '<p class="pythonerror">%r</p>' '<p class="pythonerror"><pre>%s</pre></p>' '</div>'% (action, e, e.report()) ) except Exception, e: raise errors.append( '<div class="error">' '<p>%s/ -- Error while reading paster variables:</p>' '<p class="pythonerror">%r</p>' '<p class="pythonerror">%s</p>' '</div>'% (action, e, e) ) main = get_template('templates/main_template.pt') return render_template_to_response( 'templates/process.pt', request = request, errors = errors, context=context, output=output_dir_prefix, download_path=download_path, main=main) | fe40ea8d2e545a7ce0b7a084f40e2f92abd0203c /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/13230/fe40ea8d2e545a7ce0b7a084f40e2f92abd0203c/views.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3311,
9574,
67,
2567,
12,
2472,
16,
590,
4672,
923,
67,
4905,
273,
10228,
9297,
67,
26275,
9464,
7109,
2187,
296,
9297,
67,
26275,
9464,
4625,
3546,
1334,
273,
5378,
876,
67,
1214,
67,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3311,
9574,
67,
2567,
12,
2472,
16,
590,
4672,
923,
67,
4905,
273,
10228,
9297,
67,
26275,
9464,
7109,
2187,
296,
9297,
67,
26275,
9464,
4625,
3546,
1334,
273,
5378,
876,
67,
1214,
67,
... |
This is the message given by the server in responce to the | This is the message given by the server in response to the | def quotedata(data): """Quote data for email. Double leading '.', and change Unix newline '\n', or Mac '\r' into Internet CRLF end-of-line. """ return re.sub(r'(?m)^\.', '..', re.sub(r'(?:\r\n|\n|\r(?!\n))', CRLF, data)) | 38138a69a9ce81026584c5df54a7d656104103d9 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3187/38138a69a9ce81026584c5df54a7d656104103d9/smtplib.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
9298,
396,
12,
892,
4672,
3536,
10257,
501,
364,
2699,
18,
225,
3698,
7676,
2611,
16,
471,
2549,
9480,
9472,
2337,
82,
2187,
578,
13217,
2337,
86,
11,
1368,
21352,
21791,
679,
17,
792,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
9298,
396,
12,
892,
4672,
3536,
10257,
501,
364,
2699,
18,
225,
3698,
7676,
2611,
16,
471,
2549,
9480,
9472,
2337,
82,
2187,
578,
13217,
2337,
86,
11,
1368,
21352,
21791,
679,
17,
792,
... |
struct = 3 elif self.radioButton2.GetValue(): | def GetValues(self): if self.radioButton1.GetValue(): struct = 1 elif self.radioButton3.GetValue(): struct = 3 elif self.radioButton2.GetValue(): struct = 7 name = self.IndexName.GetValue() index = eval(self.Index.GetValue()) number = eval(self.Number.GetValue()) return index, name, struct, number | f10f1674cce933505ba0bf578b3a0b475804a5a7 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/11017/f10f1674cce933505ba0bf578b3a0b475804a5a7/objdictedit.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
968,
1972,
12,
2890,
4672,
309,
365,
18,
17006,
3616,
21,
18,
967,
620,
13332,
1958,
273,
404,
1327,
365,
18,
17006,
3616,
23,
18,
967,
620,
13332,
1958,
273,
2371,
508,
273,
365,
18,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
968,
1972,
12,
2890,
4672,
309,
365,
18,
17006,
3616,
21,
18,
967,
620,
13332,
1958,
273,
404,
1327,
365,
18,
17006,
3616,
23,
18,
967,
620,
13332,
1958,
273,
2371,
508,
273,
365,
18,
... | |
instance = h24.CreatFromDOM(dom.documentElement) | instance = h24.CreateFromDOM(dom.documentElement) | def test_h24 (self): xml = '<h24></h24>' dom = minidom.parseString(xml) instance = h24.CreateFromDOM(dom.documentElement) print instance print instance.toxml() self.assertRaises(MissingContentError, h24.CreateFromDOM, dom.documentElement) for num_elt in range(0, 5): xml = '<h24>%s</h24>' % (''.join(num_elt * ['<elt/>']),) dom = minidom.parseString(xml) if 2 > num_elt: self.assertRaises(MissingContentError, h24.CreateFromDOM, dom.documentElement) elif 4 >= num_elt: instance = h24.CreatFromDOM(dom.documentElement) self.assertEqual(num_elt, len(instance.elt())) self.assertEqual(instance.toDOM().toxml(), xml) else: self.assertRaises(ExtraContentError, h24.CreateFromDOM, dom.documentElement) | 432bc85cfb05f93ce4273c510cca99b7d9fab4f9 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7171/432bc85cfb05f93ce4273c510cca99b7d9fab4f9/test-particle.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
76,
3247,
261,
2890,
4672,
2025,
273,
2368,
76,
3247,
4695,
76,
3247,
1870,
4092,
273,
1131,
350,
362,
18,
2670,
780,
12,
2902,
13,
791,
273,
366,
3247,
18,
1684,
1265,
8168,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
76,
3247,
261,
2890,
4672,
2025,
273,
2368,
76,
3247,
4695,
76,
3247,
1870,
4092,
273,
1131,
350,
362,
18,
2670,
780,
12,
2902,
13,
791,
273,
366,
3247,
18,
1684,
1265,
8168,... |
if exclude_quote: | if remove_quote: | def remove_wikicode(text, re_dotall = False, debug = False): if not text: return "" if debug: write_log(text+'\n', "copyright/wikicode.txt") text = re.sub('(?i)</?(p|u|i|b|em|div|span|font|small|big|code|tt).*?>', '', text) text = re.sub('(?i)<(/\s*)?br(\s*/)?>', '', text) text = re.sub('<!--.*?-->', '', text) text = text.replace('<', '<') text = text.replace('>', '>') # remove URL text = re.sub('(ftp|https?)://[\w/.,;:@&=%#\\\?_!~*\'|()\"+-]+', ' ', text) # remove Image tags text = reImageC.sub("", text) # replace piped wikilink text = re.sub("\[\[[^\]]*?\|(.*?)\]\]", "\\1", text) # remove unicode and polytonic template text = re.sub("(?i){{(unicode|polytonic)\|(.*?)}}", "\\1", text) if re_dotall: flags = "(?xsim)" # exclude wikitable text = re.sub('(?s){\|.*?^\|}', '', text) else: flags = "(?xim)" text = re.sub(""" %s ( <ref[^>]*?\s*/\s*> | # exclude <ref name = '' / > tags <ref.*?>.*?</ref> | # exclude <ref> notes ^[\ \t]*({\||[|!]).*?$ | # exclude wikitable </*nowiki> | # remove <nowiki> tags {{.*?}} | # remove (not nested) template <math>.*?</math> | # remove LaTeX staff [\[\]] | # remove [, ] ^[*:;]+ | # remove *, :, ; in begin of line <!-- | --> | ) """ % flags, "", text) if exclude_quote: # '' text '' # '' text ''. # '' text '' (text) # « text » # ... # italic_quoteC = re.compile("(?m)^[:*]?\s*(''.*?'')\.?\s*(\(.*?\))?\r?$") index = 0 try: import pywikiparser except ImportError: pywikiparser = False while pywikiparser: m = italic_quoteC.search(text, index) if not m: break s = pywikiparser.Parser(m.group(1)) try: xmldata = s.parse().toxml() if '<wikipage><p><i>' in xmldata and '</i></p></wikipage>' in xmldata: if xmldata.count('<i>') == 1: text = text[:m.start()] + text[m.end():] except: pass index = m.start() + 1 text = re.sub('(?m)^[:*]*\s*["][^"]+["]\.?\s*(\(.*?\))?\r?$', "", text) text = re.sub('(?m)^[:*]*\s*[«][^»]+[»]\.?\s*(\(.*?\))?\r?$', "", text) text = re.sub('(?m)^[:*]*\s*[“][^”]+[”]\.?\s*(\(.*?\))?\r?$', "", text) # remove useless spaces text = re.sub("(?m)(^[ \t]+|[ \t]+\r?$)", "", text) if debug: write_log(text+'\n', "copyright/wikicode_removed.txt") return text | e8580311da32621d1a5295ed08c368281d379fc5 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/4404/e8580311da32621d1a5295ed08c368281d379fc5/copyright.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1206,
67,
11999,
3487,
12,
955,
16,
283,
67,
9811,
454,
273,
1083,
16,
1198,
273,
1083,
4672,
309,
486,
977,
30,
327,
1408,
225,
309,
1198,
30,
1045,
67,
1330,
12,
955,
6797,
64,
82,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1206,
67,
11999,
3487,
12,
955,
16,
283,
67,
9811,
454,
273,
1083,
16,
1198,
273,
1083,
4672,
309,
486,
977,
30,
327,
1408,
225,
309,
1198,
30,
1045,
67,
1330,
12,
955,
6797,
64,
82,... |
log.msg("stun policy says yes, use STUN") | log.msg("stun policy says yes, use STUN", system='sip') | def setLocalIP(self, dest): """ Try and determine the local IP address to use. We use a ConnectedDatagramProtocol in the (faint?) hope that on a machine with multiple interfaces, we'll get the right one """ # XXX Allow over-riding from shtoom.stun import StunHook, getPolicy global _CACHED_LOCAL_IP host, port = dest getPref = self.sip.app.getPref if getPref('localip') is not None or _CACHED_LOCAL_IP is not None: ip = getPref('localip') or _CACHED_LOCAL_IP locAddress = (ip, getPref('listenport') or 5060) remAddress = ( host, port ) # Argh. Do a DNS lookup on remAddress else: # it is a hack! protocol = ConnectedDatagramProtocol() port = reactor.connectUDP(host, port, protocol) if protocol.transport: locAddress = (protocol.transport.getHost()[1], getPref('listenport') or 5060) remAddress = protocol.transport.getPeer()[1:3] port.stopListening() log.msg("discovered local address %r, remote %r"%(locAddress, remAddress)) _CACHED_LOCAL_IP = locAddress[0] else: self.compDef.errback(ValueError("couldn't connect to %s"%( host))) self.abortCall() return None pol = getPolicy() useStun = pol.checkStun(locAddress[0], remAddress[0]) if useStun is True: self._needSTUN = True log.msg("stun policy says yes, use STUN") deferred = defer.Deferred() deferred.addCallback(self.setStunnedLocalIP).addErrback(log.err) SH = StunHook(self.sip) deferred = SH.discoverStun(deferred) elif useStun is False: self._localIP, self._localPort = locAddress if not self.dialog.getCallID(): self.setCallID() self.sip.updateCallObject(self, self.getCallID()) self.setupDeferred.callback((self._localIP, self._localPort)) else: print "STUN policy %r failed for %r %r"%(pol, locAddress[0], remAddress[0]) # None. STUN stuff failed. Abort. from shtoom.exceptions import HostNotKnown d = self.compDef del self.compDef d.errback(HostNotKnown) self.setState('ABORTED') | 151c9fcc2f2d80d0ff28dab85b0bcc585be354f9 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/1265/151c9fcc2f2d80d0ff28dab85b0bcc585be354f9/sip.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
27970,
2579,
12,
2890,
16,
1570,
4672,
3536,
6161,
471,
4199,
326,
1191,
2971,
1758,
358,
999,
18,
1660,
999,
279,
6683,
828,
5139,
17049,
5752,
316,
326,
261,
507,
474,
8535,
27370,
716... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
27970,
2579,
12,
2890,
16,
1570,
4672,
3536,
6161,
471,
4199,
326,
1191,
2971,
1758,
358,
999,
18,
1660,
999,
279,
6683,
828,
5139,
17049,
5752,
316,
326,
261,
507,
474,
8535,
27370,
716... |
pass | pass | def setinputsizes(self, sizes): | 34c41b6b2a05b0c448580a7fb8f0602d4ac4c768 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8941/34c41b6b2a05b0c448580a7fb8f0602d4ac4c768/informixdb.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
444,
10029,
3128,
12,
2890,
16,
8453,
4672,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
444,
10029,
3128,
12,
2890,
16,
8453,
4672,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-10... |
1.00000000000000 | 0.999999999999998 | def sha_an_numerical(self, prec = 53, use_database=False, regproof=False): """ Return the numerical analytic order of Sha, which is a floating point number in all cases. | 0b6a7ae551afb09f819f164f6839d54ba267e0ef /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/9890/0b6a7ae551afb09f819f164f6839d54ba267e0ef/ell_rational_field.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
6056,
67,
304,
67,
2107,
10882,
12,
2890,
16,
13382,
273,
15935,
16,
999,
67,
6231,
33,
8381,
16,
960,
24207,
33,
8381,
4672,
3536,
2000,
326,
17409,
392,
7834,
335,
1353,
434,
2638,
6... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
6056,
67,
304,
67,
2107,
10882,
12,
2890,
16,
13382,
273,
15935,
16,
999,
67,
6231,
33,
8381,
16,
960,
24207,
33,
8381,
4672,
3536,
2000,
326,
17409,
392,
7834,
335,
1353,
434,
2638,
6... |
return Crystals_of_letters_type_A_element(self, value); class Letters(Element): r""" A generic class for letters | return Crystal_of_letters_type_A_element(self, value); class Letter(Element): r""" A class for letters TEST: sage: from sage.combinat.crystals import Letter sage: parent1 = 1 sage: parent2 = 2 sage: l11 = Letter(parent1, 1) sage: l12 = Letter(parent1, 2) sage: l21 = Letter(parent2, 1) sage: l22 = Letter(parent2, 2) sage: l11 == l11 True sage: l11 == l12 False sage: l11 == l21 False | def __call__(self, value): return Crystals_of_letters_type_A_element(self, value); | e2abaca418af8cb3ca8128b0fd154aed4615a689 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9890/e2abaca418af8cb3ca8128b0fd154aed4615a689/crystals.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
1991,
972,
12,
2890,
16,
460,
4672,
327,
14998,
1094,
1031,
67,
792,
67,
26054,
67,
723,
67,
37,
67,
2956,
12,
2890,
16,
460,
1769,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
1991,
972,
12,
2890,
16,
460,
4672,
327,
14998,
1094,
1031,
67,
792,
67,
26054,
67,
723,
67,
37,
67,
2956,
12,
2890,
16,
460,
1769,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
window.do_itemhit(item, event) | self._windows[window].do_itemhit(item, event) | def do_dialogevent(self, event): gotone, window, item = DialogSelect(event) if gotone: if self._windows.has_key(window): window.do_itemhit(item, event) else: print 'Dialog event for unknown dialog' | 5c3cd713b4dd144a6d58c2949a533a6da58c1cf1 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12029/5c3cd713b4dd144a6d58c2949a533a6da58c1cf1/FrameWork.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
741,
67,
12730,
2575,
12,
2890,
16,
871,
4672,
2363,
476,
16,
2742,
16,
761,
273,
17242,
3391,
12,
2575,
13,
309,
2363,
476,
30,
309,
365,
6315,
13226,
18,
5332,
67,
856,
12,
5668,
4... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
741,
67,
12730,
2575,
12,
2890,
16,
871,
4672,
2363,
476,
16,
2742,
16,
761,
273,
17242,
3391,
12,
2575,
13,
309,
2363,
476,
30,
309,
365,
6315,
13226,
18,
5332,
67,
856,
12,
5668,
4... |
print >>self, self.WARN, self.msg, | print >>self, BasicLogger.WARN, self.msg, | def warning(self, msg, *args): if self.level < 1: return print >>self, self.WARN, self.msg, print >>self, msg %args | 0e572c485ebb264ee85b8afc2865034ce368ae54 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/13054/0e572c485ebb264ee85b8afc2865034ce368ae54/logging.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3436,
12,
2890,
16,
1234,
16,
380,
1968,
4672,
309,
365,
18,
2815,
411,
404,
30,
327,
1172,
1671,
2890,
16,
7651,
3328,
18,
27999,
16,
365,
18,
3576,
16,
1172,
1671,
2890,
16,
1234,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3436,
12,
2890,
16,
1234,
16,
380,
1968,
4672,
309,
365,
18,
2815,
411,
404,
30,
327,
1172,
1671,
2890,
16,
7651,
3328,
18,
27999,
16,
365,
18,
3576,
16,
1172,
1671,
2890,
16,
1234,
... |
def exception_handler(etype, value, tb): | def exception_handler(self, etype, value, tb): | def exception_handler(etype, value, tb): get_flowcontrol().measurement_end() raise etype, value, tb | 0d70cf91480ed1fe1cfdf6e8d9207b886b2509dc /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/6471/0d70cf91480ed1fe1cfdf6e8d9207b886b2509dc/qtflow.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1520,
67,
4176,
12,
2890,
16,
24700,
16,
460,
16,
8739,
4672,
336,
67,
2426,
7098,
7675,
16944,
67,
409,
1435,
1002,
24700,
16,
460,
16,
8739,
282,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1520,
67,
4176,
12,
2890,
16,
24700,
16,
460,
16,
8739,
4672,
336,
67,
2426,
7098,
7675,
16944,
67,
409,
1435,
1002,
24700,
16,
460,
16,
8739,
282,
2,
-100,
-100,
-100,
-100,
-100,
-10... |
def value_setter(self, newvalue): | def value(self, newvalue): | def value_setter(self, newvalue): return | 4e17a7d82e2bac201ec6a68b68f6234bfd884c24 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/1534/4e17a7d82e2bac201ec6a68b68f6234bfd884c24/abc_abstractproperty_rw_deco.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
460,
12,
2890,
16,
394,
1132,
4672,
327,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
460,
12,
2890,
16,
394,
1132,
4672,
327,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
""" | """ | def _set_status(self, value): self._status = value | 8b169bdc1a9ad510d1fa7ac55cf02c0a00396ed8 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/1074/8b169bdc1a9ad510d1fa7ac55cf02c0a00396ed8/__init__.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
542,
67,
2327,
12,
2890,
16,
460,
4672,
365,
6315,
2327,
273,
460,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
542,
67,
2327,
12,
2890,
16,
460,
4672,
365,
6315,
2327,
273,
460,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-10... |
str += "<br>Using tree: %s, branch %s" % (self.tree, self.branch) | str += "<br>Using tree: %s, branch %s, hostname %s" % (self.tree, self.branch, self.machine) | def describe(self): str = "" str += "Getting changes from the Tinderbox service running at %s " \ % self.tinderboxURL str += "<br>Using tree: %s, branch %s" % (self.tree, self.branch) return str | 3431613a58fcc2f2a4363847cc084921453783e7 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/13067/3431613a58fcc2f2a4363847cc084921453783e7/tinderboxpoller.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
6401,
12,
2890,
4672,
609,
273,
1408,
609,
1011,
315,
19213,
3478,
628,
326,
399,
14055,
2147,
1156,
3549,
622,
738,
87,
315,
521,
738,
365,
18,
88,
14055,
2147,
1785,
609,
1011,
3532,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
6401,
12,
2890,
4672,
609,
273,
1408,
609,
1011,
315,
19213,
3478,
628,
326,
399,
14055,
2147,
1156,
3549,
622,
738,
87,
315,
521,
738,
365,
18,
88,
14055,
2147,
1785,
609,
1011,
3532,
... |
fileName = "default_name" | fileName = "BK_default_name" | def download(self): lhcbGroup = credentials.getSelectedGroup() if lhcbGroup == "visitor": c.result = {"success":"false","error":"You are not authorised"} return c.result cl = LHCB_BKKDBClient(getRPCClient('Bookkeeping/BookkeepingManager')) req = () if request.params.has_key("root") and len(request.params["root"]) > 0: req = str(request.params["root"]) else: c.result = {"success":"false","error":"File name field is empty"} return c.result if request.params.has_key("start") and len(request.params["start"]) > 0: startItem = request.params["start"] else: startItem = 0 if request.params.has_key("limit") and len(request.params["limit"]) > 0: maxItems = request.params["limit"] else: maxItems = 25 if request.params.has_key("type") and len(request.params["type"]) > 0: fileType = request.params["type"] else: fileType = "txt" if request.params.has_key("fname") and len(request.params["fname"]) > 0: fileName = request.params["fname"] else: fileName = "default_name" files = self.__showFiles(req,{},startItem,maxItems) files = files["result"] tmp = cl.writePythonOrJobOptions(startItem,maxItems,req,fileType) fileName = fileName + "." + fileType gLogger.info("\033[0;31m - \033[0m",fileName) response.headers['Content-type'] = 'application/x-unknown' response.headers["Content-Disposition"] = "attachment; filename=%s" % fileName return tmp | b1793c46ebc44dfbd4a938440592a639a54d7aab /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/12767/b1793c46ebc44dfbd4a938440592a639a54d7aab/BK.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4224,
12,
2890,
4672,
29143,
7358,
1114,
273,
4448,
18,
588,
7416,
1114,
1435,
309,
29143,
7358,
1114,
422,
315,
3516,
1811,
6877,
276,
18,
2088,
273,
12528,
4768,
15563,
5743,
15937,
1636... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4224,
12,
2890,
4672,
29143,
7358,
1114,
273,
4448,
18,
588,
7416,
1114,
1435,
309,
29143,
7358,
1114,
422,
315,
3516,
1811,
6877,
276,
18,
2088,
273,
12528,
4768,
15563,
5743,
15937,
1636... |
self.change = {'cash': self.data[5], 'percent': self.data[6]} self.volume = {'total': self.data[7], 'daily': self.data[8]} self.value = {'total': float(self.last_price ) * float(self.volume['total']), 'bid': self.data[9], 'ask': self.data[10], 'p_close': self.data[11], 'l_close': self.data[12]} self.range = {'day': self.data[13], 'year': self.data[14]} (self.EPS, self.PE) = (self.data[15], self.data[16]) self.dividend = {'pay_date': self.data[17], 'per_share': self.data[18], 'yeild': self.data[19]} (self.capital, self.exchange) = (self.data[20], self.data[21]) """ Extended Attributes """ (self.short_ratio, self.target_52w) = self.data[22:24] self.EPS_est = {'current_year': self.data[24], 'next_year': self.data[25], 'next_quarter': self.data[26]} self.price_EPS_est = {'current_year': self.data[27], 'next_year': self.data[28], 'next_quarter': self.data[29]} (self.PEG, self.book_value, self.price_book, self.price_sales, self.EBITDA) = self.data[29:34] self.average_move = {'d50': self.data[34], 'd200': self.data[35]} """ Real-Time Attributes """ self.realtime = {'ask': self.data[36], 'bid': self.data[37], 'change': {'percent': self.data[38], 'cash': self.data[40]}, 'last_trade': {'date': self.data[39].split(" - ")[0], 'price': self.data[39].split(" - ")[1]}, 'day_range': self.data[41], 'capital': self.data[42]} | print "Day range (real-time): %s" % Stock.realtime['day_range'] print "Market Capitalization: %s" % Stock.realtime['capital'] print | def __init__(self, symbol): """ Download stock's attributes and attribute them to this object. * Basic attributes: | 2bc56260e439cbf4ac58fccaf7d6d4f92d32f2b3 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/2318/2bc56260e439cbf4ac58fccaf7d6d4f92d32f2b3/YahooFinance.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
3273,
4672,
3536,
13059,
12480,
1807,
1677,
471,
1566,
2182,
358,
333,
733,
18,
225,
380,
7651,
1677,
30,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
3273,
4672,
3536,
13059,
12480,
1807,
1677,
471,
1566,
2182,
358,
333,
733,
18,
225,
380,
7651,
1677,
30,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
line = [ '"%s"' % x.encode(encoding) for x in line ] | line = [ '"%s"' % x.encode(encoding).replace('"', '""') for x in line ] | def to_str(self, encoding='UTF-8'): lines = [] # When schema or columns (or both) are None there is plain # string to string conversion if self.schema and self.columns: for line in self.state.lines: line = [ '"%s"' % self.schema[self.columns[i]].encode(value) for i, value in enumerate(line) ] lines.append(','.join(line)) else: for line in self.state.lines: line = [ '"%s"' % x.encode(encoding) for x in line ] lines.append(','.join(line)) return '\n'.join(lines) | 626f975eddd0c0b2fe69070c72791e25362c7b6b /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12681/626f975eddd0c0b2fe69070c72791e25362c7b6b/itools_csv.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
358,
67,
701,
12,
2890,
16,
2688,
2218,
5159,
17,
28,
11,
4672,
2362,
273,
5378,
468,
5203,
1963,
578,
2168,
261,
280,
3937,
13,
854,
599,
1915,
353,
7351,
468,
533,
358,
533,
4105,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
358,
67,
701,
12,
2890,
16,
2688,
2218,
5159,
17,
28,
11,
4672,
2362,
273,
5378,
468,
5203,
1963,
578,
2168,
261,
280,
3937,
13,
854,
599,
1915,
353,
7351,
468,
533,
358,
533,
4105,
... |
bg,fg=divmod(q,256) | if color: bg,fg=divmod(q,256) else: bg,fg=0,7 | def dumphtml(self): h=self.height w=self.width r="" span="" span_bg=-1 span_fg=-1 for i in range(h*w): q,c=divmod(self.scr[i],256) bg,fg=divmod(q,256) if i==self.cy*w+self.cx: bg=1 fg=7 if (bg!=span_bg or fg!=span_fg or i==h*w-1): if len(span): r+='<span class="f%d b%d">%s</span>'%(span_fg,span_bg,cgi.escape(span.translate(self.trhtml))) span="" span_bg=bg span_fg=fg span+=chr(c) if i%w==w-1: span+='\n' r='<?xml version="1.0" encoding="ISO-8859-1"?><pre class="term">%s</pre>'%r if self.last_html==r: return '<?xml version="1.0"?><idem></idem>' else: self.last_html=r | 9006feef9ed1aece90fd903313a881c66d0a3c47 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/13009/9006feef9ed1aece90fd903313a881c66d0a3c47/ajaxterm.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4657,
2620,
12,
2890,
4672,
366,
33,
2890,
18,
4210,
341,
33,
2890,
18,
2819,
436,
1546,
6,
4548,
1546,
6,
4548,
67,
12370,
29711,
21,
4548,
67,
2137,
29711,
21,
364,
277,
316,
1048,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4657,
2620,
12,
2890,
4672,
366,
33,
2890,
18,
4210,
341,
33,
2890,
18,
2819,
436,
1546,
6,
4548,
1546,
6,
4548,
67,
12370,
29711,
21,
4548,
67,
2137,
29711,
21,
364,
277,
316,
1048,
... |
p = "Hello, World" | p = b"Hello, World" | def test_memmove(self): # large buffers apparently increase the chance that the memory # is allocated in high address space. a = create_string_buffer(1000000) p = "Hello, World" result = memmove(a, p, len(p)) self.failUnlessEqual(a.value, "Hello, World") | d68d131d29380af5a4adffb791b80049dc737c13 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/8546/d68d131d29380af5a4adffb791b80049dc737c13/test_memfunctions.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
3917,
8501,
12,
2890,
4672,
468,
7876,
9664,
29614,
715,
10929,
326,
17920,
716,
326,
3778,
468,
353,
11977,
316,
3551,
1758,
3476,
18,
279,
273,
752,
67,
1080,
67,
4106,
12,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
3917,
8501,
12,
2890,
4672,
468,
7876,
9664,
29614,
715,
10929,
326,
17920,
716,
326,
3778,
468,
353,
11977,
316,
3551,
1758,
3476,
18,
279,
273,
752,
67,
1080,
67,
4106,
12,
... |
self.nat_discovery() | self._lock.acquire() self._running = True self._lock.release() try: self.nat_discovery() finally: self._lock.acquire() self._running = False self._lock.release() | def run(self): self.nat_discovery() | 698c4e731a473819061b683aa113c7cd992f5d04 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9686/698c4e731a473819061b683aa113c7cd992f5d04/NatCheckMsgHandler.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
12,
2890,
4672,
365,
18,
18757,
67,
22268,
1435,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
12,
2890,
4672,
365,
18,
18757,
67,
22268,
1435,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
else: conn, cursor = self.open() | def open_for_store(self): """Open and initialize a connection for storing objects. | 148069014ce5c26f86993f308add743510824dbf /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/10510/148069014ce5c26f86993f308add743510824dbf/oracle.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1696,
67,
1884,
67,
2233,
12,
2890,
4672,
3536,
3678,
471,
4046,
279,
1459,
364,
15729,
2184,
18,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1696,
67,
1884,
67,
2233,
12,
2890,
4672,
3536,
3678,
471,
4046,
279,
1459,
364,
15729,
2184,
18,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-... | |
"--expose-debug-as debugContext", | "--expose-debug-as", "debugContext", | def RunTests(build_dir=None): '''This is just a simple wrapper for running the test through v8_shell_sample. Since v8_shell_sample always returns 0 whether the test passes or fails, buildbot looks at stdout to test for failure. ''' script_dir = google.path_utils.ScriptDir() chrome_dir = google.path_utils.FindUpward(script_dir, "chrome") v8_dir = google.path_utils.FindUpward(script_dir, "v8") if build_dir: v8_shell_sample = os.path.join(build_dir, "v8_shell_sample.exe") else: v8_shell_sample = os.path.join(chrome_dir, "Debug", "v8_shell_sample.exe") # look for Debug version first if not os.path.isfile(v8_shell_sample): v8_shell_sample = os.path.join(chrome_dir, "Release", "v8_shell_sample.exe") cmd = [v8_shell_sample, "--allow-natives-syntax", "--expose-debug-as debugContext", os.path.join(chrome_dir, "browser", "debugger", "resources", "debugger_shell.js"), # TODO Change the location of mjsunit.js from the copy in this # directory to the copy in V8 when switching to use V8 from # code.google.com. # os.path.join(v8_dir, "test", "mjsunit", "mjsunit.js"), os.path.join(chrome_dir, "test", "debugger", "mjsunit.js"), os.path.join(chrome_dir, "test", "debugger", "test_protocol.js") ] (retcode, output) = google.process_utils.RunCommandFull(cmd, collect_output=True) if "Success" in output: return 0 else: return 1 | 12a489db6a540afb6726d00248de914abb33d654 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/5060/12a489db6a540afb6726d00248de914abb33d654/debugger_unittests.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1939,
14650,
12,
3510,
67,
1214,
33,
7036,
4672,
9163,
2503,
353,
2537,
279,
4143,
4053,
364,
3549,
326,
1842,
3059,
331,
28,
67,
10304,
67,
6358,
18,
7897,
331,
28,
67,
10304,
67,
635... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1939,
14650,
12,
3510,
67,
1214,
33,
7036,
4672,
9163,
2503,
353,
2537,
279,
4143,
4053,
364,
3549,
326,
1842,
3059,
331,
28,
67,
10304,
67,
6358,
18,
7897,
331,
28,
67,
10304,
67,
635... |
unicode_comment_pattern = re.compile(r'( |\n|^).. ') | unicode_comment_pattern = re.compile(r'( |\n|^)\.\. ') | def unicode_directive(name, arguments, options, content, lineno, content_offset, block_text, state, state_machine): r""" Convert Unicode character codes (numbers) to characters. Codes may be decimal numbers, hexadecimal numbers (prefixed by ``0x``, ``x``, ``\x``, ``U+``, ``u``, or ``\u``; e.g. ``U+262E``), or XML-style numeric character entities (e.g. ``☮``). Text following ".." is a comment and is ignored. Spaces are ignored, and any other text remains as-is. """ if not isinstance(state, states.SubstitutionDef): error = state_machine.reporter.error( 'Invalid context: the "%s" directive can only be used within a ' 'substitution definition.' % (name), nodes.literal_block(block_text, block_text), line=lineno) return [error] codes = unicode_comment_pattern.split(arguments[0])[0].split() element = nodes.Element() for code in codes: try: decoded = directives.unicode_code(code) except ValueError, err: error = state_machine.reporter.error( 'Invalid character code: %s\n%s: %s' % (code, err.__class__.__name__, err), nodes.literal_block(block_text, block_text), line=lineno) return [error] element += nodes.Text(decoded) return element.children | 56f60cbe4b1a38b544fd6580ce68e75fa4e4d65e /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/1278/56f60cbe4b1a38b544fd6580ce68e75fa4e4d65e/misc.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
5252,
67,
22347,
12,
529,
16,
1775,
16,
702,
16,
913,
16,
7586,
16,
913,
67,
3348,
16,
1203,
67,
955,
16,
919,
16,
919,
67,
9149,
4672,
436,
8395,
4037,
9633,
3351,
6198,
261,
13851,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
5252,
67,
22347,
12,
529,
16,
1775,
16,
702,
16,
913,
16,
7586,
16,
913,
67,
3348,
16,
1203,
67,
955,
16,
919,
16,
919,
67,
9149,
4672,
436,
8395,
4037,
9633,
3351,
6198,
261,
13851,... |
if exists("%s-prev"%dir_path): | if p.isdir("%s-prev"%dir_path): | def _mkdir_keep_prev(dir_path): if exists("%s-prev"%dir_path): local("rm -rf %s-prev"%dir_path) if exists("%s"%dir_path): local("mv %s %s-prev"%(dir_path, dir_path)) local("mkdir -p %s"%dir_path) | 23265700124e040af26c9d5696cf20f1c85a41a4 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12422/23265700124e040af26c9d5696cf20f1c85a41a4/demodata.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
26686,
67,
10102,
67,
10001,
12,
1214,
67,
803,
4672,
309,
293,
18,
291,
1214,
27188,
87,
17,
10001,
28385,
1214,
67,
803,
4672,
1191,
2932,
8864,
300,
5809,
738,
87,
17,
10001,
2... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
26686,
67,
10102,
67,
10001,
12,
1214,
67,
803,
4672,
309,
293,
18,
291,
1214,
27188,
87,
17,
10001,
28385,
1214,
67,
803,
4672,
1191,
2932,
8864,
300,
5809,
738,
87,
17,
10001,
2... |
state = "active" | state = _("active") | def run(self): | d62eb7fc2f8da5cc0b33e5d1229c2a062b1c701d /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/61/d62eb7fc2f8da5cc0b33e5d1229c2a062b1c701d/proximity.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
12,
2890,
4672,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
12,
2890,
4672,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-... |
message = 'Revert "%s"\n\nThis reverts commit %s.\n' \ % (subject, commit.get_id_hash()) | body = '' message = 'Revert "%s"\n\nThis reverts commit %s.\n\n%s\n' \ % (subject, commit.get_id_hash(), body) | def __pick_commit(commit_id, patchname, options): """Pick a commit id. """ commit = git.Commit(commit_id) if options.name: patchname = options.name elif patchname and options.revert: patchname = 'revert-' + patchname if patchname: patchname = find_patch_name(patchname, crt_series.patch_exists) if options.parent: parent = git_id(crt_series, options.parent) else: parent = commit.get_parent() if not options.revert: bottom = parent top = commit_id else: bottom = commit_id top = parent if options.fold: out.start('Folding commit %s' % commit_id) # try a direct git apply first if not git.apply_diff(bottom, top, files = options.file): if options.file: raise CmdException('Patch folding failed') else: git.merge_recursive(bottom, git.get_head(), top) out.done() elif options.update: rev1 = git_id(crt_series, 'HEAD^') rev2 = git_id(crt_series, 'HEAD') files = git.barefiles(rev1, rev2).split('\n') out.start('Updating with commit %s' % commit_id) if not git.apply_diff(bottom, top, files = files): raise CmdException, 'Patch updating failed' out.done() else: message = commit.get_log() if options.revert: if message: subject = message.splitlines()[0] else: subject = commit.get_id_hash() message = 'Revert "%s"\n\nThis reverts commit %s.\n' \ % (subject, commit.get_id_hash()) elif options.expose: message += '(imported from commit %s)\n' % commit.get_id_hash() author_name, author_email, author_date = \ name_email_date(commit.get_author()) out.start('Importing commit %s' % commit_id) newpatch = crt_series.new_patch(patchname, message = message, can_edit = False, unapplied = True, bottom = bottom, top = top, author_name = author_name, author_email = author_email, author_date = author_date) # in case the patch name was automatically generated patchname = newpatch.get_name() # find a patchlog to fork from refbranchname, refpatchname = parse_rev(patchname) if refpatchname: if refbranchname: # assume the refseries is OK, since we already resolved # commit_str to a git_id refseries = Series(refbranchname) else: refseries = crt_series patch = refseries.get_patch(refpatchname) if patch.get_log(): out.info("Log was %s" % newpatch.get_log()) out.info("Setting log to %s\n" % patch.get_log()) newpatch.set_log(patch.get_log()) out.info("Log is now %s" % newpatch.get_log()) else: out.info("No log for %s\n" % patchname) if not options.unapplied: modified = crt_series.push_patch(patchname) else: modified = False if crt_series.empty_patch(patchname): out.done('empty patch') elif modified: out.done('modified') else: out.done() | 1666c983e2260a60f9f6318a03e8cbb1cd320cba /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/12366/1666c983e2260a60f9f6318a03e8cbb1cd320cba/pick.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
11503,
67,
7371,
12,
7371,
67,
350,
16,
4729,
529,
16,
702,
4672,
3536,
17968,
279,
3294,
612,
18,
3536,
3294,
273,
5071,
18,
5580,
12,
7371,
67,
350,
13,
225,
309,
702,
18,
52... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
11503,
67,
7371,
12,
7371,
67,
350,
16,
4729,
529,
16,
702,
4672,
3536,
17968,
279,
3294,
612,
18,
3536,
3294,
273,
5071,
18,
5580,
12,
7371,
67,
350,
13,
225,
309,
702,
18,
52... |
return "ExperimentConfiguration %s for %s" % (self.name, self.experiment_metadata) | return u"ExperimentConfiguration %s for %s" % (self.name, self.experiment_metadata) | def __unicode__(self): return "ExperimentConfiguration %s for %s" % (self.name, self.experiment_metadata) | e0c246b75fbd7beea69fbdb5057dc22d60db6a17 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/7374/e0c246b75fbd7beea69fbdb5057dc22d60db6a17/models.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
9124,
972,
12,
2890,
4672,
327,
315,
22338,
1750,
738,
87,
364,
738,
87,
6,
738,
261,
2890,
18,
529,
16,
365,
18,
18142,
67,
4165,
13,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
9124,
972,
12,
2890,
4672,
327,
315,
22338,
1750,
738,
87,
364,
738,
87,
6,
738,
261,
2890,
18,
529,
16,
365,
18,
18142,
67,
4165,
13,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-1... |
self.listMatch[jobRequirements] = {'LastListMatch': now} | def submitPilot(self, jobDict, director): """ Submit pilot for the given job, this is done from the Thread Pool job """ try: job = jobDict['JobID'] ret = jobDB.getJobAttribute(job, 'Status') if ret['OK'] and not ret['Value'] == MAJOR_WAIT: self.log.warn( 'Job is no longer in %s Status:' % MAJOR_WAIT, job ) return S_ERROR( 'Job is no longer in %s Status:' % MAJOR_WAIT ) self.log.verbose( 'Submitting Pilot' ) ceMask = self.__resolveCECandidates( jobDict ) if not ceMask: return S_ERROR( 'No CE available for job' ) self.workDir = director.workDir workingDirectory = os.path.join( self.workDir, job) if os.path.isdir( workingDirectory ): shutil.rmtree( workingDirectory ) elif os.path.lexists( workingDirectory ): os.remove( workingDirectory ) os.makedirs(workingDirectory) inputSandbox = [] pilotOptions = [] if jobDict['PilotType'].lower()=='private': self.log.verbose('Job %s will be submitted with a privat pilot' % job) ownerDN = jobDict['OwnerDN'] ownerGroup = jobDict['OwnerGroup'] # User Group requirement pilotOptions.append( '-G %s' % jobDict['OwnerGroup'] ) # check if group allows jobsharing ownerGroupProperties = getPropertiesForGroup( ownerGroup ) if not 'JobSharing' in ownerGroupProperties: # Add Owner requirement to pilot pilotOptions.append( "-O '%s'" % ownerDN ) pilotOptions.append( '-o /AgentJobRequirements/PilotType=private' ) pilotOptions.append( '-o /Resources/Computing/InProcess/PilotType=private' ) else: self.log.verbose('Job %s will be submitted with a generic pilot' % job) ownerDN = self.genericPilotDN ownerGroup = self.genericPilotGroup pilotOptions.append( '-o /AgentJobRequirements/PilotType=generic' ) pilotOptions.append( '-o /Resources/Computing/InProcess/PilotType=generic' ) # Requested version of DIRAC pilotOptions.append( '-v %s' % self.diracVersion ) # Requested CPU time pilotOptions.append( '-T %s' % jobDict['MaxCPUTime'] ) # Default Setup. It may be overwriten by the Arguments in the Job JDL pilotOptions.append( '-o /DIRAC/Setup=%s' % self.diracSetup ) # Write JDL retDict = self._prepareJDL( jobDict, pilotOptions, ceMask ) jdl = retDict['JDL'] jobRequirements = retDict['Requirements'] pilots = retDict['Pilots'] if not jdl: try: shutil.rmtree( workingDirectory ) except: pass updateJobStatus( self.log, AGENT_NAME, job, MAJOR_WAIT, MINOR_SUBMIT, logRecord=True ) return S_ERROR( 'Could not create JDL:', job ) # get a valid proxy ret = gProxyManager.getPilotProxyFromDIRACGroup( ownerDN, ownerGroup ) if not ret['OK']: self.log.error( ret['Message'] ) self.log.error( 'Could not get proxy:', 'User "%s", Group "%s"' % ( ownerDN, ownerGroup ) ) try: shutil.rmtree( workingDirectory ) except: pass updateJobStatus( self.log, AGENT_NAME, job, MAJOR_WAIT, MINOR_SUBMIT, logRecord=True ) return S_ERROR( 'Could not get proxy' ) proxy = ret['Value'] # Need to get VOMS extension for the later interctions with WMS ret = gProxyManager.getVOMSAttributes(proxy) if not ret['OK']: updateJobStatus( self.log, AGENT_NAME, job, MAJOR_WAIT, MINOR_SUBMIT, logRecord=True ) return ret if not ret['Value']: updateJobStatus( self.log, AGENT_NAME, job, MAJOR_WAIT, MINOR_SUBMIT, logRecord=True ) return S_ERROR( 'getPilotProxyFromDIRACGroup returns a proxy without VOMS Extensions' ) vomsGroup = ret['Value'][0] # Check that there are available queues for the Job: if self.enableListMatch: availableCEs = [] now = Time.dateTime() if not jobRequirements in self.listMatch: availableCEs = self._listMatch( proxy, job, jdl ) if availableCEs != False: self.listMatch[jobRequirements] = {'LastListMatch': now} self.listMatch[jobRequirements]['AvailableCEs'] = availableCEs else: availableCEs = self.listMatch[jobRequirements]['AvailableCEs'] if not Time.timeInterval( self.listMatch[jobRequirements]['LastListMatch'], self.listMatchDelay * Time.minute ).includes( now ): availableCEs = self._listMatch( proxy, job, jdl ) if availableCEs != False: self.listMatch[jobRequirements]['AvailableCEs'] = availableCEs self.listMatch[jobRequirements] = {'LastListMatch': now} else: del self.listMatch[jobRequirements] else: self.log.verbose( 'LastListMatch', self.listMatch[jobRequirements]['LastListMatch'] ) self.log.verbose( 'AvailableCEs ', availableCEs ) if type(availableCEs) == types.ListType : jobDB.setJobParameter( job, 'Available_CEs' , '%s CEs returned from list-match on %s' % ( len(availableCEs), Time.toString() ) ) if not availableCEs: # FIXME: set Job Minor Status try: shutil.rmtree( workingDirectory ) except: pass updateJobStatus( self.log, AGENT_NAME, job, MAJOR_WAIT, MINOR_SUBMIT, logRecord=True ) return S_ERROR( 'No queue available for job' ) # Now we are ready for the actual submission, so self.log.verbose('Submitting Pilot Agent for job:', job ) submitRet = self._submitPilot( proxy, job, jdl ) try: shutil.rmtree( workingDirectory ) except: pass if not submitRet: updateJobStatus( self.log, AGENT_NAME, job, MAJOR_WAIT, MINOR_SUBMIT, logRecord=True ) return S_ERROR( 'Pilot Submission Failed' ) pilotReference, resourceBroker = submitRet # Now, update the job Minor Status if pilots > 1 : pilotReference = self._getChildrenReferences( proxy, pilotReference, job ) else: pilotReference = [pilotReference] pilotAgentsDB.addPilotReference( pilotReference, job, ownerDN, vomsGroup, broker=resourceBroker, gridType=self.gridMiddleware, requirements=jobRequirements ) ret = jobDB.getJobAttribute(job, 'Status') if ret['OK'] and ret['Value'] == MAJOR_WAIT: updateJobStatus( self.log, AGENT_NAME, job, MAJOR_WAIT, MINOR_RESPONSE, logRecord=True ) else: self.log.warn( 'Job is no longer in %s Status:' % MAJOR_WAIT, job ) | d109b88106e7666d141cb418ec5b6770850724aa /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/12864/d109b88106e7666d141cb418ec5b6770850724aa/Director.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4879,
52,
22797,
12,
2890,
16,
1719,
5014,
16,
2657,
280,
4672,
3536,
17320,
293,
22797,
364,
326,
864,
1719,
16,
333,
353,
2731,
628,
326,
4884,
8828,
1719,
3536,
775,
30,
1719,
273,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4879,
52,
22797,
12,
2890,
16,
1719,
5014,
16,
2657,
280,
4672,
3536,
17320,
293,
22797,
364,
326,
864,
1719,
16,
333,
353,
2731,
628,
326,
4884,
8828,
1719,
3536,
775,
30,
1719,
273,
... | |
'L,TimeDateStamp', 'L,PointerToSymbolTable', 'L,NumberOfSymbols', 'H,SizeOfOptionalHeader', | 'I,TimeDateStamp', 'I,PointerToSymbolTable', 'I,NumberOfSymbols', 'H,SizeOfOptionalHeader', | def __init__(self, **args): for key, value in args.items(): setattr(self, key, value) | d0ca75794f69f1d09fd6d31d40ba2f1e9b3b7fb4 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/213/d0ca75794f69f1d09fd6d31d40ba2f1e9b3b7fb4/pefile.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
2826,
1968,
4672,
364,
498,
16,
460,
316,
833,
18,
3319,
13332,
9241,
12,
2890,
16,
498,
16,
460,
13,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
2826,
1968,
4672,
364,
498,
16,
460,
316,
833,
18,
3319,
13332,
9241,
12,
2890,
16,
498,
16,
460,
13,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
if 'extraCredentials' in credDict : if type( credDict[ 'extraCredentials' ] ) not in ( types.StringType, types.UnicodeType ): self.__authLogger.warn( "The credentials seem to be forwarded by a host, but it is not a trusted one" ) | if 'extraCredentials' in credDict and not 'group' in credDict: credDict[ 'group' ] = credDict[ 'extraCredentials' ] if 'DN' in credDict: if credDict[ 'group' ] == self.__hostsGroup: if not self.getHostNickName( credDict ): self.__authLogger.warn( "Host is invalid" ) | def authQuery( self, methodQuery, credDict ): """ Check if the query is authorized for a credentials dictionary | 0dafdb1dfadfa20ea51341368a5cdf30a132733c /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/12864/0dafdb1dfadfa20ea51341368a5cdf30a132733c/AuthManager.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1357,
1138,
12,
365,
16,
707,
1138,
16,
6197,
5014,
262,
30,
3536,
2073,
309,
326,
843,
353,
10799,
364,
279,
4448,
3880,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1357,
1138,
12,
365,
16,
707,
1138,
16,
6197,
5014,
262,
30,
3536,
2073,
309,
326,
843,
353,
10799,
364,
279,
4448,
3880,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
self.outer_state.openBracket() | def parseToken(self, token): # This is an anonymous class if and only if the next # non-whitespace token is a bracket if token == '\n': return self if token == '{': self.outer_state.openBracket() self.outer_state.addAnonClass() elif token == '}': self.outer_state.closeBracket() elif token in ['"', "'"]: return IgnoreState(token, self) return self.outer_state | 8bafac57fb9af3eec6fb2ed77be499c561049f35 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/7126/8bafac57fb9af3eec6fb2ed77be499c561049f35/JavaCommon.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1109,
1345,
12,
2890,
16,
1147,
4672,
468,
1220,
353,
392,
13236,
667,
309,
471,
1338,
309,
326,
1024,
468,
1661,
17,
18777,
1147,
353,
279,
9843,
309,
1147,
422,
2337,
82,
4278,
327,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1109,
1345,
12,
2890,
16,
1147,
4672,
468,
1220,
353,
392,
13236,
667,
309,
471,
1338,
309,
326,
1024,
468,
1661,
17,
18777,
1147,
353,
279,
9843,
309,
1147,
422,
2337,
82,
4278,
327,
... | |
except LookupCanceled(): | except LookupCanceled: | def update_completion_worker(self): while True: start_word, to_select = self.update_completion_q.get() self.update_completion_t0 = time.time() try: lang_word_list = self.do_lookup(start_word, to_select) except LookupCanceled(): pass else: gobject.idle_add(self.update_completion_callback, lang_word_list, to_select, start_word, time.time() - self.update_completion_t0) self.update_completion_t0 = None self.update_completion_q.task_done() | a02fdbce87b3eaeebe5fd2ba19e0ba63975467eb /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/14305/a02fdbce87b3eaeebe5fd2ba19e0ba63975467eb/ui.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1089,
67,
11469,
67,
10124,
12,
2890,
4672,
1323,
1053,
30,
787,
67,
1095,
16,
358,
67,
4025,
273,
365,
18,
2725,
67,
11469,
67,
85,
18,
588,
1435,
365,
18,
2725,
67,
11469,
67,
88,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1089,
67,
11469,
67,
10124,
12,
2890,
4672,
1323,
1053,
30,
787,
67,
1095,
16,
358,
67,
4025,
273,
365,
18,
2725,
67,
11469,
67,
85,
18,
588,
1435,
365,
18,
2725,
67,
11469,
67,
88,
... |
'start=%s' % self.start), 'reason=%s' % self.reason), 'args=%s' % self.args), 'end=%s' % self.end), 'object=%s' % self.object), | 'start=%s' % self.start, 'reason=%s' % self.reason, 'args=%s' % self.args, 'end=%s' % self.end, 'object=%s' % self.object, | def __str__(self): # this is a bad hack, please supply an implementation res = ' '.join([ 'start=%s' % self.start), 'reason=%s' % self.reason), 'args=%s' % self.args), 'end=%s' % self.end), 'object=%s' % self.object), ]) return res | 7d942c2f4a7e8f483ad5b90e36ffc05a872090d2 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/6934/7d942c2f4a7e8f483ad5b90e36ffc05a872090d2/exceptions.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
701,
972,
12,
2890,
4672,
468,
333,
353,
279,
5570,
11769,
16,
9582,
14467,
392,
4471,
400,
273,
296,
2418,
5701,
3816,
296,
1937,
5095,
87,
11,
738,
365,
18,
1937,
16,
296,
1057... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
701,
972,
12,
2890,
4672,
468,
333,
353,
279,
5570,
11769,
16,
9582,
14467,
392,
4471,
400,
273,
296,
2418,
5701,
3816,
296,
1937,
5095,
87,
11,
738,
365,
18,
1937,
16,
296,
1057... |
bosh.dirs['Custom1'], Image(r'images/custom1.png'), _(bosh.inisettings['custom1txt']))) | bosh.dirs['Custom1'], Image(r'images/custom1.png'), _(bosh.inisettings['custom1txt']))) | def InitStatusBar(): """Initialize status bar links.""" #--Bash Status/LinkBar #BashStatusBar.buttons.append(App_Fallout3()) BashStatusBar.buttons.append(Fose_Button()) BashStatusBar.buttons.append(AutoQuit_Button()) BashStatusBar.buttons.append( Fallout3_Button( bosh.dirs['app'].join('Fallout3.exe'), Image(r'images/fallout3.png'), _("Launch Fallout3"), _("Launch Fallout3 + FOSE"), '')) BashStatusBar.buttons.append( App_Button( bosh.dirs['app'].join('GECK.exe'), Image(r'images/geck.png'), _("Launch GECK"), _("Launch GECK + FOSE"), '-editor')) BashStatusBar.buttons.append( App_Button( bosh.dirs['app'].join('fomm\\fomm.exe'), Image(r'images/database_connect.png'), _("Launch FOMM"))) BashStatusBar.buttons.append( App_FO3Edit( (bosh.dirs['FO3EditPath'], '-view'), Image(r'images/brick_edit.png'), _("Launch FO3View"))) BashStatusBar.buttons.append( App_FO3Edit( dirs['FO3EditPath']), Image(r'images/brick.png'), _("Launch FO3Edit"))) BashStatusBar.buttons.append( App_FO3Edit( (dirs['FO3EditPath'],'-translate'), Image(r'images/brick_error.png'), _("Launch FO3Trans"))) BashStatusBar.buttons.append( App_BOSS( bosh.dirs['app'].join('Data\\BOSS-F.bat'), Image(r'images/Boss1.png'), _("Launch BOSS"))) BashStatusBar.buttons.append( App_BOSS( bosh.dirs['app'].join('Data\\BOSS-F.bat'), Image(r'images/boss2.png'), _("Launch BOSS"))) if bosh.inisettings['showmodelingtoollaunchers']: BashStatusBar.buttons.append( App_Button( bosh.dirs['BlenderPath'], Image(r'images/blender.png'), _("Launch Blender"))) BashStatusBar.buttons.append( App_Button( bosh.dirs['GmaxPath'], Image(r'images/gmax.png'), _("Launch Gmax"))) BashStatusBar.buttons.append( App_Button( bosh.dirs['MayaPath'], Image(r'images/maya.png'), _("Launch Maya"))) BashStatusBar.buttons.append( App_Button( bosh.dirs['MaxPath'], Image(r'images/max.png'), _("Launch 3dsMax"))) if bosh.inisettings['showmodelingtoollaunchers'] or bosh.inisettings['showtexturetoollaunchers']: BashStatusBar.buttons.append( App_Button( bosh.dirs['NifskopePath'], Image(r'images/nifskope.png'), _("Launch Nifskope"))) if bosh.inisettings['showtexturetoollaunchers']: BashStatusBar.buttons.append( App_Button( bosh.dirs['GIMP'], Image(r'images/gimp.png'), _("Launch GIMP"))) BashStatusBar.buttons.append( App_Button( bosh.dirs['Photoshop'], Image(r'images/photoshop.png'), _("Launch Photoshop"))) BashStatusBar.buttons.append( App_Button( bosh.dirs['Artweaver'], Image(r'images/artweaver.png'), _("Launch Artweaver"))) BashStatusBar.buttons.append( App_Button( bosh.dirs['PaintNET'], Image(r'images/paint.net.png'), _("Launch Paint.NET"))) BashStatusBar.buttons.append( App_Button( bosh.dirs['DDSConverter'], Image(r'images/ddsconverter.png'), _("Launch DDSConverter"))) if bosh.inisettings['showaudiotoollaunchers']: BashStatusBar.buttons.append( App_Button( bosh.dirs['Audacity'], Image(r'images/audacity.png'), _("Launch Audacity"))) BashStatusBar.buttons.append( App_Button( bosh.dirs['Fraps'], Image(r'images/fraps.png'), _("Launch Fraps"))) BashStatusBar.buttons.append( App_Button( bosh.dirs['NPP'], Image(r'images/notepad++.png'), _("Launch Notepad++"))) BashStatusBar.buttons.append( App_Button( bosh.dirs['Custom1'], Image(r'images/custom1.png'), _(bosh.inisettings['custom1txt']))) BashStatusBar.buttons.append( App_Button( bosh.dirs['Custom2'], Image(r'images/custom2.png'), _(bosh.inisettings['custom2txt']))) BashStatusBar.buttons.append( App_Button( bosh.dirs['Custom3'], Image(r'images/custom3.png'), _(bosh.inisettings['custom3txt']))) BashStatusBar.buttons.append( App_Button( bosh.dirs['Custom4'], Image(r'images/custom4.png'), _(bosh.inisettings['custom4txt']))) BashStatusBar.buttons.append(App_BashMon()) BashStatusBar.buttons.append(App_DocBrowser()) BashStatusBar.buttons.append(App_ModChecker()) BashStatusBar.buttons.append(App_Help()) | 3ac43907fa076fea1e8d682219e2b28fc7419f7b /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/6682/3ac43907fa076fea1e8d682219e2b28fc7419f7b/basher.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4378,
1482,
5190,
13332,
3536,
7520,
1267,
4653,
4716,
12123,
468,
413,
38,
961,
2685,
19,
2098,
5190,
468,
38,
961,
1482,
5190,
18,
16016,
18,
6923,
12,
3371,
67,
42,
287,
14533,
23,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4378,
1482,
5190,
13332,
3536,
7520,
1267,
4653,
4716,
12123,
468,
413,
38,
961,
2685,
19,
2098,
5190,
468,
38,
961,
1482,
5190,
18,
16016,
18,
6923,
12,
3371,
67,
42,
287,
14533,
23,
... |
self.log_info ('sending %s' % repr(data)) | self.log_info('sending %s' % repr(data)) | def send (self, data): if self.debug: self.log_info ('sending %s' % repr(data)) self.out_buffer = self.out_buffer + data self.initiate_send() | 52c4de7c23c31563e4f0316b68865e17bb8c0f33 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/8125/52c4de7c23c31563e4f0316b68865e17bb8c0f33/asyncore.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1366,
261,
2890,
16,
501,
4672,
309,
365,
18,
4148,
30,
365,
18,
1330,
67,
1376,
2668,
87,
2846,
738,
87,
11,
738,
8480,
12,
892,
3719,
365,
18,
659,
67,
4106,
273,
365,
18,
659,
6... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1366,
261,
2890,
16,
501,
4672,
309,
365,
18,
4148,
30,
365,
18,
1330,
67,
1376,
2668,
87,
2846,
738,
87,
11,
738,
8480,
12,
892,
3719,
365,
18,
659,
67,
4106,
273,
365,
18,
659,
6... |
sage: LCM([1,2,3,4,5/3]) | sage: LCM([1,2,3,4,5]) | def lcm(a, b=None, integer=False): """ The least common multiple of a and b, or if a is a list and b is omitted the least common multiple of all elements of v. NOTE: Use integer=True to make this vastly faster if you are working with lists of integers. INPUT: a -- number b -- number (optional) integer -- (default: False); if True, do an integer LCM or v -- vector integer -- (default: False); if True, do an integer LCM NOTE -- this is *vastly* faster than doing the generic LCM EXAMPLES: sage: LCM(97,100) 9700 sage: LCM(0,2) 0 sage: LCM(-3,-5) 15 sage: LCM([1,2,3,4,5/3]) 60 sage: v = LCM(range(1,10000),integer=True) # *very* fast! sage: len(str(v)) 4349 """ if integer: if isinstance(a,(list,tuple)): return sage.rings.integer.LCM_list(a) else: return integer_ring.ZZ(a).lcm(\ integer_ring.ZZ(b)) if isinstance(a, (list,tuple)): return __LCM_list(a) if not isinstance(a, RingElement): a = integer_ring.ZZ(a) return a.lcm(b) | bdc03f71082d359479de239409797484223e36e0 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9417/bdc03f71082d359479de239409797484223e36e0/arith.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
328,
7670,
12,
69,
16,
324,
33,
7036,
16,
3571,
33,
8381,
4672,
3536,
1021,
4520,
2975,
3229,
434,
279,
471,
324,
16,
578,
309,
279,
353,
279,
666,
471,
324,
353,
16195,
326,
4520,
2... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
328,
7670,
12,
69,
16,
324,
33,
7036,
16,
3571,
33,
8381,
4672,
3536,
1021,
4520,
2975,
3229,
434,
279,
471,
324,
16,
578,
309,
279,
353,
279,
666,
471,
324,
353,
16195,
326,
4520,
2... |
rsp = self.apiPost("reviewrequests/%s/draft" % review_request.id) self.assertEqual(rsp['stat'], 'ok') | rsp = self.apiPost("reviewrequests/%s/reviews" % review_request.id) self.assertEqual(rsp['stat'], 'ok') self.assertTrue('review' in rsp) review_id = rsp['review']['id'] | def test_get_review_screenshot_comments(self): """Testing the GET reviewrequests/<id>/reviews/draft/screenshot-comments API""" screenshot_comment_text = "Test screenshot comment" x, y, w, h = 2, 2, 10, 10 | 6f62171a9876a999f22f6ebd5ed31ea85071c2f8 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/1600/6f62171a9876a999f22f6ebd5ed31ea85071c2f8/tests.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
588,
67,
16041,
67,
29435,
67,
9231,
12,
2890,
4672,
3536,
22218,
326,
4978,
10725,
11420,
28177,
350,
16893,
266,
7061,
19,
17153,
19,
29435,
17,
9231,
1491,
8395,
17179,
67,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
588,
67,
16041,
67,
29435,
67,
9231,
12,
2890,
4672,
3536,
22218,
326,
4978,
10725,
11420,
28177,
350,
16893,
266,
7061,
19,
17153,
19,
29435,
17,
9231,
1491,
8395,
17179,
67,
... |
x509 = store.get_current_cert() | pem = store.get_current_cert().as_pem() | def verifyCallback(self, ok, store): # Returning 1 means any error is ignored and SSL checking continues log.debug('TwistedProtocolWrapper.verifyCallback') global trusted_until_shutdown_site_certs, \ trusted_until_shutdown_invalid_site_certs, \ unknown_issuer if not ok: try: err = store.get_error() x509 = store.get_current_cert() # Check temporarily trusted certificates pem = x509.as_pem() if err not in unknown_issuer: # Check if we are temporarily ignoring errors with this # certificate. acceptedErrList = trusted_until_shutdown_invalid_site_certs.get(pem) if acceptedErrList is not None and err in acceptedErrList: log.debug('Ignoring certificate error %d' %err) return 1 self.untrustedCertificates.append(pem) return ok if pem in trusted_until_shutdown_site_certs: log.debug('Found temporarily trusted site cert') return 1 # Check permanently trusted certificates # XXX Why does this need to be commit()? refresh() does not # XXX seem pick up changes made in main thread. # Perhaps this should just use the main view (or the main # view should be passed in) and not do any refresh or commit # self.repositoryView.commit() q = schema.ns('osaf.framework.certstore', self.repositoryView).sslTrustedSiteCertificatesQuery for cert in q: if cert.pemAsString() == pem: log.debug('Found permanently trusted site cert') return 1 self.untrustedCertificates.append(pem) except: # This is ok, we MUST return a value and not raise log.exception('SSL verifyCallback raised exception') | 89df4286345b6769b8583f5b9eb368a23770c242 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/9228/89df4286345b6769b8583f5b9eb368a23770c242/ssl.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3929,
2428,
12,
2890,
16,
1529,
16,
1707,
4672,
468,
2000,
310,
404,
4696,
1281,
555,
353,
5455,
471,
7419,
6728,
29359,
613,
18,
4148,
2668,
23539,
25444,
5752,
3611,
18,
8705,
2428,
61... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3929,
2428,
12,
2890,
16,
1529,
16,
1707,
4672,
468,
2000,
310,
404,
4696,
1281,
555,
353,
5455,
471,
7419,
6728,
29359,
613,
18,
4148,
2668,
23539,
25444,
5752,
3611,
18,
8705,
2428,
61... |
return self.search_package_directories(qualified_name, pyx_suffixes, pos) | print "Pyrex.Compiler.Main.find_pyx_file:", qualified_name, pos result = self.search_package_directories(qualified_name, pyx_suffixes, pos) print "Pyrex.Compiler.Main.find_pyx_file: result =", result return result | def find_pyx_file(self, qualified_name, pos): # Search include path for the .pyx file corresponding to the # given fully-qualified module name, as for find_pxd_file(). return self.search_package_directories(qualified_name, pyx_suffixes, pos) | 664108392cf067623c44e45621a7e50e7fdc07c6 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/2099/664108392cf067623c44e45621a7e50e7fdc07c6/Main.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1104,
67,
2074,
92,
67,
768,
12,
2890,
16,
9348,
67,
529,
16,
949,
4672,
468,
225,
5167,
2341,
589,
364,
326,
263,
2074,
92,
585,
4656,
358,
326,
468,
225,
864,
7418,
17,
19724,
1605... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1104,
67,
2074,
92,
67,
768,
12,
2890,
16,
9348,
67,
529,
16,
949,
4672,
468,
225,
5167,
2341,
589,
364,
326,
263,
2074,
92,
585,
4656,
358,
326,
468,
225,
864,
7418,
17,
19724,
1605... |
tempdata = tempdata.encode(self.setup['encoding']) | try: tempdata = tempdata.encode(self.setup['encoding']) except UnicodeEncodeError, e: self.logger.error("Error encoding file %s:\n %s" % \ (entry.get('name'), e)) return False | def VerifyConfigFile(self, entry, _): '''Install ConfigFile Entry''' # configfile verify is permissions check + content check permissionStatus = self.VerifyDirectory(entry, _) tbin = False if entry.get('encoding', 'ascii') == 'base64': tempdata = binascii.a2b_base64(entry.text) tbin = True elif entry.get('empty', 'false') == 'true': tempdata = '' else: if entry.text == None: self.logger.error("Cannot verify incomplete ConfigFile %s" % \ (entry.get('name'))) return False tempdata = entry.text if type(tempdata) == unicode: tempdata = tempdata.encode(self.setup['encoding']) try: content = open(entry.get('name')).read() except IOError, error: if error.strerror == "No such file or directory": # print diff for files that don't exist (yet) content = '' else: self.logger.error("Failed to read %s: %s" % \ (error.filename, error.strerror)) return False # comparison should be done with fingerprints or # md5sum so it would be faster for big binary files contentStatus = content == tempdata if not contentStatus: if tbin or not isString(content): entry.set('current_bfile', binascii.b2a_base64(content)) nqtext = entry.get('qtext', '') nqtext += '\nBinary file, no printable diff' else: do_diff = True rawdiff = [] start = time.time() longtime = False for x in difflib.ndiff(content.split('\n'), tempdata.split('\n')): now = time.time() rawdiff.append(x) if now - start > 5 and not longtime: self.logger.info("Diff of %s taking a long time" % \ (entry.get('name'))) longtime = True elif now - start > 30: self.logger.error("Diff of %s took too long; giving up" % \ (entry.get('name'))) do_diff = False break if do_diff: diff = '\n'.join(rawdiff) entry.set("current_bdiff", binascii.b2a_base64(diff)) | f940deb8419060c6a5685ad460efa232ba1a8aad /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/11867/f940deb8419060c6a5685ad460efa232ba1a8aad/POSIX.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
8553,
13705,
12,
2890,
16,
1241,
16,
389,
4672,
9163,
6410,
31988,
3841,
26418,
468,
642,
768,
3929,
353,
4371,
866,
397,
913,
866,
4132,
1482,
273,
365,
18,
8097,
2853,
12,
4099,
16,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
8553,
13705,
12,
2890,
16,
1241,
16,
389,
4672,
9163,
6410,
31988,
3841,
26418,
468,
642,
768,
3929,
353,
4371,
866,
397,
913,
866,
4132,
1482,
273,
365,
18,
8097,
2853,
12,
4099,
16,
... |
res = run_sql(query, (recID, str(digit1)+str(digit2))) | res = run_sql(query, (recID, str(digit1)+str(digit2)+'%')) | def print_record(recID, format='hb', ot='', ln=CFG_SITE_LANG, decompress=zlib.decompress, search_pattern=None, user_info=None, verbose=0): """Prints record 'recID' formatted accoding to 'format'.""" _ = gettext_set_language(ln) out = "" # sanity check: record_exist_p = record_exists(recID) if record_exist_p == 0: # doesn't exist return out # New Python BibFormat procedure for formatting # Old procedure follows further below # We must still check some special formats, but these # should disappear when BibFormat improves. if not (CFG_BIBFORMAT_USE_OLD_BIBFORMAT \ or format.lower().startswith('t') \ or format.lower().startswith('hm') \ or str(format[0:3]).isdigit() \ or ot): # Unspecified format is hd if format == '': format = 'hd' if record_exist_p == -1 and get_output_format_content_type(format) == 'text/html': # HTML output displays a default value for deleted records. # Other format have to deal with it. out += _("The record has been deleted.") else: out += call_bibformat(recID, format, ln, search_pattern=search_pattern, user_info=user_info, verbose=verbose) # at the end of HTML brief mode, print the "Detailed record" functionality: if format.lower().startswith('hb') and \ format.lower() != 'hb_p': out += websearch_templates.tmpl_print_record_brief_links( ln = ln, recID = recID, ) return out # Old PHP BibFormat procedure for formatting # print record opening tags, if needed: if format == "marcxml" or format == "oai_dc": out += " <record>\n" out += " <header>\n" for oai_id in get_fieldvalues(recID, CFG_OAI_ID_FIELD): out += " <identifier>%s</identifier>\n" % oai_id out += " <datestamp>%s</datestamp>\n" % get_modification_date(recID) out += " </header>\n" out += " <metadata>\n" if format.startswith("xm") or format == "marcxml": # look for detailed format existence: query = "SELECT value FROM bibfmt WHERE id_bibrec=%s AND format=%s" res = run_sql(query, (recID, format), 1) if res and record_exist_p == 1: # record 'recID' is formatted in 'format', so print it out += "%s" % decompress(res[0][0]) else: # record 'recID' is not formatted in 'format' -- they are not in "bibfmt" table; so fetch all the data from "bibXXx" tables: if format == "marcxml": out += """ <record xmlns="http://www.loc.gov/MARC21/slim">\n""" out += " <controlfield tag=\"001\">%d</controlfield>\n" % int(recID) elif format.startswith("xm"): out += """ <record>\n""" out += " <controlfield tag=\"001\">%d</controlfield>\n" % int(recID) if record_exist_p == -1: # deleted record, so display only OAI ID and 980: oai_ids = get_fieldvalues(recID, CFG_OAI_ID_FIELD) if oai_ids: out += "<datafield tag=\"%s\" ind1=\"%s\" ind2=\"%s\"><subfield code=\"%s\">%s</subfield></datafield>\n" % \ (CFG_OAI_ID_FIELD[0:3], CFG_OAI_ID_FIELD[3:4], CFG_OAI_ID_FIELD[4:5], CFG_OAI_ID_FIELD[5:6], oai_ids[0]) out += "<datafield tag=\"980\" ind1=\"\" ind2=\"\"><subfield code=\"c\">DELETED</subfield></datafield>\n" else: # controlfields query = "SELECT b.tag,b.value,bb.field_number FROM bib00x AS b, bibrec_bib00x AS bb "\ "WHERE bb.id_bibrec=%s AND b.id=bb.id_bibxxx AND b.tag LIKE '00%%' "\ "ORDER BY bb.field_number, b.tag ASC" res = run_sql(query, (recID, )) for row in res: field, value = row[0], row[1] value = encode_for_xml(value) out += """ <controlfield tag="%s" >%s</controlfield>\n""" % \ (encode_for_xml(field[0:3]), value) # datafields i = 1 # Do not process bib00x and bibrec_bib00x, as # they are controlfields. So start at bib01x and # bibrec_bib00x (and set i = 0 at the end of # first loop) for digit1 in range(0, 10): for digit2 in range(i, 10): bx = "bib%d%dx" % (digit1, digit2) bibx = "bibrec_bib%d%dx" % (digit1, digit2) query = "SELECT b.tag,b.value,bb.field_number FROM %s AS b, %s AS bb "\ "WHERE bb.id_bibrec=%%s AND b.id=bb.id_bibxxx AND b.tag LIKE %%s%%%% "\ "ORDER BY bb.field_number, b.tag ASC" % (bx, bibx) res = run_sql(query, (recID, str(digit1)+str(digit2))) field_number_old = -999 field_old = "" for row in res: field, value, field_number = row[0], row[1], row[2] ind1, ind2 = field[3], field[4] if ind1 == "_" or ind1 == "": ind1 = " " if ind2 == "_" or ind2 == "": ind2 = " " # print field tag if field_number != field_number_old or field[:-1] != field_old[:-1]: if field_number_old != -999: out += """ </datafield>\n""" out += """ <datafield tag="%s" ind1="%s" ind2="%s">\n""" % \ (encode_for_xml(field[0:3]), encode_for_xml(ind1), encode_for_xml(ind2)) field_number_old = field_number field_old = field # print subfield value value = encode_for_xml(value) out += """ <subfield code="%s">%s</subfield>\n""" % \ (encode_for_xml(field[-1:]), value) # all fields/subfields printed in this run, so close the tag: if field_number_old != -999: out += """ </datafield>\n""" i = 0 # Next loop should start looking at bib%0 and bibrec_bib00x # we are at the end of printing the record: out += " </record>\n" elif format == "xd" or format == "oai_dc": # XML Dublin Core format, possibly OAI -- select only some bibXXx fields: out += """ <dc xmlns="http://purl.org/dc/elements/1.1/" xmlns:xsi="http://www.w3.org/2001/XMLSchema-instance" xsi:schemaLocation="http://purl.org/dc/elements/1.1/ http://www.openarchives.org/OAI/1.1/dc.xsd">\n""" if record_exist_p == -1: out += "" else: for f in get_fieldvalues(recID, "041__a"): out += " <language>%s</language>\n" % f for f in get_fieldvalues(recID, "100__a"): out += " <creator>%s</creator>\n" % encode_for_xml(f) for f in get_fieldvalues(recID, "700__a"): out += " <creator>%s</creator>\n" % encode_for_xml(f) for f in get_fieldvalues(recID, "245__a"): out += " <title>%s</title>\n" % encode_for_xml(f) for f in get_fieldvalues(recID, "65017a"): out += " <subject>%s</subject>\n" % encode_for_xml(f) for f in get_fieldvalues(recID, "8564_u"): out += " <identifier>%s</identifier>\n" % encode_for_xml(f) for f in get_fieldvalues(recID, "520__a"): out += " <description>%s</description>\n" % encode_for_xml(f) out += " <date>%s</date>\n" % get_creation_date(recID) out += " </dc>\n" elif len(format) == 6 and str(format[0:3]).isdigit(): # user has asked to print some fields only if format == "001": out += "<!--%s-begin-->%s<!--%s-end-->\n" % (format, recID, format) else: vals = get_fieldvalues(recID, format) for val in vals: out += "<!--%s-begin-->%s<!--%s-end-->\n" % (format, val, format) elif format.startswith('t'): ## user directly asked for some tags to be displayed only if record_exist_p == -1: out += get_fieldvalues_alephseq_like(recID, ["001", CFG_OAI_ID_FIELD, "980"]) else: out += get_fieldvalues_alephseq_like(recID, ot) elif format == "hm": if record_exist_p == -1: out += "<pre>" + cgi.escape(get_fieldvalues_alephseq_like(recID, ["001", CFG_OAI_ID_FIELD, "980"])) + "</pre>" else: out += "<pre>" + cgi.escape(get_fieldvalues_alephseq_like(recID, ot)) + "</pre>" elif format.startswith("h") and ot: ## user directly asked for some tags to be displayed only if record_exist_p == -1: out += "<pre>" + get_fieldvalues_alephseq_like(recID, ["001", CFG_OAI_ID_FIELD, "980"]) + "</pre>" else: out += "<pre>" + get_fieldvalues_alephseq_like(recID, ot) + "</pre>" elif format == "hd": # HTML detailed format if record_exist_p == -1: out += _("The record has been deleted.") else: # look for detailed format existence: query = "SELECT value FROM bibfmt WHERE id_bibrec=%s AND format=%s" res = run_sql(query, (recID, format), 1) if res: # record 'recID' is formatted in 'format', so print it out += "%s" % decompress(res[0][0]) else: # record 'recID' is not formatted in 'format', so try to call BibFormat on the fly or use default format: out_record_in_format = call_bibformat(recID, format, ln, search_pattern=search_pattern, user_info=user_info, verbose=verbose) if out_record_in_format: out += out_record_in_format else: out += websearch_templates.tmpl_print_record_detailed( ln = ln, recID = recID, ) elif format.startswith("hb_") or format.startswith("hd_"): # underscore means that HTML brief/detailed formats should be called on-the-fly; suitable for testing formats if record_exist_p == -1: out += _("The record has been deleted.") else: out += call_bibformat(recID, format, ln, search_pattern=search_pattern, user_info=user_info, verbose=verbose) elif format.startswith("hx"): # BibTeX format, called on the fly: if record_exist_p == -1: out += _("The record has been deleted.") else: out += call_bibformat(recID, format, ln, search_pattern=search_pattern, user_info=user_info, verbose=verbose) elif format.startswith("hs"): # for citation/download similarity navigation links: if record_exist_p == -1: out += _("The record has been deleted.") else: out += '<a href="%s">' % websearch_templates.build_search_url(recid=recID, ln=ln) # firstly, title: titles = get_fieldvalues(recID, "245__a") if titles: for title in titles: out += "<strong>%s</strong>" % title else: # usual title not found, try conference title: titles = get_fieldvalues(recID, "111__a") if titles: for title in titles: out += "<strong>%s</strong>" % title else: # just print record ID: out += "<strong>%s %d</strong>" % (get_field_i18nname("record ID", ln), recID) out += "</a>" # secondly, authors: authors = get_fieldvalues(recID, "100__a") + get_fieldvalues(recID, "700__a") if authors: out += " - %s" % authors[0] if len(authors) > 1: out += " <em>et al</em>" # thirdly publication info: publinfos = get_fieldvalues(recID, "773__s") if not publinfos: publinfos = get_fieldvalues(recID, "909C4s") if not publinfos: publinfos = get_fieldvalues(recID, "037__a") if not publinfos: publinfos = get_fieldvalues(recID, "088__a") if publinfos: out += " - %s" % publinfos[0] else: # fourthly publication year (if not publication info): years = get_fieldvalues(recID, "773__y") if not years: years = get_fieldvalues(recID, "909C4y") if not years: years = get_fieldvalues(recID, "260__c") if years: out += " (%s)" % years[0] else: # HTML brief format by default if record_exist_p == -1: out += _("The record has been deleted.") else: query = "SELECT value FROM bibfmt WHERE id_bibrec=%s AND format=%s" res = run_sql(query, (recID, format)) if res: # record 'recID' is formatted in 'format', so print it out += "%s" % decompress(res[0][0]) else: # record 'recID' is not formatted in 'format', so try to call BibFormat on the fly: or use default format: if CFG_WEBSEARCH_CALL_BIBFORMAT: out_record_in_format = call_bibformat(recID, format, ln, search_pattern=search_pattern, user_info=user_info, verbose=verbose) if out_record_in_format: out += out_record_in_format else: out += websearch_templates.tmpl_print_record_brief( ln = ln, recID = recID, ) else: out += websearch_templates.tmpl_print_record_brief( ln = ln, recID = recID, ) # at the end of HTML brief mode, print the "Detailed record" functionality: if format == 'hp' or format.startswith("hb_") or format.startswith("hd_"): pass # do nothing for portfolio and on-the-fly formats else: out += websearch_templates.tmpl_print_record_brief_links( ln = ln, recID = recID, ) # print record closing tags, if needed: if format == "marcxml" or format == "oai_dc": out += " </metadata>\n" out += " </record>\n" return out | b0a7a42e5f62a4288226e4e7ee86b0635ca9725d /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/12027/b0a7a42e5f62a4288226e4e7ee86b0635ca9725d/search_engine.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1172,
67,
3366,
12,
3927,
734,
16,
740,
2218,
76,
70,
2187,
15835,
2218,
2187,
7211,
33,
19727,
67,
20609,
67,
10571,
16,
16824,
33,
94,
2941,
18,
323,
14706,
16,
1623,
67,
4951,
33,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1172,
67,
3366,
12,
3927,
734,
16,
740,
2218,
76,
70,
2187,
15835,
2218,
2187,
7211,
33,
19727,
67,
20609,
67,
10571,
16,
16824,
33,
94,
2941,
18,
323,
14706,
16,
1623,
67,
4951,
33,
... |
buildno = int(buildno) | def _sys_version(): """ Returns a parsed version of Python's sys.version as tuple (version, buildno, builddate, compiler) referring to the Python version, build number, build date/time as string and the compiler identification string. Note that unlike the Python sys.version, the returned value for the Python version will always include the patchlevel (it defaults to '.0'). """ global _sys_version_cache if _sys_version_cache is not None: return _sys_version_cache version, buildno, builddate, buildtime, compiler = \ _sys_version_parser.match(sys.version).groups() buildno = int(buildno) builddate = builddate + ' ' + buildtime l = string.split(version, '.') if len(l) == 2: l.append('0') version = string.join(l, '.') _sys_version_cache = (version, buildno, builddate, compiler) return _sys_version_cache | edc8c8b0f98d7d995167ee059f8b3e1fe33dc85f /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/3187/edc8c8b0f98d7d995167ee059f8b3e1fe33dc85f/platform.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
9499,
67,
1589,
13332,
225,
3536,
2860,
279,
2707,
1177,
434,
6600,
1807,
2589,
18,
1589,
487,
3193,
261,
1589,
16,
1361,
2135,
16,
25666,
330,
449,
340,
16,
5274,
13,
1278,
20245,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
9499,
67,
1589,
13332,
225,
3536,
2860,
279,
2707,
1177,
434,
6600,
1807,
2589,
18,
1589,
487,
3193,
261,
1589,
16,
1361,
2135,
16,
25666,
330,
449,
340,
16,
5274,
13,
1278,
20245,
... | |
self.root = None | self.root = self.prefix = None | def initialize_options(self): self.zip_ok = None self.install_dir = self.script_dir = self.exclude_scripts = None self.index_url = None self.find_links = None self.build_directory = None self.args = None self.optimize = self.record = None self.upgrade = self.always_copy = self.multi_version = None self.editable = self.no_deps = self.allow_hosts = None self.root = None | ee68f241feb1442ac6f1288cb945258b7ed0e70a /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/495/ee68f241feb1442ac6f1288cb945258b7ed0e70a/easy_install.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4046,
67,
2116,
12,
2890,
4672,
365,
18,
4450,
67,
601,
273,
599,
365,
18,
5425,
67,
1214,
273,
365,
18,
4263,
67,
1214,
273,
365,
18,
10157,
67,
12827,
273,
599,
365,
18,
1615,
67,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
4046,
67,
2116,
12,
2890,
4672,
365,
18,
4450,
67,
601,
273,
599,
365,
18,
5425,
67,
1214,
273,
365,
18,
4263,
67,
1214,
273,
365,
18,
10157,
67,
12827,
273,
599,
365,
18,
1615,
67,
... |
c_call = c_call + " %s->%s = xmlStrdup(%s);\n" % (args[0][0], | c_call = c_call + " %s->%s = xmlStrdup((const xmlChar *)%s);\n" % (args[0][0], | def print_function_wrapper(name, output, export, include): global py_types global unknown_types global functions global skipped_modules try: (desc, ret, args, file) = functions[name] except: print "failed to get function %s infos" return if skipped_modules.has_key(file): return 0 if skip_function(name) == 1: return 0 c_call = ""; format="" format_args="" c_args="" c_return="" c_convert="" for arg in args: # This should be correct if arg[1][0:6] == "const ": arg[1] = arg[1][6:] c_args = c_args + " %s %s;\n" % (arg[1], arg[0]) if py_types.has_key(arg[1]): (f, t, n, c) = py_types[arg[1]] if f != None: format = format + f if t != None: format_args = format_args + ", &pyobj_%s" % (arg[0]) c_args = c_args + " PyObject *pyobj_%s;\n" % (arg[0]) c_convert = c_convert + \ " %s = (%s) Py%s_Get(pyobj_%s);\n" % (arg[0], arg[1], t, arg[0]); else: format_args = format_args + ", &%s" % (arg[0]) if c_call != "": c_call = c_call + ", "; c_call = c_call + "%s" % (arg[0]) else: if skipped_types.has_key(arg[1]): return 0 if unknown_types.has_key(arg[1]): lst = unknown_types[arg[1]] lst.append(name) else: unknown_types[arg[1]] = [name] return -1 if format != "": format = format + ":%s" % (name) if ret[0] == 'void': if file == "python_accessor": if args[1][1] == "char *" or args[1][1] == "xmlChar *": c_call = "\n if (%s->%s != NULL) xmlFree(%s->%s);\n" % ( args[0][0], args[1][0], args[0][0], args[1][0]) c_call = c_call + " %s->%s = xmlStrdup(%s);\n" % (args[0][0], args[1][0], args[1][0]) else: c_call = "\n %s->%s = %s;\n" % (args[0][0], args[1][0], args[1][0]) else: c_call = "\n %s(%s);\n" % (name, c_call); ret_convert = " Py_INCREF(Py_None);\n return(Py_None);\n" elif py_types.has_key(ret[0]): (f, t, n, c) = py_types[ret[0]] c_return = " %s c_retval;\n" % (ret[0]) if file == "python_accessor" and ret[2] != None: c_call = "\n c_retval = %s->%s;\n" % (args[0][0], ret[2]) else: c_call = "\n c_retval = %s(%s);\n" % (name, c_call); ret_convert = " py_retval = libxml_%sWrap((%s) c_retval);\n" % (n,c) ret_convert = ret_convert + " return(py_retval);\n" elif py_return_types.has_key(ret[0]): (f, t, n, c) = py_return_types[ret[0]] c_return = " %s c_retval;\n" % (ret[0]) c_call = "\n c_retval = %s(%s);\n" % (name, c_call); ret_convert = " py_retval = libxml_%sWrap((%s) c_retval);\n" % (n,c) ret_convert = ret_convert + " return(py_retval);\n" else: if skipped_types.has_key(ret[0]): return 0 if unknown_types.has_key(ret[0]): lst = unknown_types[ret[0]] lst.append(name) else: unknown_types[ret[0]] = [name] return -1 include.write("PyObject * ") include.write("libxml_%s(PyObject *self, PyObject *args);\n" % (name)) export.write(" { \"%s\", libxml_%s, METH_VARARGS, NULL },\n" % (name, name)) if file == "python": # Those have been manually generated return 1 if file == "python_accessor" and ret[0] != "void" and ret[2] == None: # Those have been manually generated return 1 output.write("PyObject *\n") output.write("libxml_%s(PyObject *self, PyObject *args) {\n" % (name)) if ret[0] != 'void': output.write(" PyObject *py_retval;\n") if c_return != "": output.write(c_return) if c_args != "": output.write(c_args) if format != "": output.write("\n if (!PyArg_ParseTuple(args, \"%s\"%s))\n" % (format, format_args)) output.write(" return(NULL);\n") if c_convert != "": output.write(c_convert) output.write(c_call) output.write(ret_convert) output.write("}\n\n") return 1 | d2379010b1aa04ff87876d5b6dfd42ce2e12928f /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12659/d2379010b1aa04ff87876d5b6dfd42ce2e12928f/generator.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1172,
67,
915,
67,
8376,
12,
529,
16,
876,
16,
3359,
16,
2341,
4672,
2552,
2395,
67,
2352,
2552,
5917,
67,
2352,
2552,
4186,
2552,
9700,
67,
6400,
225,
775,
30,
261,
5569,
16,
325,
1... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1172,
67,
915,
67,
8376,
12,
529,
16,
876,
16,
3359,
16,
2341,
4672,
2552,
2395,
67,
2352,
2552,
5917,
67,
2352,
2552,
4186,
2552,
9700,
67,
6400,
225,
775,
30,
261,
5569,
16,
325,
1... |
print "open", askopenfilename(filetypes=[("all filez", "*")]).encode(enc) | print "open", askopenfilename(filetypes=[("all files", "*")]).encode(enc) | def askdirectory (**options): "Ask for a directory, and return the file name" return Directory(**options).show() | 74c94771d2936383f0e28e732dfaf5b272657cc8 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12029/74c94771d2936383f0e28e732dfaf5b272657cc8/tkFileDialog.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
6827,
5149,
261,
636,
2116,
4672,
315,
23663,
364,
279,
1867,
16,
471,
327,
326,
585,
508,
6,
327,
8930,
12,
636,
2116,
2934,
4500,
1435,
225,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
6827,
5149,
261,
636,
2116,
4672,
315,
23663,
364,
279,
1867,
16,
471,
327,
326,
585,
508,
6,
327,
8930,
12,
636,
2116,
2934,
4500,
1435,
225,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-1... |
'staff': u.status==STAFF, | 'administrateur': u.status>COACH, | def liste_echeances(request, **kwargs): """View: list of echeances of this administrator's groups or users. Filters on group and cours.""" # préparation des paramètres # les clés de ce dictionnaire *doivent* être des chaînes params = dict(request.GET.items()) for key, value in params.items(): if not isinstance(key, str): del params[key] params[smart_str(key)] = value # u = utilisateur courant u = request.session['v'] # recup de la liste des echéances gérées par cet admin le = Echeance.objects.filter(groupe__in=u.groupes_list()) # recup de la liste des objets correspondant aux paramètres passés le = le.filter(**params) # construction des filtres lg = {'nom': 'groupe', 'title': _('group'), 'valeurs': [{'id': g.id, 'libel':g.nom} for g in u.groupes_list()]} lc = {'nom': 'cours', 'title': _('course'), 'valeurs': [{'id': c.id, 'libel':c.titre(u.langue)} for c in u.cours_list()]} lf = [filtre for filtre in [lg,lc] if len(filtre['valeurs'])>1] lf = makefilters(request.GET, lf) return render_to_response('coaching/liste_echeance.html', {'visiteur': u.prenom_nom(), 'staff': u.status==STAFF, 'client': u.groupe.client, 'liste_echeances': le, 'liste_filtres': lf,}) | 5622003147f995391ac97c1031fcc18f24ef8d30 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/11552/5622003147f995391ac97c1031fcc18f24ef8d30/views.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
31614,
67,
8987,
73,
6872,
12,
2293,
16,
2826,
4333,
4672,
3536,
1767,
30,
666,
434,
425,
18706,
6872,
434,
333,
22330,
1807,
3252,
578,
3677,
18,
225,
14475,
603,
1041,
471,
276,
4390,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
31614,
67,
8987,
73,
6872,
12,
2293,
16,
2826,
4333,
4672,
3536,
1767,
30,
666,
434,
425,
18706,
6872,
434,
333,
22330,
1807,
3252,
578,
3677,
18,
225,
14475,
603,
1041,
471,
276,
4390,
... |
dc.SetPen(drawingPen) | def Draw(self, dc, boundingRect, styles, bitmapBrush): item = self._item | eb038a8d592a9472b37a26775af15c1eb21b246f /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9228/eb038a8d592a9472b37a26775af15c1eb21b246f/CalendarCanvas.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
10184,
12,
2890,
16,
6744,
16,
7552,
6120,
16,
5687,
16,
9389,
19114,
1218,
4672,
761,
273,
365,
6315,
1726,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
10184,
12,
2890,
16,
6744,
16,
7552,
6120,
16,
5687,
16,
9389,
19114,
1218,
4672,
761,
273,
365,
6315,
1726,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-1... | |
for n in xrange(0, len(nodes)): p = gv.getv(nodes[n], "pos") | for n, n_gv in nodes.iteritems(): p = gv.getv(n_gv, "pos") | def graph_draw(g, pos=None, size=(15, 15), pin=False, layout= "neato", maxiter=None, ratio= "fill", overlap="prism", sep=None, splines=False, vsize=0.1, penwidth=1.0, elen=None, gprops={}, vprops={}, eprops={}, vcolor=None, ecolor=None, vcmap=matplotlib.cm.jet, vnorm=True, ecmap=matplotlib.cm.jet, enorm=True, output= "", output_format= "auto", returngv=False, fork=False, seed=0): r"""Draw a graph using graphviz. Parameters ---------- g : Graph Graph to be used. pos : PropertyMap or tuple of PropertyMaps (optional, default: None) Vertex property maps containing the x and y coordinates of the vertices. size : tuple of scalars (optional, default: (15,15)) Size (in centimeters) of the canvas. pin : bool (default: False) If True, the vertices are not moved from their initial position. layout : string (default: "neato") Layout engine to be used. Possible values are "neato", "fdp", "dot", "circo", "twopi" and "arf". maxiter : int (default: None) If specified, limits the maximum number of iterations. ratio : string or float (default: "fill") Sets the aspect ratio (drawing height/drawing width) for the drawing. Note that this is adjusted before the 'size' attribute constraints are enforced. If ratio is numeric, it is taken as the desired aspect ratio. Then, if the actual aspect ratio is less than the desired ratio, the drawing height is scaled up to achieve the desired ratio; if the actual ratio is greater than that desired ratio, the drawing width is scaled up. If ratio = "fill" and the size attribute is set, node positions are scaled, separately in both x and y, so that the final drawing exactly fills the specified size. If ratio = "compress" and the size attribute is set, dot attempts to compress the initial layout to fit in the given size. This achieves a tighter packing of nodes but reduces the balance and symmetry. This feature only works in dot. If ratio = "expand", the size attribute is set, and both the width and the height of the graph are less than the value in size, node positions are scaled uniformly until at least one dimension fits size exactly. Note that this is distinct from using size as the desired size, as here the drawing is expanded before edges are generated and all node and text sizes remain unchanged. If ratio = "auto", the page attribute is set and the graph cannot be drawn on a single page, then size is set to an "ideal" value. In particular, the size in a given dimension will be the smallest integral multiple of the page size in that dimension which is at least half the current size. The two dimensions are then scaled independently to the new size. This feature only works in dot. overlap : bool or string (default: "prism") Determines if and how node overlaps should be removed. Nodes are first enlarged using the sep attribute. If True, overlaps are retained. If the value is "scale", overlaps are removed by uniformly scaling in x and y. If the value is False, node overlaps are removed by a Voronoi-based technique. If the value is "scalexy", x and y are separately scaled to remove overlaps. If sfdp is available, one can set overlap to "prism" to use a proximity graph-based algorithm for overlap removal. This is the preferred technique, though "scale" and False can work well with small graphs. This technique starts with a small scaling up, controlled by the overlap_scaling attribute, which can remove a significant portion of the overlap. The prism option also accepts an optional non-negative integer suffix. This can be used to control the number of attempts made at overlap removal. By default, overlap="prism" is equivalent to overlap="prism1000". Setting overlap="prism0" causes only the scaling phase to be run. If the value is "compress", the layout will be scaled down as much as possible without introducing any overlaps, obviously assuming there are none to begin with. sep : float (default: None) Specifies margin to leave around nodes when removing node overlap. This guarantees a minimal non-zero distance between nodes. splines : bool (default: False) If True, the edges are drawn as splines and routed around the vertices. vsize : float, PropertyMap, or tuple (default: 0.1) Default vertex size (width and height). If a tuple is specified, the first value should be a property map, and the second is a scale factor. penwidth : float, PropertyMap or tuple (default: 1.0) Specifies the width of the pen, in points, used to draw lines and curves, including the boundaries of edges and clusters. It has no effect on text. Default vertex size (width and height). If a tuple is specified, the first value should be a property map, and the second is a scale factor. elen : float or PropertyMap (default: None) Preferred edge length, in inches. gprops : dict (default: {}) Additional graph properties, as a dictionary. The keys are the property names, and the values must be convertible to string. vprops : dict (default: {}) Additional vertex properties, as a dictionary. The keys are the property names, and the values must be convertible to string, or vertex property maps, with values convertible to strings. eprops : dict (default: {}) Additional edge properties, as a dictionary. The keys are the property names, and the values must be convertible to string, or edge property maps, with values convertible to strings. vcolor : string or PropertyMap (default: None) Drawing color for vertices. If the valued supplied is a property map, the values must be scalar types, whose color values are obtained from the 'vcmap' argument. ecolor : string or PropertyMap (default: None) Drawing color for edges. If the valued supplied is a property map, the values must be scalar types, whose color values are obtained from the 'ecmap' argument. vcmap : matplotlib.colors.Colormap (default: matplotlib.cm.jet) Vertex color map. vnorm : bool (default: True) Normalize vertex color values to the [0,1] range. ecmap : matplotlib.colors.Colormap (default: matplotlib.cm.jet) Edge color map. enorm : bool (default: True) Normalize edge color values to the [0,1] range. output : string (default: "") Output file name. output_format : string (default: "auto") Output file format. Possible values are "auto", "xlib", "ps", "svg", "svgz", "fig", "mif", "hpgl", "pcl", "png", "gif", "dia", "imap", "cmapx". If the value is "auto", the format is guessed from the 'output' parameter, or 'xlib' if it is empty. If the value is None, no output is produced. returngv : bool (default: False) Return the graph object used internally with the gv module. fork : bool (default: False) If true, the program is forked before drawing. This is used as a work-around for a bug in graphviz, where the exit() function is called, which would cause the calling program to end. This is always assumed 'True', if output_format = 'xlib'. seed : int (default: 0) Seed for the random number generator. If the value 0, a different random value is used each time. Returns ------- pos : PropertyMap Vector vertex property map with the x and y coordinates of the vertices. gv : gv.digraph or gv.graph (optional, only if returngv == True) Internally used graphviz graph. Notes ----- This function is a wrapper for the [graphviz] python routines. Extensive additional documentation for the graph, vertex and edge properties is available at: http://www.graphviz.org/doc/info/attrs.html. Examples -------- >>> from numpy import * >>> from numpy.random import seed, zipf >>> seed(42) >>> g = gt.random_graph(1000, lambda: min(zipf(2.4), 40), ... lambda i,j: exp(abs(i-j)), directed=False) >>> # extract largest component >>> comp = gt.label_components(g) >>> h = gt.vertex_hist(g, comp) >>> max_comp = h[1][list(h[0]).index(max(h[0]))] >>> g.remove_vertex_if(lambda v: comp[v] != max_comp) >>> >>> deg = g.degree_property_map("out") >>> deg.get_array()[:] = 2*(sqrt(deg.get_array()[:])*0.5 + 0.4) >>> ebet = gt.betweenness(g)[1] >>> ebet.get_array()[:] *= 4000 >>> ebet.get_array()[:] += 10 >>> gt.graph_draw(g, vsize=deg, vcolor=deg, elen=10, ecolor=ebet, ... penwidth=ebet, overlap="prism", output="graph-draw.png") <...> .. figure:: graph-draw.png :align: center Kamada-Kawai force-directed layout of a graph with a power-law degree distribution, and dissortative degree correlation. The vertex size and color indicate the degree, and the edge color and width the edge betweeness centrality. References ---------- .. [graphviz] http://www.graphviz.org """ if output != "" and output != None: output = os.path.expanduser(output) # check opening file for writing, since graphviz will bork if it is not # possible to open file if os.path.dirname(output) != "" and \ not os.access(os.path.dirname(output), os.W_OK): raise IOError("cannot write to " + os.path.dirname(output)) if g.is_directed(): gvg = gv.digraph("G") else: gvg = gv.graph("G") if layout == "arf": layout = "neato" pos = arf_layout(g, pos=pos) pin = True if pos != None: # copy user-supplied property if isinstance(pos, PropertyMap): pos = ungroup_vector_property(g, pos, [0,1]) else: pos = (g.copy_property(pos[0]), g.copy_property(pos[1])) if type(vsize) == tuple: s = g.new_vertex_property("double") g.copy_property(vsize[0], s) s.a *= vsize[1] vsize = s if type(penwidth) == tuple: s = g.new_edge_property("double") g.copy_property(penwidth[0], s) s.a *= penwidth[1] penwidth = s # main graph properties gv.setv(gvg,"outputorder", "edgesfirst") gv.setv(gvg,"mode", "major") if overlap == False: overlap = "false" else: overlap = "true" if isinstance(overlap,str): gv.setv(gvg,"overlap", overlap) if sep != None: gv.setv(gvg,"sep", str(sep)) if splines: gv.setv(gvg,"splines", "true") gv.setv(gvg,"ratio", str(ratio)) gv.setv(gvg,"size", "%f,%f" % (size[0]/2.54,size[1]/2.54)) # centimeters if maxiter != None: gv.setv(gvg,"maxiter", str(maxiter)) if seed == 0: seed = numpy.random.randint(sys.maxint) if type(seed) == int: gv.setv(gvg, "start", "%d" % seed) else: gv.setv(gvg, "start", seed) # apply all user supplied properties for k,val in gprops.iteritems(): if isinstance(val, PropertyMap): gv.setv(gvg, k, str(val[g])) else: gv.setv(gvg, k, str(val)) # normalize color properties if vcolor != None and not isinstance(vcolor, str): minmax = [float("inf"), -float("inf")] for v in g.vertices(): c = vcolor[v] minmax[0] = min(c,minmax[0]) minmax[1] = max(c,minmax[1]) if minmax[0] == minmax[1]: minmax[1] += 1 if vnorm: vnorm = matplotlib.colors.normalize(vmin=minmax[0], vmax=minmax[1]) else: vnorm = lambda x: x if ecolor != None and not isinstance(ecolor, str): minmax = [float("inf"), -float("inf")] for e in g.edges(): c = ecolor[e] minmax[0] = min(c,minmax[0]) minmax[1] = max(c,minmax[1]) if minmax[0] == minmax[1]: minmax[1] += 1 if enorm: enorm = matplotlib.colors.normalize(vmin=minmax[0], vmax=minmax[1]) else: enorm = lambda x: x nodes = [] edges = [] # add nodes for v in g.vertices(): n = gv.node(gvg,str(g.vertex_index[v])) if type(vsize) == PropertyMap: vw = vh = vsize[v] else: vw = vh = vsize gv.setv(n, "width", "%g" % vw) gv.setv(n, "height", "%g" % vh) gv.setv(n, "style", "filled") gv.setv(n, "color", "black") # apply color if vcolor != None: if isinstance(vcolor,str): gv.setv(n, "fillcolor", vcolor) else: color = tuple([int(c*255.0) for c in vcmap(vnorm(vcolor[v]))]) gv.setv(n, "fillcolor", "#%.2x%.2x%.2x%.2x" % color) else: gv.setv(n, "fillcolor", "red") gv.setv(n, "label", "") # user supplied position if pos != None: gv.setv(n, "pos", "%f,%f" % (pos[0][v],pos[1][v])) gv.setv(n, "pin", str(pin)) # apply all user supplied properties for k,val in vprops.iteritems(): if isinstance(val, PropertyMap): gv.setv(n, k, str(val[v])) else: gv.setv(n, k, str(val)) nodes.append(n) for e in g.edges(): ge = gv.edge(nodes[g.vertex_index[e.source()]], nodes[g.vertex_index[e.target()]]) gv.setv(ge, "arrowsize", "0.3") if g.is_directed(): gv.setv(ge, "arrowhead", "vee") # apply color if ecolor != None: if isinstance(ecolor,str): gv.setv(ge, "color", ecolor) else: color = tuple([int(c*255.0) for c in ecmap(enorm(ecolor[e]))]) gv.setv(ge, "color", "#%.2x%.2x%.2x%.2x" % color) # apply edge length if elen != None: if isinstance(elen, PropertyMap): gv.setv(ge, "len", str(elen[e])) else: gv.setv(ge, "len", str(elen)) # apply width if penwidth != None: if isinstance(penwidth, PropertyMap): gv.setv(ge, "penwidth", str(penwidth[e])) else: gv.setv(ge, "penwidth", str(penwidth)) # apply all user supplied properties for k,v in eprops.iteritems(): if isinstance(v, PropertyMap): gv.setv(ge, k, str(v[e])) else: gv.setv(ge, k, str(v)) edges.append(ge) gv.layout(gvg, layout) gv.render(gvg, "dot", "/dev/null") # retrieve positions if pos == None: pos = (g.new_vertex_property("double"), g.new_vertex_property("double")) for n in xrange(0, len(nodes)): p = gv.getv(nodes[n], "pos") p = p.split(",") pos[0][g.vertex(n)] = float(p[0]) pos[1][g.vertex(n)] = float(p[1]) if output_format == "auto": if output == "": output_format = "xlib" elif output != None: output_format = output.split(".")[-1] # if using xlib we need to fork the process, otherwise good ol' graphviz # will call exit() when the window is closed if output_format == "xlib" or fork: pid = os.fork() if pid == 0: gv.render(gvg, output_format, output) os._exit(0) # since we forked, it's good to be sure if output_format != "xlib": os.wait() elif output != None: gv.render(gvg, output_format, output) # I don't get this, but it seems necessary pos[0].get_array()[:] /= 100 pos[1].get_array()[:] /= 100 pos = group_vector_property(g, pos) if returngv: return pos, gv else: gv.rm(gvg) del gvg return pos | 19b68a13db29808d40d53cad6e35f01a3e20fe34 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/3516/19b68a13db29808d40d53cad6e35f01a3e20fe34/__init__.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2667,
67,
9446,
12,
75,
16,
949,
33,
7036,
16,
963,
28657,
3600,
16,
4711,
3631,
7092,
33,
8381,
16,
3511,
33,
315,
4644,
31093,
3113,
25743,
33,
7036,
16,
7169,
33,
315,
5935,
3113,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
2667,
67,
9446,
12,
75,
16,
949,
33,
7036,
16,
963,
28657,
3600,
16,
4711,
3631,
7092,
33,
8381,
16,
3511,
33,
315,
4644,
31093,
3113,
25743,
33,
7036,
16,
7169,
33,
315,
5935,
3113,
... |
self.thread_count = MAX_SANE_THREAD_COUNT | self.thread_count = thread_count | def __init__(self): self._ips = set() self.thread_count = MAX_SANE_THREAD_COUNT super(NameServers, self).__init__() | 1430af6a930a8f31d1cf827010d7c93e35ea3017 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/4170/1430af6a930a8f31d1cf827010d7c93e35ea3017/nameserver_list.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
4672,
365,
6315,
7146,
273,
444,
1435,
365,
18,
5930,
67,
1883,
273,
2650,
67,
1883,
2240,
12,
461,
11913,
16,
365,
2934,
972,
2738,
972,
1435,
2,
0,
0,
0,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
4672,
365,
6315,
7146,
273,
444,
1435,
365,
18,
5930,
67,
1883,
273,
2650,
67,
1883,
2240,
12,
461,
11913,
16,
365,
2934,
972,
2738,
972,
1435,
2,
-100,
-100... |
lines = [s.replace('\t', SPACE8) for s in text.split('\n')] | lines = [s.replace('\t', SPACE8) for s in text.splitlines()] | def _split_header(self, text): maxheaderlen = self.__maxheaderlen # Find out whether any lines in the header are really longer than # maxheaderlen characters wide. There could be continuation lines # that actually shorten it. Also, replace hard tabs with 8 spaces. lines = [s.replace('\t', SPACE8) for s in text.split('\n')] for line in lines: if len(line) > maxheaderlen: break else: # No line was actually longer than maxheaderlen characters, so # just return the original unchanged. return text rtn = [] for line in text.split('\n'): splitline = [] # Short lines can remain unchanged if len(line.replace('\t', SPACE8)) <= maxheaderlen: splitline.append(line) rtn.append(SEMINLTAB.join(splitline)) else: oldlen = len(line) # Try to break the line on semicolons, but if that doesn't # work, try to split on folding whitespace. while len(line) > maxheaderlen: i = line.rfind(';', 0, maxheaderlen) if i < 0: break splitline.append(line[:i]) line = line[i+1:].lstrip() if len(line) <> oldlen: # Splitting on semis worked splitline.append(line) rtn.append(SEMINLTAB.join(splitline)) continue # Splitting on semis didn't help, so try to split on # whitespace. parts = re.split(r'(\s+)', line) # Watch out though for "Header: longnonsplittableline" if parts[0].endswith(':') and len(parts) == 3: rtn.append(line) continue first = parts.pop(0) sublines = [first] acc = len(first) while parts: len0 = len(parts[0]) len1 = len(parts[1]) if acc + len0 + len1 < maxheaderlen: sublines.append(parts.pop(0)) sublines.append(parts.pop(0)) acc += len0 + len1 else: # Split it here, but don't forget to ignore the # next whitespace-only part splitline.append(EMPTYSTRING.join(sublines)) del parts[0] first = parts.pop(0) sublines = [first] acc = len(first) splitline.append(EMPTYSTRING.join(sublines)) rtn.append(NLTAB.join(splitline)) return NL.join(rtn) | 9c54d8a782b36e843b35e10a09a65914f90f8757 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12029/9c54d8a782b36e843b35e10a09a65914f90f8757/Generator.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
4939,
67,
3374,
12,
2890,
16,
977,
4672,
943,
3374,
1897,
273,
365,
16186,
1896,
3374,
1897,
468,
4163,
596,
2856,
1281,
2362,
316,
326,
1446,
854,
8654,
7144,
2353,
468,
943,
3374,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
4939,
67,
3374,
12,
2890,
16,
977,
4672,
943,
3374,
1897,
273,
365,
16186,
1896,
3374,
1897,
468,
4163,
596,
2856,
1281,
2362,
316,
326,
1446,
854,
8654,
7144,
2353,
468,
943,
3374,... |
self.out.append(r'\DUfootnotemark{%s}{%s}{' % | self.out.append(r'\DUfootnotemark{%s}{%s}{' % | def visit_footnote_reference(self, node): href = '' if 'refid' in node: href = node['refid'] elif 'refname' in node: href = self.document.nameids[node['refname']] # if self.use_latex_footnotes: # TODO: insert footnote content at (or near) this place # print "footnote-ref to", node['refid'] # footnotes = (self.document.footnotes + # self.document.autofootnotes + # self.document.symbol_footnotes) # for footnote in footnotes: # # print footnote['ids'] # if node.get('refid', '') in footnote['ids']: # print 'matches', footnote['ids'] format = self.settings.footnote_references if format == 'brackets': self.append_hypertargets(node) self.out.append('\\hyperlink{%s}{[' % href) self.context.append(']}') else: self.fallbacks['footnotes'] = PreambleCmds.footnotes # TODO: second argument = backlink id self.out.append(r'\DUfootnotemark{%s}{%s}{' % (node['ids'][0], href)) self.context.append('}') | d1a332cc00a1b10df4c2b8fa6efdd3ca79dbf763 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/1278/d1a332cc00a1b10df4c2b8fa6efdd3ca79dbf763/__init__.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3757,
67,
10923,
7652,
67,
6180,
12,
2890,
16,
756,
4672,
3897,
273,
875,
309,
296,
1734,
350,
11,
316,
756,
30,
3897,
273,
756,
3292,
1734,
350,
3546,
1327,
296,
1734,
529,
11,
316,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3757,
67,
10923,
7652,
67,
6180,
12,
2890,
16,
756,
4672,
3897,
273,
875,
309,
296,
1734,
350,
11,
316,
756,
30,
3897,
273,
756,
3292,
1734,
350,
3546,
1327,
296,
1734,
529,
11,
316,
... |
if (len(flagrow) == 0) and (not clip): if (len(maskflag) == 0): raise Exception, 'maskflag is undefined' masks = s.create_mask(maskflag) | def sdflag(sdfile, antenna, scanlist, field, iflist, pollist, maskflag, flagrow, clip, clipminmax, clipoutside, flagmode, outfile, outform, overwrite, plotlevel): casalog.origin('sdflag') ### ### Now the actual task code ### try: myp=None if sdfile=='': raise Exception, 'sdfile is undefined' filename = os.path.expandvars(sdfile) filename = os.path.expanduser(filename) if not os.path.exists(filename): s = "File '%s' not found." % (filename) raise Exception, s # Default file name if ( outfile == '' ): project = sdfile.rstrip('/') + '_f' else: project = outfile outfilename = os.path.expandvars(project) outfilename = os.path.expanduser(outfilename) if not overwrite and flagmode!='restore': if os.path.exists(outfilename): s = "Output file '%s' exist." % (outfilename) raise Exception, s s = sd.scantable(sdfile,average=False,antenna=antenna) #check the format of the infile if isinstance(sdfile, str): if os.path.isdir(filename) and os.path.exists(filename+'/table.info'): if os.path.exists(filename+'/table.f1'): format = 'MS2' else: format = 'ASAP' else: format = 'SDFITS' if ( abs(plotlevel) > 1 ): # print summary of input data #print "Initial Scantable:" #print s casalog.post( "Initial Scantable:" ) casalog.post( s._summary() ) casalog.post( '--------------------------------------------------------------------------------' ) # Default file name #if ( outfile == '' ): # project = sdfile + '_f' #elif ( outfile == 'none' ): # project = sdfile # outform = format # overwrite=True #else: # project = outfile # get telescope name #'ATPKSMB', 'ATPKSHOH', 'ATMOPRA', 'DSS-43' (Tid), 'CEDUNA', and 'HOBART' #Currently only channel based flag is allowed # so make sure the input data is channel unit_in=s.get_unit() if unit_in!='channel': s.set_unit('channel') antennaname = s.get_antennaname() # Select scan and field sel = sd.selector() # Set up scanlist if ( type(scanlist) == list ): # is a list scans = scanlist else: # is a single int, make into list scans = [ scanlist ] # Now select them if ( len(scans) > 0 ): sel.set_scans(scans) # Select source names if ( field != '' ): sel.set_name(field) # NOTE: currently can only select one # set of names this way, will probably # need to do a set_query eventually # Select IFs if ( type(iflist) == list ): # is a list ifs = iflist else: # is a single int, make into list ifs = [ iflist ] if ( len(ifs) > 0 ): # Do any IF selection sel.set_ifs(ifs) # Select Pol if ( type(pollist) == list ): pols = pollist else: pols = [ pollist ] if ( len(pols) > 0 ): sel.set_polarisations(pols) try: #Apply the selection s.set_selection(sel) except Exception, instance: #print '***Error***',instance casalog.post( str(instance), priority = 'ERROR' ) return # flag mode if ( flagmode.lower() == 'restore' ): show_flag_history(s) maskflag=restore_flag(s) return maskflag elif (flagmode.lower() == 'flag'): flgmode=flagmode unflag=False elif (flagmode.lower() == 'unflag'): flgmode=flagmode unflag=True else: raise Exception, 'unexpected flagmode' scanns = s.getscannos() sn=list(scanns) nr=s.nrow() if clip: casalog.post("Number of spectra to be flagged: %d" % (nr) ) casalog.post("Applying clipping...") else: if (len(flagrow) == 0): #print "Number of scans to be flagged:", len(sn) #print "Number of spectra to be flagged:", nr casalog.post( "Number of spectra to be flagged: %d" % (nr) ) casalog.post( "Applying channel flagging..." ) else: casalog.post( "Number of rows to be flagged: %d" % (len(flagrow)) ) casalog.post( "Applying row flagging..." ) dthres = uthres = None if isinstance(clipminmax, list): if (len(clipminmax) == 2): dthres = min(clipminmax) uthres = max(clipminmax) #for row in range(ns): if ( abs(plotlevel) > 0 ): if (len(flagrow) == 0) and (not clip): #channel flag if (len(maskflag) == 0): raise Exception, 'maskflag is undefined' masks = s.create_mask(maskflag) #sc=s.copy() # Plot final spectrum if nr < 3: nrow=1 ncol=nr elif nr < 5: nrow=2 ncol=2 elif nr < 7: nrow=2 ncol=3 elif nr < 10: nrow=3 ncol=3 else: nrow=4 ncol=4 casalog.post( "nrow,ncol= %d,%d" % (nrow, ncol) ) if nr >16: casalog.post( "Only first 16 spectra is plotted.", priority = 'WARN' ) if not myp or myp.is_dead: if sd.rcParams['plotter.gui']: from asap.asaplotgui import asaplotgui as asaplot else: from asap.asaplot import asaplot myp = asaplot() myp.hold() myp.clear() myp.set_panels(nrow,ncol) colours = ["green","red","#dddddd","#777777"] from matplotlib.numerix import ma, array, logical_not, logical_and if nr <17: rowlist=range(nr) else: rowlist=range(16) for row in rowlist: myp.subplot(row) myp.palette(0,colours) if row==rowlist[0]: myp.set_line(label='Spectrum') else: myp.set_line(label='spec') x = s._getabcissa(row) y = s._getspectrum(row) if s._getflagrow(row): oldmskarr = array([False]*(s.nchan())) else: oldmskarr = array(s._getmask(row)) if clip: if (uthres != None) and (dthres != None) and (uthres > dthres): masks = array(s._getclipmask(row, uthres, dthres, clipoutside, unflag)) else: masks = [False]*(s.nchan()) elif (len(flagrow) > 0): found = False for i in range(0, len(flagrow)): if (row == flagrow[i]): found = True break masks = [found and not(unflag)]*(s.nchan()) #marr = array(masks) marr = logical_not(array(masks)) allmsk = logical_and(marr,oldmskarr) ym = ma.masked_array(y,mask=logical_not(allmsk)) myp.plot(x,ym) myp.palette(2) if row==rowlist[0]: myp.set_line(label='previously flagged data') else: myp.set_line(label='prev') #oldmskarr = logical_not(oldmskarr) ym = ma.masked_array(y,mask=oldmskarr) myp.plot(x,ym) myp.palette(1) if row==rowlist[0]: myp.set_line(label='current flag masks') else: myp.set_line(label='flag') ym = ma.masked_array(y,mask=marr) myp.plot(x,ym) xlim=[min(x),max(x)] myp.axes.set_xlim(xlim) myp.release() #Apply flag ans=raw_input("Apply %s (y/n)?: " % flgmode) if ans.upper() == 'Y': if (clip): if (uthres != None) and (dthres != None) and (uthres > dthres): s.clip(uthres, dthres, clipoutside, unflag) else: if (len(flagrow) == 0): s.flag(masks,unflag) else: s.flag_row(flagrow, unflag) params={} if ( vars()['pols'] == [] ): params['pols']=list(s.getpolnos()) else: params['pols']=vars()['pols'] if ( vars()['ifs'] == [] ): params['ifs']=list(s.getifnos()) else: params['ifs']=vars()['ifs'] if ( vars()['scans'] == [] ): params['scans']=list(s.getscannos()) else: params['scans']=vars()['scans'] params['mode']=vars()['flagmode'] params['maskflag']=vars()['maskflag'] #print "input parameters:\n", params s._add_history( "sdflag", params ) else: return #Plot the result #print "Showing only the first spectrum..." casalog.post( "Showing only the first spectrum..." ) row=rowlist[0] if not myp or myp.is_dead: if sd.rcParams['plotter.gui']: from asap.asaplotgui import asaplotgui as asaplot else: from asap.asaplot import asaplot myp = asaplot() myp.hold() myp.clear() myp.palette(0,colours) myp.set_panels() myp.set_line(label="Spectrum after %s" % flgmode+'ging') x = s._getabcissa(row) y = s._getspectrum(row) allmskarr=array(s._getmask(row)) #y = ma.masked_array(y,mask=marr) ym = ma.masked_array(y,mask=logical_not(allmskarr)) myp.plot(x,ym) myp.palette(2) myp.set_line(label="Flagged") #allmsk = logical_and(marr,oldmskarr) #y = ma.masked_array(y,mask=logical_not(allmsk)) ym = ma.masked_array(y,mask=allmskarr) myp.plot(x,ym) myp.axes.set_xlim(xlim) if ( plotlevel < 0 ): # Hardcopy - currently no way w/o screen display first pltfile=project+'_flag.eps' myp.save(pltfile) myp.release() else: ans=raw_input("Apply %s (y/n)?: " % flgmode) if ans.upper() == 'Y': if clip: if (uthres != None) and (dthres != None) and (uthres > dthres): s.clip(uthres, dthres, clipoutside, unflag) else: if (len(flagrow) == 0): s.flag(masks,unflag) else: s.flag_row(flagrow, unflag) params={} if ( vars()['pols'] == [] ): params['pols']=list(s.getpolnos()) else: params['pols']=vars()['pols'] if ( vars()['ifs'] == [] ): params['ifs']=list(s.getifnos()) else: params['ifs']=vars()['ifs'] if ( vars()['scans'] == [] ): params['scans']=list(s.getscannos()) else: params['scans']=vars()['scans'] params['mode']=vars()['flagmode'] params['maskflag']=vars()['maskflag'] #print "input parameters:\n", params s._add_history( "sdflag", params ) else: return sel.reset() s.set_selection(sel) # Now save the spectrum and write out final ms if ( (outform == 'ASCII') or (outform == 'ascii') ): outform = 'ASCII' spefile = project + '_' elif ( (outform == 'ASAP') or (outform == 'asap') ): outform = 'ASAP' spefile = project elif ( (outform == 'SDFITS') or (outform == 'sdfits') ): outform = 'SDFITS' spefile = project elif ( (outform == 'MS') or (outform == 'ms') or (outform == 'MS2') or (outform == 'ms2') ): outform = 'MS2' spefile = project else: outform = 'ASAP' spefile = project if overwrite and os.path.exists(outfilename): os.system('rm -rf %s' % outfilename) #put back original spectral unit s.set_unit(unit_in) s.save(spefile,outform,overwrite) if outform!='ASCII': #print "Wrote output "+outform+" file "+spefile casalog.post( "Wrote output "+outform+" file "+spefile ) #if outfile!= 'none' or ( outfile=='none' and outform!='ASAP') : # print "Wrote output "+outform+" file "+spefile #else: # print "Updated current flag information to "+spefile # Clean up scantable del s # DONE except Exception, instance: #print '***Error***',instance casalog.post( str(instance), priority = 'ERROR' ) finally: casalog.post('') | 15d1302fcdfcb9607e06e0aa60d7ec0f82aa04c9 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/2098/15d1302fcdfcb9607e06e0aa60d7ec0f82aa04c9/task_sdflag.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
31789,
22532,
12,
6427,
768,
16,
392,
2253,
6582,
16,
4135,
1098,
16,
652,
16,
309,
1098,
16,
2952,
1098,
16,
3066,
6420,
16,
2982,
492,
16,
6807,
16,
6807,
1154,
1896,
16,
6807,
659,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
31789,
22532,
12,
6427,
768,
16,
392,
2253,
6582,
16,
4135,
1098,
16,
652,
16,
309,
1098,
16,
2952,
1098,
16,
3066,
6420,
16,
2982,
492,
16,
6807,
16,
6807,
1154,
1896,
16,
6807,
659,
... | |
if Config != "NT386": | if Config != "NT386" and not Target.endswith("_NT"): | def _CopyCompiler(From, To): CreateDirectory(To) from_cm3 = os.path.join(From, "cm3") from_cm3exe = os.path.join(From, "cm3.exe") from_cm3cg = os.path.join(From, "cm3cg") from_cm3cgexe = os.path.join(From, "cm3cg.exe") if (Config != "NT386" and not FileExists(from_cm3) and not FileExists(from_cm3exe) and not FileExists(from_cm3cg) and not FileExists(from_cm3cgexe)): FatalError("none of " + from_cm3 + ", " + from_cm3exe + ", " + from_cm3cg + ", " + from_cm3cgexe + " exist") # # check .exe first to avoid being fooled by Cygwin # if FileExists(from_cm3exe): from_cm3 = from_cm3exe elif FileExists(from_cm3): pass else: from_cm3 = None from_cm3exe = None if from_cm3: # # delete .exe first to avoid being fooled by Cygwin # DeleteFile(os.path.join(To, "cm3.exe")) DeleteFile(os.path.join(To, "cm3")) CopyFile(from_cm3, To) or FatalError("3") CopyFileIfExist(os.path.join(From, "cm3.pdb"), To) or FatalError("5") if Config != "NT386": if FileExists(from_cm3cgexe): from_cm3cg = from_cm3cgexe elif FileExists(from_cm3cg): pass else: from_cm3cg = None from_cm3cgexe = None if from_cm3cg: # # delete .exe first to avoid being fooled by Cygwin # DeleteFile(os.path.join(To, "cm3cg.exe")) DeleteFile(os.path.join(To, "cm3cg")) CopyFile(from_cm3cg, To) or FatalError("4") return True | 14e6facff06b2fc91279af79db1d8f8435565022 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9328/14e6facff06b2fc91279af79db1d8f8435565022/pylib.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2951,
9213,
12,
1265,
16,
2974,
4672,
225,
1788,
2853,
12,
774,
13,
225,
628,
67,
7670,
23,
273,
1140,
18,
803,
18,
5701,
12,
1265,
16,
315,
7670,
23,
7923,
628,
67,
7670,
23,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2951,
9213,
12,
1265,
16,
2974,
4672,
225,
1788,
2853,
12,
774,
13,
225,
628,
67,
7670,
23,
273,
1140,
18,
803,
18,
5701,
12,
1265,
16,
315,
7670,
23,
7923,
628,
67,
7670,
23,
... |
return self._decomp.Decompress(len(data) / self._nchannels, | return self._decomp.Decompress(len(data) // self._nchannels, | def _decomp_data(self, data): import cl dummy = self._decomp.SetParam(cl.FRAME_BUFFER_SIZE, len(data) * 2) return self._decomp.Decompress(len(data) / self._nchannels, data) | bca7b4894a1e338b53bc65b01a09484e5f2da318 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/8546/bca7b4894a1e338b53bc65b01a09484e5f2da318/aifc.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
323,
2919,
67,
892,
12,
2890,
16,
501,
4672,
1930,
927,
9609,
273,
365,
6315,
323,
2919,
18,
694,
786,
12,
830,
18,
15072,
67,
11302,
67,
4574,
16,
562,
12,
892,
13,
380,
576,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
323,
2919,
67,
892,
12,
2890,
16,
501,
4672,
1930,
927,
9609,
273,
365,
6315,
323,
2919,
18,
694,
786,
12,
830,
18,
15072,
67,
11302,
67,
4574,
16,
562,
12,
892,
13,
380,
576,
... |
elif flags & Client._FLAG_PICKLE: | elif flags & Client._FLAG_SERIALIZED: | def _recv_value(self, server, flags, rlen): rlen += 2 # include \r\n buf = server.recv(rlen) if len(buf) != rlen: raise _Error("received %d bytes when expecting %d" % (len(buf), rlen)) | e5ed3a4ffff0eaa456d776e74df6c63995141ed2 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/11935/e5ed3a4ffff0eaa456d776e74df6c63995141ed2/memcache.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
18334,
67,
1132,
12,
2890,
16,
1438,
16,
2943,
16,
436,
1897,
4672,
436,
1897,
1011,
576,
468,
2341,
521,
86,
64,
82,
1681,
273,
1438,
18,
18334,
12,
1321,
275,
13,
309,
562,
12... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
18334,
67,
1132,
12,
2890,
16,
1438,
16,
2943,
16,
436,
1897,
4672,
436,
1897,
1011,
576,
468,
2341,
521,
86,
64,
82,
1681,
273,
1438,
18,
18334,
12,
1321,
275,
13,
309,
562,
12... |
uri = req.uri if uri == '/': path = [''] else: uri = re_slashes.sub('/', uri) path = uri[1:].split('/') return root._traverse(req, path) except TraversalError: | if re_bibdoc_uri.match(req.uri): allowed_methods = ("GET", "POST", "HEAD", "OPTIONS") else: allowed_methods = ("GET", "POST", "OPTIONS") req.allow_methods(allowed_methods) if req.method not in allowed_methods: raise apache.SERVER_RETURN, apache.HTTP_METHOD_NOT_ALLOWED if req.method == 'OPTIONS': req.headers_out['Allow'] = ', '.join(allowed_methods) raise apache.SERVER_RETURN, apache.OK os.environ["HTTP_USER_AGENT"] = req.headers_in.get('User-Agent', '') guest_p = isGuestUser(getUid(req)) if guest_p: cache_control = "public" else: cache_control = "private" req.headers_out['Cache-Control'] = cache_control try: uri = req.uri if uri == '/': path = [''] else: uri = re_slashes.sub('/', uri) path = uri[1:].split('/') return root._traverse(req, path) except TraversalError: return apache.HTTP_NOT_FOUND except apache.SERVER_RETURN: raise except IOError, exc: if 'Write failed, client closed connection' in "%s" % exc: raise else: register_exception(req=req, alert_admin=True) except Exception: register_exception(req=req, alert_admin=True) raise | def _handler(req): """ This handler is invoked by mod_python with the apache request.""" | 60e09ca1bba94cdd1e4efbf4a45ba163ebdd84f0 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/2111/60e09ca1bba94cdd1e4efbf4a45ba163ebdd84f0/webinterface_handler.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
4176,
12,
3658,
4672,
3536,
1220,
1838,
353,
8187,
635,
681,
67,
8103,
598,
326,
12291,
590,
12123,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
4176,
12,
3658,
4672,
3536,
1220,
1838,
353,
8187,
635,
681,
67,
8103,
598,
326,
12291,
590,
12123,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
"""SMTP 'verify' command. Checks for address validity.""" | """SMTP 'verify' command -- checks for address validity.""" | def expn(self, address): """SMTP 'verify' command. Checks for address validity.""" self.putcmd("expn", quoteaddr(address)) return self.getreply() | 58a0b3fcac35fd85d6794fbcb62a4f4925cb00da /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12029/58a0b3fcac35fd85d6794fbcb62a4f4925cb00da/smtplib.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1329,
82,
12,
2890,
16,
1758,
4672,
3536,
55,
14636,
296,
8705,
11,
1296,
1493,
4271,
364,
1758,
13800,
12123,
365,
18,
458,
4172,
2932,
2749,
82,
3113,
4914,
684,
3069,
12,
2867,
3719,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1329,
82,
12,
2890,
16,
1758,
4672,
3536,
55,
14636,
296,
8705,
11,
1296,
1493,
4271,
364,
1758,
13800,
12123,
365,
18,
458,
4172,
2932,
2749,
82,
3113,
4914,
684,
3069,
12,
2867,
3719,
... |
return self.cache.setdefault(args,self.func(*args)) | if not args in self.cache: self.cache[args] = self.func(*args) return self.cache[args] | def __call__(self, *args): try: return self.cache.setdefault(args,self.func(*args)) except TypeError: # uncachable -- for instance, passing a list as an argument. # Better to not cache than to blow up entirely. return self.func(*args) | bd341109a466cb120a2ee3494d6276ae7cea8cd7 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/10307/bd341109a466cb120a2ee3494d6276ae7cea8cd7/tracer.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
1991,
972,
12,
2890,
16,
380,
1968,
4672,
775,
30,
309,
486,
833,
316,
365,
18,
2493,
30,
365,
18,
2493,
63,
1968,
65,
273,
365,
18,
644,
30857,
1968,
13,
327,
365,
18,
2493,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
1991,
972,
12,
2890,
16,
380,
1968,
4672,
775,
30,
309,
486,
833,
316,
365,
18,
2493,
30,
365,
18,
2493,
63,
1968,
65,
273,
365,
18,
644,
30857,
1968,
13,
327,
365,
18,
2493,
... |
userDefaultsFile = resolvePath("~/.grid-control.conf", check = False) | userDefaultsFile = utils.resolvePath("~/.grid-control.conf", check = False) | def parseFileInt(fn): try: parser.readfp(open(fn, 'r')) except IOError: raise ConfigError("Error while reading configuration file '%s'!" % fn) except cp.Error: print "Configuration file `%s' contains an error:" % fn raise | 545a17b18e0b5208d69b9ae3f55d6501cf3e90a0 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/8443/545a17b18e0b5208d69b9ae3f55d6501cf3e90a0/config.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
31986,
1702,
12,
4293,
4672,
775,
30,
2082,
18,
896,
7944,
12,
3190,
12,
4293,
16,
296,
86,
26112,
1335,
8340,
30,
1002,
1903,
668,
2932,
668,
1323,
6453,
1664,
585,
1995,
87,
11,
4442... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
31986,
1702,
12,
4293,
4672,
775,
30,
2082,
18,
896,
7944,
12,
3190,
12,
4293,
16,
296,
86,
26112,
1335,
8340,
30,
1002,
1903,
668,
2932,
668,
1323,
6453,
1664,
585,
1995,
87,
11,
4442... |
parser.set_defaults(dpi=96,quiet=False,sources={},animate=None,output=None,expressions=[],lastsource=None,id=()) | parser.set_defaults(dpi=96,quiet=False,sources={},animate=None,output=None,expressions=[],lastsource=None,id=[]) | def newexpression(option, opt_str, value, parser): if parser.values.lastsource is None: raise optparse.OptionValueError('%s must be preceded by a -s/--source option.' % option) if not isinstance(value,tuple): value = (None,value) parser.values.expressions.append((value[0],parser.values.lastsource,value[1])) | 4e2feb7d61fc667b6752e77b007b7f6c16ba25cb /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/383/4e2feb7d61fc667b6752e77b007b7f6c16ba25cb/multiplot.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
394,
8692,
12,
3482,
16,
2153,
67,
701,
16,
460,
16,
2082,
4672,
309,
2082,
18,
2372,
18,
2722,
3168,
353,
599,
30,
1002,
2153,
2670,
18,
13366,
668,
29909,
87,
1297,
506,
29953,
635,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
394,
8692,
12,
3482,
16,
2153,
67,
701,
16,
460,
16,
2082,
4672,
309,
2082,
18,
2372,
18,
2722,
3168,
353,
599,
30,
1002,
2153,
2670,
18,
13366,
668,
29909,
87,
1297,
506,
29953,
635,
... |
if plugin.closeWaitActive : | if plugin.closeWaitActive and plugin.eventHandlingIsAlive : | def run(self) : if self.watchDogTime == 0 : return False plugin = self.plugin CoInitialize() | 73adb38badb3754f4a03bf320c9c362e091da22a /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/8076/73adb38badb3754f4a03bf320c9c362e091da22a/__init__.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
12,
2890,
13,
294,
309,
365,
18,
7585,
40,
717,
950,
422,
374,
294,
327,
1083,
1909,
273,
365,
18,
4094,
7695,
7520,
1435,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1086,
12,
2890,
13,
294,
309,
365,
18,
7585,
40,
717,
950,
422,
374,
294,
327,
1083,
1909,
273,
365,
18,
4094,
7695,
7520,
1435,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100... |
label1 = gtk.Label("\t\tRecording is finished.\nrecordMyDesktop has exited with status %d\nDescription:%s\n"%(status,self.exit_status[int(status)])) | label1 = gtk.Label("\t\t"+_("Recording is finished.\nrecordMyDesktop has exited with status")+ ": %d\n"%(status)+_("Description")+":%s\n"%(self.exit_status[int(status)])) | def __exit_status_dialog(self,status): dialog = gtk.Dialog(title=None, parent=None, flags=0, buttons=None) label1=None try: label1 = gtk.Label("\t\tRecording is finished.\nrecordMyDesktop has exited with status %d\nDescription:%s\n"%(status,self.exit_status[int(status)])) except: label1 = gtk.Label("\t\tRecording is finished.\nrecordMyDesktop has exited with uknown\nerror code: %d \n"%(status)) dialog.vbox.pack_start(label1, True, True, 0) label1.show() button = gtk.Button(label=None, stock=gtk.STOCK_OK) dialog.vbox.pack_start(button, True, True, 0) def __destroy(button): dialog.destroy() button.connect("clicked", __destroy) button.show() dialog.set_size_request(300,128) dialog.show() | 575d85d8c50e1f5700bfbc519eaa7fe92903e36f /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/9403/575d85d8c50e1f5700bfbc519eaa7fe92903e36f/rmdTrayIcon.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
8593,
67,
2327,
67,
12730,
12,
2890,
16,
2327,
4672,
6176,
273,
22718,
18,
6353,
12,
2649,
33,
7036,
16,
982,
33,
7036,
16,
2943,
33,
20,
16,
9502,
33,
7036,
13,
1433,
21,
33,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
8593,
67,
2327,
67,
12730,
12,
2890,
16,
2327,
4672,
6176,
273,
22718,
18,
6353,
12,
2649,
33,
7036,
16,
982,
33,
7036,
16,
2943,
33,
20,
16,
9502,
33,
7036,
13,
1433,
21,
33,
... |
self.assertEqual(hash(desc.summary), 4980424973015496725) | self.assertEqual(hash(desc.summary), 588429333) | def testChangeset(self): """Testing PerforceTool.get_changeset""" | 99d9ebb5cebb8bbb44232b44aafe778f866d1580 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/1600/99d9ebb5cebb8bbb44232b44aafe778f866d1580/tests.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
29743,
12,
2890,
4672,
3536,
22218,
5722,
5734,
6364,
18,
588,
67,
6329,
278,
8395,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
29743,
12,
2890,
4672,
3536,
22218,
5722,
5734,
6364,
18,
588,
67,
6329,
278,
8395,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,... |
value &= 0xFFFFFFFF | value &= 0xFFFFFFFF | def SetRegValue(value, name): """ Set register value @param name: the register name @param value: new register value @note: The debugger should be running It is not necessary to use this function to set register values. A register name in the left side of an assignment will do too. """ rv = idaapi.regval_t() if type(value)==types.StringType: value = int(value) elif type(value)!=types.IntType: print "SetRegValue: value must be integer!" return BADADDR if value<0: #ival_set cannot handle negative numbers value &= 0xFFFFFFFF rv.ival = value return idaapi.set_reg_val(name, rv) | 14fac67483b08dd1be32c0afae368de44b30373e /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/6984/14fac67483b08dd1be32c0afae368de44b30373e/idc.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1000,
1617,
620,
12,
1132,
16,
508,
4672,
3536,
1000,
1744,
460,
225,
632,
891,
508,
30,
326,
1744,
508,
632,
891,
460,
30,
394,
1744,
460,
225,
632,
7652,
30,
1021,
19977,
1410,
506,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1000,
1617,
620,
12,
1132,
16,
508,
4672,
3536,
1000,
1744,
460,
225,
632,
891,
508,
30,
326,
1744,
508,
632,
891,
460,
30,
394,
1744,
460,
225,
632,
7652,
30,
1021,
19977,
1410,
506,
... |
height as the master) bordermode="inside" or "outside" - whether to take border width of master widget into account """ for k in ['in_']: if k in kw: kw[k[:-1]] = kw[k] del kw[k] | height as the master) bordermode="inside" or "outside" - whether to take border width of master widget into account """ | def place_configure(self, cnf={}, **kw): """Place a widget in the parent widget. Use as options: in=master - master relative to which the widget is placed. x=amount - locate anchor of this widget at position x of master y=amount - locate anchor of this widget at position y of master relx=amount - locate anchor of this widget between 0.0 and 1.0 relative to width of master (1.0 is right edge) rely=amount - locate anchor of this widget between 0.0 and 1.0 relative to height of master (1.0 is bottom edge) anchor=NSEW (or subset) - position anchor according to given direction width=amount - width of this widget in pixel height=amount - height of this widget in pixel relwidth=amount - width of this widget between 0.0 and 1.0 relative to width of master (1.0 is the same width as the master) relheight=amount - height of this widget between 0.0 and 1.0 relative to height of master (1.0 is the same height as the master) bordermode="inside" or "outside" - whether to take border width of master widget into account """ for k in ['in_']: if k in kw: kw[k[:-1]] = kw[k] del kw[k] self.tk.call( ('place', 'configure', self._w) + self._options(cnf, kw)) | 27856c1fab2add060e43569aba04994d8d21c029 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/8125/27856c1fab2add060e43569aba04994d8d21c029/__init__.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3166,
67,
14895,
12,
2890,
16,
23118,
28793,
2826,
9987,
4672,
3536,
6029,
279,
3604,
316,
326,
982,
3604,
18,
2672,
487,
702,
30,
316,
33,
7525,
300,
4171,
3632,
358,
1492,
326,
3604,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3166,
67,
14895,
12,
2890,
16,
23118,
28793,
2826,
9987,
4672,
3536,
6029,
279,
3604,
316,
326,
982,
3604,
18,
2672,
487,
702,
30,
316,
33,
7525,
300,
4171,
3632,
358,
1492,
326,
3604,
... |
assert res == (a == "hello") | assert_equal(res,(a == "hello")) | def check_char(self,level=5): a = "hello" res = inline_tools.inline('return_val = (a == "hello");',['a']) assert res == (a == "hello") | 267b08d9c3e7613c126729a179f4503c91625e9c /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/7416/267b08d9c3e7613c126729a179f4503c91625e9c/test_scxx_object.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
866,
67,
3001,
12,
2890,
16,
2815,
33,
25,
4672,
279,
273,
315,
23711,
6,
400,
273,
6370,
67,
6642,
18,
10047,
2668,
2463,
67,
1125,
273,
261,
69,
422,
315,
23711,
8863,
2187,
3292,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
866,
67,
3001,
12,
2890,
16,
2815,
33,
25,
4672,
279,
273,
315,
23711,
6,
400,
273,
6370,
67,
6642,
18,
10047,
2668,
2463,
67,
1125,
273,
261,
69,
422,
315,
23711,
8863,
2187,
3292,
... |
Let `M=` ``self.lattice().dual()`` be the lattice dual to the | Let `M=` ``self.dual_lattice()`` be the lattice dual to the | def orthogonal_sublattice(self, *args, **kwds): r""" The sublattice (in the dual lattice) orthogonal to the sublattice spanned by the cone. Let `M=` ``self.lattice().dual()`` be the lattice dual to the ambient lattice of the given cone `\sigma`. Then, in the notation of [Fulton]_, this method returns the sublattice | 80a9bb2c425ebeec4f69e966b1ed7b66d21d86f7 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/9890/80a9bb2c425ebeec4f69e966b1ed7b66d21d86f7/cone.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
16208,
28396,
67,
87,
440,
24966,
12,
2890,
16,
380,
1968,
16,
2826,
25577,
4672,
436,
8395,
1021,
17809,
24966,
261,
267,
326,
24557,
16690,
13,
16208,
28396,
358,
326,
17809,
24966,
1694... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
16208,
28396,
67,
87,
440,
24966,
12,
2890,
16,
380,
1968,
16,
2826,
25577,
4672,
436,
8395,
1021,
17809,
24966,
261,
267,
326,
24557,
16690,
13,
16208,
28396,
358,
326,
17809,
24966,
1694... |
(?P<title>[^[]+)? | (?P<title>[^[]+)? | def as_attachment(self): """Return this message as an attachment.""" return (self.getname(), self.gettype(), self.getbody()) | 5007a594beb2621dd1ac05e40b59df6d65936de3 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/1906/5007a594beb2621dd1ac05e40b59df6d65936de3/mailgw.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
487,
67,
11461,
12,
2890,
4672,
3536,
990,
333,
883,
487,
392,
6042,
12123,
327,
261,
2890,
18,
588,
529,
9334,
365,
18,
588,
723,
9334,
365,
18,
588,
3432,
10756,
2,
0,
0,
0,
0,
0... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
487,
67,
11461,
12,
2890,
4672,
3536,
990,
333,
883,
487,
392,
6042,
12123,
327,
261,
2890,
18,
588,
529,
9334,
365,
18,
588,
723,
9334,
365,
18,
588,
3432,
10756,
2,
-100,
-100,
-100,... |
print >> DEBUGSTREAM, 'we got some refusals' | print >> DEBUGSTREAM, 'we got some refusals:', refused | def process_message(self, peer, mailfrom, rcpttos, data): lines = data.split('\n') # Look for the last header i = 0 for line in lines: if not line: break i += 1 lines.insert(i, 'X-Peer: %s' % peer[0]) data = NEWLINE.join(lines) refused = self._deliver(mailfrom, rcpttos, data) # TBD: what to do with refused addresses? print >> DEBUGSTREAM, 'we got some refusals' | be3e352be45b63d1168727d245f93e7eecbfdba4 /local1/tlutelli/issta_data/temp/all_python//python/2006_temp/2006/12029/be3e352be45b63d1168727d245f93e7eecbfdba4/smtpd.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1207,
67,
2150,
12,
2890,
16,
4261,
16,
4791,
2080,
16,
4519,
337,
14627,
16,
501,
4672,
2362,
273,
501,
18,
4939,
2668,
64,
82,
6134,
468,
10176,
364,
326,
1142,
1446,
277,
273,
374,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1207,
67,
2150,
12,
2890,
16,
4261,
16,
4791,
2080,
16,
4519,
337,
14627,
16,
501,
4672,
2362,
273,
501,
18,
4939,
2668,
64,
82,
6134,
468,
10176,
364,
326,
1142,
1446,
277,
273,
374,
... |
self.tag = tag | self._revspec = None self.revspec = (tag, revspec) def get_revspec(self): return self._revspec def set_revspec(self, (tag, revspec)): if revspec: self._revspec = ['-r%s' % revspec] elif tag: logging.info('tag ' + _('attribute is deprecated. Use revspec instead.')) self._revspec = ['-rtag:%s' % tag] elif self.config.sticky_date: self._revspec = ['-rdate:%s' % self.config.sticky_date] else: self._revspec = [] revspec = property(get_revspec, set_revspec) | def __init__(self, repository, module_href, checkoutdir, tag): Branch.__init__(self, repository, module_href, checkoutdir) self.tag = tag | 6bb6439fa76a3dadc9bdb5c0828322c6c7dd6015 /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/4596/6bb6439fa76a3dadc9bdb5c0828322c6c7dd6015/bzr.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
3352,
16,
1605,
67,
7547,
16,
13926,
1214,
16,
1047,
4672,
15449,
16186,
2738,
972,
12,
2890,
16,
3352,
16,
1605,
67,
7547,
16,
13926,
1214,
13,
365,
18,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
3352,
16,
1605,
67,
7547,
16,
13926,
1214,
16,
1047,
4672,
15449,
16186,
2738,
972,
12,
2890,
16,
3352,
16,
1605,
67,
7547,
16,
13926,
1214,
13,
365,
18,... |
print "Couldn't find msvc 9.0 compiler installation directory!" | print "Couldn't find msvc compiler installation directory!" | def get_msvs8_install_dirs(version = None, vs8suite = None): """ Get installed locations for various msvc-related products, like the .NET SDK and the Platform SDK. """ if not SCons.Util.can_read_reg: return {} if not version: versions = get_visualstudio_versions() if versions: version = versions[0] #use highest version by default else: return {} rv = {} version_num, suite = msvs_parse_version(version) K = 'Software\\Microsoft\\VisualStudio\\' + str(version_num) if (version_num >= 8.0): if vs8suite == None: # We've been given no guidance about which Visual Studio 8 # suite to use, so attempt to autodetect. suites = get_visualstudio8_suites(version) if suites: vs8suite = suites[0] rv['VS8SUITE'] = vs8suite if vs8suite == 'EXPRESS' or not suites: K = 'Software\\Microsoft\\VCExpress\\' + str(version_num) # vc++ install dir if (version_num < 7.0): key = K + r'\Setup\Microsoft Visual C++\ProductDir' else: key = K + r'\Setup\VC\ProductDir' try: (rv['VCINSTALLDIR'], t) = SCons.Util.RegGetValue(SCons.Util.HKEY_LOCAL_MACHINE, key) except SCons.Util.RegError: print "Couldn't find msvc 9.0 compiler installation directory!" pass # visual studio install dir if (version_num < 7.0): try: (rv['VSINSTALLDIR'], t) = SCons.Util.RegGetValue(SCons.Util.HKEY_LOCAL_MACHINE, K + r'\Setup\Microsoft Visual Studio\ProductDir') except SCons.Util.RegError: pass if not rv.has_key('VSINSTALLDIR') or not rv['VSINSTALLDIR']: if rv.has_key('VCINSTALLDIR') and rv['VCINSTALLDIR']: rv['VSINSTALLDIR'] = os.path.dirname(rv['VCINSTALLDIR']) else: rv['VSINSTALLDIR'] = os.path.join(SCons.Platform.win32.get_program_files_dir(),'Microsoft Visual Studio') else: try: (rv['VSINSTALLDIR'], t) = SCons.Util.RegGetValue(SCons.Util.HKEY_LOCAL_MACHINE, K + r'\Setup\VS\ProductDir') except SCons.Util.RegError: pass # .NET framework install dir try: (rv['FRAMEWORKDIR'], t) = SCons.Util.RegGetValue(SCons.Util.HKEY_LOCAL_MACHINE, r'Software\Microsoft\.NETFramework\InstallRoot') except SCons.Util.RegError: pass if rv.has_key('FRAMEWORKDIR'): # try and enumerate the installed versions of the .NET framework. contents = os.listdir(rv['FRAMEWORKDIR']) l = re.compile('v[0-9]+.*') versions = [] for entry in contents: if l.match(entry): versions.append(entry) def versrt(a,b): # since version numbers aren't really floats... aa = a[1:] bb = b[1:] aal = aa.split('.') bbl = bb.split('.') c = int(bbl[0]) - int(aal[0]) if c == 0: c = int(bbl[1]) - int(aal[1]) if c == 0: c = int(bbl[2]) - int(aal[2]) return c versions.sort(versrt) rv['FRAMEWORKVERSIONS'] = versions # assume that the highest version is the latest version installed rv['FRAMEWORKVERSION'] = versions[0] # .NET framework SDK install dir try: if rv.has_key('FRAMEWORKVERSION') and rv['FRAMEWORKVERSION'][:4] == 'v1.1': key = r'Software\Microsoft\.NETFramework\sdkInstallRootv1.1' else: key = r'Software\Microsoft\.NETFramework\sdkInstallRoot' (rv['FRAMEWORKSDKDIR'], t) = SCons.Util.RegGetValue(SCons.Util.HKEY_LOCAL_MACHINE,key) except SCons.Util.RegError: pass # MS Platform SDK dir try: (rv['PLATFORMSDKDIR'], t) = SCons.Util.RegGetValue(SCons.Util.HKEY_LOCAL_MACHINE, r'Software\Microsoft\MicrosoftSDK\Directories\Install Dir') except SCons.Util.RegError: pass # check for existence if rv.has_key('PLATFORMSDKDIR'): if not os.path.exists(rv['PLATFORMSDKDIR']): del rv['PLATFORMSDKDIR'] # ok now look for something like this: # Software\Microsoft\MicrosoftSDK\InstalledSDKs\HKEY_CURRENT_USER\Software\Microsoft\MicrosoftSDK\InstalledSDKs\8F9E5EF3-A9A5-491B-A889-C58EFFECE8B3\Install Dir if not rv.has_key('PLATFORMSDKDIR'): try: loc = r'Software\Microsoft\MicrosoftSDK\InstalledSDKs' k = SCons.Util.RegOpenKeyEx(SCons.Util.HKEY_LOCAL_MACHINE,loc) i = 0 while 1: try: key = SCons.Util.RegEnumKey(k,i) try: (rv['PLATFORMSDKDIR'], t) = SCons.Util.RegGetValue(SCons.Util.HKEY_LOCAL_MACHINE, r'Software\Microsoft\MicrosoftSDK\InstalledSDKs\\' + key + r'\Install Dir') except SCons.Util.RegError: pass i = i + 1 except SCons.Util.RegError: break; except SCons.Util.RegError: pass if rv.has_key('PLATFORMSDKDIR'): # if we have a platform SDK, try and get some info on it. vers = {} try: loc = r'Software\Microsoft\MicrosoftSDK\InstalledSDKs' k = SCons.Util.RegOpenKeyEx(SCons.Util.HKEY_LOCAL_MACHINE,loc) i = 0 while 1: try: key = SCons.Util.RegEnumKey(k,i) sdk = SCons.Util.RegOpenKeyEx(k,key) j = 0 name = '' date = '' version = '' while 1: try: (vk,vv,t) = SCons.Util.RegEnumValue(sdk,j) if vk.lower() == 'keyword': name = vv if vk.lower() == 'propagation_date': date = vv if vk.lower() == 'version': version = vv j = j + 1 except SCons.Util.RegError: break if name: vers[name] = (date, version) i = i + 1 except SCons.Util.RegError: break rv['PLATFORMSDK_MODULES'] = vers except SCons.Util.RegError: pass # MS Platform SDKs dir for Visual Studio 2008 try: (rv['PLATFORMSDKSDIR'], t) = SCons.Util.RegGetValue(SCons.Util.HKEY_LOCAL_MACHINE, r'Software\Microsoft\Microsoft SDKs\Windows\CurrentInstallFolder') except SCons.Util.RegError: pass return rv | b0f2ea3adf6bc7a051d46942cf2c0e96c1cbe2bf /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/6357/b0f2ea3adf6bc7a051d46942cf2c0e96c1cbe2bf/msvc.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
67,
959,
6904,
28,
67,
5425,
67,
8291,
12,
1589,
273,
599,
16,
6195,
28,
30676,
273,
599,
4672,
3536,
968,
5876,
7838,
364,
11191,
4086,
4227,
17,
9243,
10406,
16,
3007,
326,
263,... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
67,
959,
6904,
28,
67,
5425,
67,
8291,
12,
1589,
273,
599,
16,
6195,
28,
30676,
273,
599,
4672,
3536,
968,
5876,
7838,
364,
11191,
4086,
4227,
17,
9243,
10406,
16,
3007,
326,
263,... |
def get(self, zstr, no_abbrev=False): | def get(self, zstr, allow_abbreviations=True): | def get(self, zstr, no_abbrev=False): state = { 'curr_alpha': 0, 'prev_alpha': 0, 'zscii': [], 'allow_abbreviations': no_abbrev, } | be9765d04303c4e0433ce98a269ff5a061cf5c2d /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/2052/be9765d04303c4e0433ce98a269ff5a061cf5c2d/zstring.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
12,
2890,
16,
998,
701,
16,
1699,
67,
378,
7379,
1012,
33,
5510,
4672,
919,
273,
288,
296,
17016,
67,
5429,
4278,
374,
16,
296,
10001,
67,
5429,
4278,
374,
16,
296,
94,
1017,
28... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
336,
12,
2890,
16,
998,
701,
16,
1699,
67,
378,
7379,
1012,
33,
5510,
4672,
919,
273,
288,
296,
17016,
67,
5429,
4278,
374,
16,
296,
10001,
67,
5429,
4278,
374,
16,
296,
94,
1017,
28... |
site.lookupFunctionName (p) | site.lookupFunctionName (p, inln) | def lookupFunctionNames (self): | 16c76ad61122d4de2a1adf56a24d755d2c6f00b3 /local1/tlutelli/issta_data/temp/all_python//python/2009_temp/2009/11834/16c76ad61122d4de2a1adf56a24d755d2c6f00b3/profile.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3689,
2083,
1557,
261,
2890,
4672,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
3689,
2083,
1557,
261,
2890,
4672,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
... |
if local: globals().update( locals() ) | def test_ExampleMaster(self): | 132c02d8264e957001d16b0a1459aa51f23e5916 /local1/tlutelli/issta_data/temp/all_python//python/2007_temp/2007/482/132c02d8264e957001d16b0a1459aa51f23e5916/ExampleMaster.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
10908,
7786,
12,
2890,
4672,
2,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0,
0... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1842,
67,
10908,
7786,
12,
2890,
4672,
2,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-100,
-1... | |
for user, authenticated, pages in cursor.execute(sql).fetchall(): | cursor.execute(sql) results = cursor.fetchall() if not results: return 'admin_user_notifications.html', {'wpages': notified_users, 'wikiurl':req.href.wiki()} for user, authenticated, pages in results: | def _do_users(self, req, cat, page, path_info): sql = "SELECT sid,authenticated,value " + \ "FROM session_attribute WHERE name = 'watched_pages';" db = self.env.get_db_cnx() cursor = db.cursor() notified_users = [] for user, authenticated, pages in cursor.execute(sql).fetchall(): attrs = {} attrs['sid'] = user attrs['authenticated'] = authenticated attrs['pages'] = pages.strip(',').split(',') notified_users.append(attrs) | a94e4a8b2e95643a9de35c9b9264b0fc945a3e86 /local1/tlutelli/issta_data/temp/all_python//python/2008_temp/2008/9414/a94e4a8b2e95643a9de35c9b9264b0fc945a3e86/admin.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2896,
67,
5577,
12,
2890,
16,
1111,
16,
6573,
16,
1363,
16,
589,
67,
1376,
4672,
1847,
273,
315,
4803,
7348,
16,
14454,
16,
1132,
315,
397,
521,
315,
11249,
1339,
67,
4589,
4852,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
389,
2896,
67,
5577,
12,
2890,
16,
1111,
16,
6573,
16,
1363,
16,
589,
67,
1376,
4672,
1847,
273,
315,
4803,
7348,
16,
14454,
16,
1132,
315,
397,
521,
315,
11249,
1339,
67,
4589,
4852,
... |
self._x, self._y = 0,0 self._sx, self._sy = 0,0 self._ax, self._ay = 0, 0 | self._x, self._y = 0,0 self._sx, self._sy = 0,0 self._ax, self._ay = 0, 0 self._tx, _ty = 0,0 | def __init__(self, name=None): self.name = name if name else "Unitled %s"%self.__name__ self._motion_queue = [] #actor's deltas for moving on the screen in the near-future self.action = None self.actions = {} self._alpha = 255 self._alpha_target = 255 self.font_speech = None self._x, self._y = 0,0 # place in scene self._sx, self._sy = 0,0 # stand point self._ax, self._ay = 0, 0 # anchor point self.speed = 10 #speed at which actor moves per frame self.inventory = {} self._scale = 1.0 self.scene = None self._walk_area = [0,0,0,0] self._solid_area = [0,0,0,0] self._clickable_area = [0,0,0,0] self._image = None self._tx, _ty = 0,0 #target for when moving self._rect = None self.game = None | 84c2c9bf76da2251a32e2f767e2d3a2d4532345f /local1/tlutelli/issta_data/temp/all_python//python/2010_temp/2010/14331/84c2c9bf76da2251a32e2f767e2d3a2d4532345f/__init__.py | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
508,
33,
7036,
4672,
365,
18,
529,
273,
508,
309,
508,
469,
315,
2802,
1259,
738,
87,
28385,
2890,
16186,
529,
972,
365,
6315,
81,
8240,
67,
4000,
273,
... | [
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1,
1... | [
1,
8585,
326,
22398,
316,
326,
981,
30,
1652,
1001,
2738,
972,
12,
2890,
16,
508,
33,
7036,
4672,
365,
18,
529,
273,
508,
309,
508,
469,
315,
2802,
1259,
738,
87,
28385,
2890,
16186,
529,
972,
365,
6315,
81,
8240,
67,
4000,
273,
... |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.