bugged stringlengths 4 228k | fixed stringlengths 0 96.3M | __index_level_0__ int64 0 481k |
|---|---|---|
def _log (fun, msg, args, tb=False): """ Log a message with given function and an optional traceback. @return: None """ fun(msg, *args) if tb: # note: get rid of last parts of the stack s = _stack_format(inspect.stack()[2:]) fun(s) | def _log (fun, msg, args, tb=False): """ Log a message with given function and an optional traceback. @return: None """ fun(msg, *args) if tb: # note: get rid of last parts of the stack fun(_stack_format(inspect.stack()[2:])) | 15,500 |
def debug (log, msg, *args, **kwargs): """ Log a debug message. return: None """ _log(logging.getLogger(log).debug, msg, args, tb=kwargs.get("tb")) | def debug (logname, msg, *args, **kwargs): """ Log a debug message. return: None """ _log(logging.getLogger(log).debug, msg, args, tb=kwargs.get("tb")) | 15,501 |
def debug (log, msg, *args, **kwargs): """ Log a debug message. return: None """ _log(logging.getLogger(log).debug, msg, args, tb=kwargs.get("tb")) | def debug (log, msg, *args, **kwargs): """ Log a debug message. return: None """ log = logging.getLogger(logname) if log.isEnabledFor(logging.DEBUG): _log(log.debug, msg, args, tb=kwargs.get("tb")) | 15,502 |
def info (log, msg, *args, **kwargs): """ Log an informational message. return: None """ _log(logging.getLogger(log).info, msg, args, tb=kwargs.get("tb")) | def info (logname, msg, *args, **kwargs): """ Log an informational message. return: None """ _log(logging.getLogger(log).info, msg, args, tb=kwargs.get("tb")) | 15,503 |
def info (log, msg, *args, **kwargs): """ Log an informational message. return: None """ _log(logging.getLogger(log).info, msg, args, tb=kwargs.get("tb")) | def info (log, msg, *args, **kwargs): """ Log an informational message. return: None """ log = logging.getLogger(logname) if log.isEnabledFor(logging.INFO): _log(log.info, msg, args, tb=kwargs.get("tb")) | 15,504 |
def warn (log, msg, *args, **kwargs): """ Log a warning. return: None """ _log(logging.getLogger(log).warn, msg, args, tb=kwargs.get("tb")) | def warn (logname, msg, *args, **kwargs): """ Log a warning. return: None """ _log(logging.getLogger(log).warn, msg, args, tb=kwargs.get("tb")) | 15,505 |
def warn (log, msg, *args, **kwargs): """ Log a warning. return: None """ _log(logging.getLogger(log).warn, msg, args, tb=kwargs.get("tb")) | def warn (log, msg, *args, **kwargs): """ Log a warning. return: None """ log = logging.getLogger(logname) if log.isEnabledFor(logging.WARN): _log(log.warn, msg, args, tb=kwargs.get("tb")) | 15,506 |
def error (log, msg, *args, **kwargs): """ Log an error. return: None """ _log(logging.getLogger(log).error, msg, args, tb=kwargs.get("tb")) | def error (logname, msg, *args, **kwargs): """ Log an error. return: None """ _log(logging.getLogger(log).error, msg, args, tb=kwargs.get("tb")) | 15,507 |
def error (log, msg, *args, **kwargs): """ Log an error. return: None """ _log(logging.getLogger(log).error, msg, args, tb=kwargs.get("tb")) | def error (log, msg, *args, **kwargs): """ Log an error. return: None """ log = logging.getLogger(logname) if log.isEnabledFor(logging.ERROR): _log(log.error, msg, args, tb=kwargs.get("tb")) | 15,508 |
def critical (log, msg, *args, **kwargs): """ Log a critical error. return: None """ _log(logging.getLogger(log).critical, msg, args, tb=kwargs.get("tb")) | def critical (logname, msg, *args, **kwargs): """ Log a critical error. return: None """ _log(logging.getLogger(log).critical, msg, args, tb=kwargs.get("tb")) | 15,509 |
def critical (log, msg, *args, **kwargs): """ Log a critical error. return: None """ _log(logging.getLogger(log).critical, msg, args, tb=kwargs.get("tb")) | def critical (log, msg, *args, **kwargs): """ Log a critical error. return: None """ log = logging.getLogger(logname) if log.isEnabledFor(logging.CRITICAL): _log(log.critical, msg, args, tb=kwargs.get("tb")) | 15,510 |
def exception (log, msg, *args, **kwargs): """ Log an exception. return: None """ _log(logging.getLogger(log).exception, msg, args, tb=kwargs.get("tb")) | def exception (logname, msg, *args, **kwargs): """ Log an exception. return: None """ _log(logging.getLogger(log).exception, msg, args, tb=kwargs.get("tb")) | 15,511 |
def exception (log, msg, *args, **kwargs): """ Log an exception. return: None """ _log(logging.getLogger(log).exception, msg, args, tb=kwargs.get("tb")) | def exception (log, msg, *args, **kwargs): """ Log an exception. return: None """ log = logging.getLogger(logname) if log.isEnabledFor(logging.EXCEPTION): _log(log.exception, msg, args, tb=kwargs.get("tb")) | 15,512 |
def init_dns_resolver_nt (): import winreg key = None try: key = winreg.key_handle(winreg.HKEY_LOCAL_MACHINE, r"SYSTEM\CurrentControlSet\Services\Tcpip\Parameters") except EnvironmentError: try: # for Windows ME key = winreg.key_handle(winreg.HKEY_LOCAL_MACHINE, r"SYSTEM\CurrentControlSet\Services\VxD\MSTCP") except En... | def init_dns_resolver_nt (): import winreg key = None try: key = winreg.key_handle(winreg.HKEY_LOCAL_MACHINE, r"SYSTEM\CurrentControlSet\Services\Tcpip\Parameters") except EnvironmentError: try: # for Windows ME key = winreg.key_handle(winreg.HKEY_LOCAL_MACHINE, r"SYSTEM\CurrentControlSet\Services\VxD\MSTCP") except En... | 15,513 |
def _fatalError (self, msg): """signal a fatal filter/parser error""" self._errorfun(msg, "fatalError") | def _fatalError (self, msg): """signal a fatal filter/parser error""" self._errorfun(msg, "fatalError") | 15,514 |
def _debugbuf (self): """print debugging information about data buffer status""" #self._debug(NIGHTMARE, "self.buf", `self.buf`) #self._debug(NIGHTMARE, "self.waitbuf", `self.waitbuf`) #self._debug(NIGHTMARE, "self.inbuf", `self.inbuf.getvalue()`) #self._debug(NIGHTMARE, "self.outbuf", `self.outbuf.getvalue()`) | def_debugbuf(self):"""printdebugginginformationaboutdatabufferstatus"""#self._debug(NIGHTMARE,"self.buf",`self.buf`)#self._debug(NIGHTMARE,"self.waitbuf",`self.waitbuf`)#self._debug(NIGHTMARE,"self.inbuf",`self.inbuf.getvalue()`)#self._debug(NIGHTMARE,"self.outbuf",`self.outbuf.getvalue()`) | 15,515 |
def feed (self, data): """feed some data to the parser""" if self.state=='parse': # look if we must replay something if self.waited: self.waited = 0 waitbuf, self.waitbuf = self.waitbuf, [] self.replay(waitbuf) if self.state!='parse': self._debug(ALWAYS, "self.inbuf", `self.inbuf.getvalue()`) return data = self.inbuf.g... | def feed (self, data): """feed some data to the parser""" if self.state=='parse': # look if we must replay something if self.waited: self.waited = 0 waitbuf, self.waitbuf = self.waitbuf, [] self.replay(waitbuf) if self.state!='parse': self._debug(ALWAYS, "self.inbuf", `self.inbuf.getvalue()`) return data = self.inbuf.g... | 15,516 |
def jsScript (self, script, ver, item): """execute given script with javascript version ver""" #self._debug(NIGHTMARE, "JS: jsScript", ver, `script`) assert self.state == 'parse' assert len(self.buf) >= 2 self.js_output = 0 self.js_env.attachListener(self) # start recursive html filter (used by jsProcessData) self.js_h... | def jsScript (self, script, ver, item): """execute given script with javascript version ver""" #self._debug(NIGHTMARE, "JS: jsScript", ver, `script`) assert self.state == 'parse' assert len(self.buf) >= 2 self.js_output = 0 self.js_env.attachListener(self) # start recursive html filter (used by jsProcessData) self.js_h... | 15,517 |
def _broken (): p = HtmlPrinter() p.feed("<img bo\\\nrder=0>") p.flush() | def _broken (): p = HtmlPrinter() p.debug(1) for c in """</td <td a="b" >""": p.feed(c) p.flush() | 15,518 |
def __init__ (self): Htmlparser.__init__(self) if wc.config['showerrors']: self.error = self._error self.warning = self._warning self.fatalError = self._fatalError self.outbuf = StringIO() self.buf = [] | def __init__ (self): HtmlParser.__init__(self) if wc.config['showerrors']: self.error = self._error self.warning = self._warning self.fatalError = self._fatalError self.outbuf = StringIO() self.buf = [] | 15,519 |
def filter (self, data, **attrs): if not attrs.has_key('buffer'): # we do not block this image # or we do not have enough buffer data return data buf = attrs['buffer'] buf.write(data) if buf.tell() > self.min_bufsize: if self.check_sizes(buf): # size is ok data = buf.getvalue() buf.close() del attrs['buffer'] return re... | def filter (self, data, **attrs): if attrs['buffer'].closed: # we do not block this image # or we do not have enough buffer data return data buf = attrs['buffer'] buf.write(data) if buf.tell() > self.min_bufsize: if self.check_sizes(buf): # size is ok data = buf.getvalue() buf.close() del attrs['buffer'] return return ... | 15,520 |
def filter (self, data, **attrs): if not attrs.has_key('buffer'): # we do not block this image # or we do not have enough buffer data return data buf = attrs['buffer'] buf.write(data) if buf.tell() > self.min_bufsize: if self.check_sizes(buf): # size is ok data = buf.getvalue() buf.close() del attrs['buffer'] return re... | def filter (self, data, **attrs): if not attrs.has_key('buffer'): # we do not block this image # or we do not have enough buffer data return data buf = attrs['buffer'] buf.write(data) if buf.tell() > self.min_bufsize: if self.check_sizes(buf): # size is ok data = buf.getvalue() buf.close() return data return '' | 15,521 |
def check_sizes (self, buf): try: img = Image.open(buf) for size, formats in sizes: if size==img.size: # size matches, look for format restriction if not formats: return False elif img.format.lower() in formats: return False except IOError: # XXX print error pass return True | def check_sizes (self, buf): try: img = Image.open(buf, 'r') for size, formats in sizes: if size==img.size: # size matches, look for format restriction if not formats: return False elif img.format.lower() in formats: return False except IOError: # XXX print error pass return True | 15,522 |
def testUrl (self): url = "http://server/..%5c..%5c..%5c..%5c..%5c..%5..%5c..%5ccskin.zip" nurl = wc.url.url_quote(wc.url.url_norm(url)) self.assertEquals(nurl, "http://server/cskin.zip") url = "http://groups.google.com/groups?hl=en&lr=&ie=UTF-8&threadm=3845B54D.E546F9BD%40monmouth.com&rnum=2&prev=/groups%3Fq%3Dlogitec... | def testUrlPathAttack (self): url = "http://server/..%5c..%5c..%5c..%5c..%5c..%5..%5c..%5ccskin.zip" nurl = wc.url.url_quote(wc.url.url_norm(url)) self.assertEquals(nurl, "http://server/cskin.zip") url = "http://groups.google.com/groups?hl=en&lr=&ie=UTF-8&threadm=3845B54D.E546F9BD%40monmouth.com&rnum=2&prev=/groups%3Fq... | 15,523 |
def testUrl (self): url = "http://server/..%5c..%5c..%5c..%5c..%5c..%5..%5c..%5ccskin.zip" nurl = wc.url.url_quote(wc.url.url_norm(url)) self.assertEquals(nurl, "http://server/cskin.zip") url = "http://groups.google.com/groups?hl=en&lr=&ie=UTF-8&threadm=3845B54D.E546F9BD%40monmouth.com&rnum=2&prev=/groups%3Fq%3Dlogitec... | def testUrl (self): url = "http://server/..%5c..%5c..%5c..%5c..%5c..%5..%5c..%5ccskin.zip" nurl = "http://server/cskin.zip" self.assertEquals(wc.url.url_quote(wc.url.url_norm(url)), nurl) def testUrlQuoting (self): url = "http://groups.google.com/groups?hl=en&lr=&ie=UTF-8&threadm=3845B54D.E546F9BD%40monmouth.com&rnum=... | 15,524 |
def testUrl (self): url = "http://server/..%5c..%5c..%5c..%5c..%5c..%5..%5c..%5ccskin.zip" nurl = wc.url.url_quote(wc.url.url_norm(url)) self.assertEquals(nurl, "http://server/cskin.zip") url = "http://groups.google.com/groups?hl=en&lr=&ie=UTF-8&threadm=3845B54D.E546F9BD%40monmouth.com&rnum=2&prev=/groups%3Fq%3Dlogitec... | def testUrl (self): url = "http://server/..%5c..%5c..%5c..%5c..%5c..%5..%5c..%5ccskin.zip" nurl = wc.url.url_quote(wc.url.url_norm(url)) self.assertEquals(nurl, "http://server/cskin.zip") url = "http://groups.google.com/groups?hl=en&lr=&ie=UTF-8&threadm=3845B54D.E546F9BD%40monmouth.com&rnum=2&prev=/groups%3Fq%3Dlogitec... | 15,525 |
def jsScriptData (self, data, url, ver): """Callback for loading <script src=""> data in the background If downloading is finished, data is None""" assert self.htmlparser.state[0] == 'wait', "non-wait state" wc.log.debug(wc.LOG_JS, "%s jsScriptData %r", self, data) if data is None: if not self.js_script: wc.log.warn(wc... | def jsScriptData (self, data, url, ver): """Callback for loading <script src=""> data in the background If downloading is finished, data is None""" assert self.htmlparser.state[0] == 'wait', "non-wait state" wc.log.debug(wc.LOG_JS, "%s jsScriptData %r", self, data) if data is None: if not self.js_script: wc.log.warn(wc... | 15,526 |
def new_instance (self, **opts): """generate new JSFilter instance""" return JSFilter(self.url, opts) | def new_instance (self, **opts): """generate new JSFilter instance""" return JSFilter(self.url, opts) | 15,527 |
def handle_read (self): """read data from SSL connection, put it into recv_buffer and call process_read""" assert self.connected debug(PROXY, '%s Connection.handle_read', self) | def handle_read (self): """read data from SSL connection, put it into recv_buffer and call process_read""" assert self.connected debug(PROXY, '%s Connection.handle_read', self) | 15,528 |
def fill_attrs (self, attrs, name): if name=='service': self._service = unxmlify(attrs.get('name')).encode('iso8859-1') self.ratings[self._service] = {} elif name=='category': assert self._service self._category = unxmlify(attrs.get('name')).encode('iso8859-1') elif name=='url': pass elif name=='pics': UrlRule.fill_att... | def fill_attrs (self, attrs, name): if name=='service': self._service = unxmlify(attrs.get('name')).encode('iso8859-1') self.ratings[self._service] = {} elif name=='category': assert self._service self._category = unxmlify(attrs.get('name')).encode('iso8859-1') elif name=='url': pass elif name=='pics': UrlRule.fill_att... | 15,529 |
def _form_apply_rating (form): # rating categories for catname, value in _get_prefix_vals(form, 'category_'): category = _get_category(catname) if category is None: # unknown category error['categoryvalue'] = True return False if category.iterable: realvalue = value else: realvalue = _intrange_from_string(value) if not... | def _form_apply_rating (form): # rating categories for catname, value in _get_prefix_vals(form, 'category_'): category = _get_category(catname) if category is None: # unknown category error['categoryvalue'] = True return if category.iterable: realvalue = value else: realvalue = _intrange_from_string(value) if not categ... | 15,530 |
def _form_apply_rating (form): # rating categories for catname, value in _get_prefix_vals(form, 'category_'): category = _get_category(catname) if category is None: # unknown category error['categoryvalue'] = True return False if category.iterable: realvalue = value else: realvalue = _intrange_from_string(value) if not... | def _form_apply_rating (form): # rating categories for catname, value in _get_prefix_vals(form, 'category_'): category = _get_category(catname) if category is None: # unknown category error['categoryvalue'] = True return False if category.iterable: realvalue = value else: realvalue = _intrange_from_string(value) if not... | 15,531 |
def _form_apply_rating (form): # rating categories for catname, value in _get_prefix_vals(form, 'category_'): category = _get_category(catname) if category is None: # unknown category error['categoryvalue'] = True return False if category.iterable: realvalue = value else: realvalue = _intrange_from_string(value) if not... | def _form_apply_rating (form): # rating categories for catname, value in _get_prefix_vals(form, 'category_'): category = _get_category(catname) if category is None: # unknown category error['categoryvalue'] = True return False if category.iterable: realvalue = value else: realvalue = _intrange_from_string(value) if not... | 15,532 |
def proxy_poll (timeout=0.0): """ Look for sockets with pending data and call the appropriate connection handlers. """ handlerCount = 0 if wc.proxy.Dispatcher.socket_map: e = wc.proxy.Dispatcher.socket_map.values() r = [x for x in e if x.readable()] w = [x for x in e if x.writable()] assert wc.log.debug(wc.LOG_PROXY, "... | def proxy_poll (timeout=0.0): """ Look for sockets with pending data and call the appropriate connection handlers. """ handlerCount = 0 if wc.proxy.Dispatcher.socket_map: e = wc.proxy.Dispatcher.socket_map.values() r = [x for x in e if x.readable()] w = [x for x in e if x.writable()] assert wc.log.debug(wc.LOG_PROXY, "... | 15,533 |
def _exec_form (form): # reset info/error global filterenabled, filterdisabled filterenabled = "" filterdisabled = "" info.clear() error.clear() res = [None] # proxy port if form.has_key('port'): _form_proxyport(_getval(form, 'port')) # proxy user if form.has_key('proxyuser'): _form_proxyuser(_getval(form, 'proxyuser')... | def _exec_form (form): # reset info/error global filterenabled, filterdisabled filterenabled = "" filterdisabled = "" info.clear() error.clear() res = [None] # proxy port if form.has_key('port'): _form_proxyport(_getval(form, 'port')) # proxy user if form.has_key('proxyuser'): _form_proxyuser(_getval(form, 'proxyuser')... | 15,534 |
def _exec_form (form): # reset info/error global filterenabled, filterdisabled filterenabled = "" filterdisabled = "" info.clear() error.clear() res = [None] # proxy port if form.has_key('port'): _form_proxyport(_getval(form, 'port')) # proxy user if form.has_key('proxyuser'): _form_proxyuser(_getval(form, 'proxyuser')... | def _exec_form (form): # reset info/error global filterenabled, filterdisabled filterenabled = "" filterdisabled = "" info.clear() error.clear() res = [None] # proxy port if form.has_key('port'): _form_proxyport(_getval(form, 'port')) # proxy user if form.has_key('proxyuser'): _form_proxyuser(_getval(form, 'proxyuser')... | 15,535 |
def _exec_form (form): # reset info/error global filterenabled, filterdisabled filterenabled = "" filterdisabled = "" info.clear() error.clear() res = [None] # proxy port if form.has_key('port'): _form_proxyport(_getval(form, 'port')) # proxy user if form.has_key('proxyuser'): _form_proxyuser(_getval(form, 'proxyuser')... | def _exec_form (form): # reset info/error global filterenabled, filterdisabled filterenabled = "" filterdisabled = "" info.clear() error.clear() res = [None] # proxy port if form.has_key('port'): _form_proxyport(_getval(form, 'port')) # proxy user if form.has_key('proxyuser'): _form_proxyuser(_getval(form, 'proxyuser')... | 15,536 |
def _form_proxyuser (proxyuser, res): if proxyuser != config['proxyuser']: res[0] = 407 config['proxyuser'] = proxyuser info['proxyuser'] = True | def _form_proxyuser (proxyuser, res): if proxyuser != config['proxyuser']: config['proxyuser'] = proxyuser info['proxyuser'] = True | 15,537 |
def _form_proxypass (proxypass, res): if proxypass != config['proxypass'] and proxypass!='__dummy__': res[0] = 407 config['proxypass'] = base64.encodestring(proxypass) info['proxypass'] = True | def _form_proxypass (proxypass, res): if proxypass != config['proxypass']: config['proxypass'] = proxypass info['proxypass'] = True | 15,538 |
def get_conf_filename (self, directory): return os.path.join(directory, "_%s_configdata.py"%self.get_name()) | def get_conf_filename (self, directory): return os.path.join(directory, "_%s_configdata.py"%self.get_name()) | 15,539 |
def expire_servers (self): """expire server connection that have been unused for too long""" debug(PROXY, "expire servers") expire_time = time.time() - 300 # Unused for five minutes to_expire = [] for addr,set in self.smap.items(): for server,status in set.items(): if status[0] == 'available' and status[1] < expire_tim... | def expire_servers (self): """expire server connection that have been unused for too long""" debug(PROXY, "pool expire servers") expire_time = time.time() - 300 # Unused for five minutes to_expire = [] for addr,set in self.smap.items(): for server,status in set.items(): if status[0] == 'available' and status[1] < expir... | 15,540 |
def connect (self, addr): """ Connect to addr. """ self.socket = socket.socket() self.socket.connect(addr) | def connect (self, addr): """ Connect to given address. """ debug("Client connect to %s" % str(addr)) self.socket = socket.socket() self.socket.connect(addr) | 15,541 |
def send_data (self, data): """ Send complete data to socket. """ self.socket.sendall(data) | def send_data (self, data): """ Send complete data to socket. """ self.socket.sendall(data) | 15,542 |
def read_data (self): """ Read until no more data is available. """ data = "" while wc.proxy.readable_socket(self.socket): s = self.socket.recv(4096) if not s: break data += s return data | def read_data (self): """ Read until no more data is available. """ data = "" while wc.proxy.readable_socket(self.socket): s = self.socket.recv(4096) if not s: break data += s return data | 15,543 |
def get_request (self): """ Get HttpRequest from internal data """ method = self.command uri = self.path vparts = self.request_version.split('/', 1)[1].split(".") version = (int(vparts[0]), int(vparts[1])) headers = [line[:-1] for line in self.headers.headers] if self.headers.has_key("Content-Length"): clen = int(self.... | def get_request (self): """ Get HttpRequest from internal data """ method = self.command uri = self.path vparts = self.request_version.split('/', 1)[1].split(".") version = (int(vparts[0]), int(vparts[1])) headers = [line[:-1] for line in self.headers.headers] if self.headers.has_key("Content-Length"): clen = int(self.... | 15,544 |
def get_request (self): """ Get HttpRequest from internal data """ method = self.command uri = self.path vparts = self.request_version.split('/', 1)[1].split(".") version = (int(vparts[0]), int(vparts[1])) headers = [line[:-1] for line in self.headers.headers] if self.headers.has_key("Content-Length"): clen = int(self.... | def get_request (self): """ Get HttpRequest from internal data """ method = self.command uri = self.path vparts = self.request_version.split('/', 1)[1].split(".") version = (int(vparts[0]), int(vparts[1])) headers = [line[:-1] for line in self.headers.headers] if self.headers.has_key("Content-Length"): clen = int(self.... | 15,545 |
def start_server (self): """ Start a HTTP server which is ready for use. @return: http server @rtype: BaseHTTPServer.HTTPServer """ port = 8000 server_address = ('', port) HandlerClass = HttpRequestHandler HandlerClass.protocol_version = "HTTP/1.1" ServerClass = BaseHTTPServer.HTTPServer httpd = ServerClass(server_addr... | def start_server (self): """ Start a HTTP server which is ready for use. @return: http server @rtype: BaseHTTPServer.HTTPServer """ port = 8000 server_address = ('', port) HandlerClass = HttpRequestHandler HandlerClass.protocol_version = "HTTP/1.1" ServerClass = HttpServer httpd = ServerClass(server_address, HandlerCla... | 15,546 |
def construct_response_data (self, response): """ Construct valid HTTP response data string. """ lines = [] version = "HTTP/%d.%d" % response.version lines.append("%s %d %s" % (version, response.status, response.msg)) lines.extend(response.headers) # an empty line ends the headers lines.extend(("", "")) data = "\r\n".j... | def construct_response_data (self, response): """ Construct a HTTP response data string. """ lines = [] version = "HTTP/%d.%d" % response.version lines.append("%s %d %s" % (version, response.status, response.msg)) lines.extend(response.headers) # an empty line ends the headers lines.extend(("", "")) data = "\r\n".join(... | 15,547 |
def get_response_data (response, url): """parse a response status line into tokens (protocol, status, msg)""" parts = response.split(None, 2) if len(parts) == 2: wc.log.warn(wc.LOG_PROXY, "Empty response message from %r", url) parts += ['Bummer'] elif len(parts) != 3: wc.log.error(wc.LOG_PROXY, "Invalid response %r fro... | def get_response_data (response, url): """parse a response status line into tokens (protocol, status, msg)""" parts = response.split(None, 2) if len(parts) == 2: wc.log.warn(wc.LOG_PROXY, "Empty response message from %r", url) parts += ['Bummer'] elif len(parts) != 3: wc.log.error(wc.LOG_PROXY, "Invalid response %r fro... | 15,548 |
def get_response_data (response, url): """parse a response status line into tokens (protocol, status, msg)""" parts = response.split(None, 2) if len(parts) == 2: wc.log.warn(wc.LOG_PROXY, "Empty response message from %r", url) parts += ['Bummer'] elif len(parts) != 3: wc.log.error(wc.LOG_PROXY, "Invalid response %r fro... | def get_response_data (response, url): """parse a response status line into tokens (protocol, status, msg)""" parts = response.split(None, 2) if len(parts) == 2: wc.log.warn(wc.LOG_PROXY, "Empty response message from %r", url) parts += ['Bummer'] elif len(parts) != 3: wc.log.error(wc.LOG_PROXY, "Invalid response %r fro... | 15,549 |
def __str__ (self): """return rule data as string""" return super(HeaderRule, self).__str__() + \ ("\nname: %r\nvalue: %r" % (self.name, self.value)) | def __str__ (self): """return rule data as string""" return super(HeaderRule, self).__str__() + \ ("\nname: %r\nvalue: %r" % (self.name, self.value)) | 15,550 |
def _main (): """USAGE: test/run.sh test/filterfile.py <config dir> <.html file>""" if len(sys.argv)!=3: print _main.__doc__ sys.exit(1) confdir = sys.argv[1] fname = sys.argv[2] if fname=="-": f = sys.stdin else: f = file(fname) logfile = os.path.join(confdir, "logging.conf") wc.initlog(logfile, wc.Name, filelogs=Fals... | def _main (): """USAGE: test/run.sh test/filterfile.py <config dir> <.html file>""" if len(sys.argv)!=3: print _main.__doc__ sys.exit(1) confdir = sys.argv[1] fname = sys.argv[2] if fname=="-": f = sys.stdin else: f = file(fname) logfile = os.path.join(confdir, "logging.conf") wc.initlog(logfile, wc.Name, filelogs=Fals... | 15,551 |
def jsProcessData (self, data): """process data produced by document.write() JavaScript""" self._debug(NIGHTMARE, "JS: document.write", `data`) self.js_output += 1 # parse recursively self.js_html.feed(data) | def jsProcessData (self, data): """process data produced by document.write() JavaScript""" self.js_output += 1 # parse recursively self.js_html.feed(data) | 15,552 |
def jsProcessPopup (self): """process javascript popup""" self._debug(NIGHTMARE, "JS: popup") self.js_popup += 1 | def jsProcessPopup (self): """process javascript popup""" self.js_popup += 1 | 15,553 |
def buf_append_data (self, data): """we have to make sure that we have no two following DATA things in the tag buffer. Why? To be 100% sure that an ENCLOSED match really matches enclosed data. """ self._debug(NIGHTMARE, "buf_append_data") if data[0]==DATA and self.buf and self.buf[-1][0]==DATA: self.buf[-1][1] += data[... | def buf_append_data (self, data): """we have to make sure that we have no two following DATA things in the tag buffer. Why? To be 100% sure that an ENCLOSED match really matches enclosed data. """ if data[0]==DATA and self.buf and self.buf[-1][0]==DATA: self.buf[-1][1] += data[1] else: self.buf.append(data) | 15,554 |
def flushbuf (self): """clear and return the output buffer""" self._debug(NIGHTMARE, "flushbuf") data = self.outbuf.getvalue() self.outbuf.close() self.outbuf = StringIO() return data | def flushbuf (self): """clear and return the output buffer""" data = self.outbuf.getvalue() self.outbuf.close() self.outbuf = StringIO() return data | 15,555 |
def _debugbuf (self): """print debugging information about data buffer status""" self._debug(NIGHTMARE, "self.outbuf", `self.outbuf.getvalue()`) self._debug(NIGHTMARE, "self.buf", `self.buf`) self._debug(NIGHTMARE, "self.waitbuf", `self.waitbuf`) self._debug(NIGHTMARE, "self.inbuf", `self.inbuf.getvalue()`) | def _debugbuf (self): """print debugging information about data buffer status""" self._debug(NIGHTMARE, "self.outbuf", `self.outbuf.getvalue()`) self._debug(NIGHTMARE, "self.buf", `self.buf`) self._debug(NIGHTMARE, "self.waitbuf", `self.waitbuf`) self._debug(NIGHTMARE, "self.inbuf", `self.inbuf.getvalue()`) | 15,556 |
def feed (self, data): """feed some data to the parser""" if self.state=='parse': # look if we must replay something if self.waited: self.waited = 0 waitbuf, self.waitbuf = self.waitbuf, [] self.replay(waitbuf) if self.state!='parse': self.inbuf.write(data) return data = self.inbuf.getvalue() + data self.inbuf.close() ... | def feed (self, data): """feed some data to the parser""" if self.state=='parse': # look if we must replay something if self.waited: self.waited = 0 waitbuf, self.waitbuf = self.waitbuf, [] self.replay(waitbuf) if self.state!='parse': self.inbuf.write(data) return data = self.inbuf.getvalue() + data self.inbuf.close() ... | 15,557 |
def feed (self, data): """feed some data to the parser""" if self.state=='parse': # look if we must replay something if self.waited: self.waited = 0 waitbuf, self.waitbuf = self.waitbuf, [] self.replay(waitbuf) if self.state!='parse': self.inbuf.write(data) return data = self.inbuf.getvalue() + data self.inbuf.close() ... | def feed (self, data): """feed some data to the parser""" if self.state=='parse': # look if we must replay something if self.waited: self.waited = 0 waitbuf, self.waitbuf = self.waitbuf, [] self.replay(waitbuf) if self.state!='parse': self.inbuf.write(data) return data = self.inbuf.getvalue() + data self.inbuf.close() ... | 15,558 |
def feed (self, data): """feed some data to the parser""" if self.state=='parse': # look if we must replay something if self.waited: self.waited = 0 waitbuf, self.waitbuf = self.waitbuf, [] self.replay(waitbuf) if self.state!='parse': self.inbuf.write(data) return data = self.inbuf.getvalue() + data self.inbuf.close() ... | def feed (self, data): """feed some data to the parser""" if self.state=='parse': # look if we must replay something if self.waited: self.waited = 0 waitbuf, self.waitbuf = self.waitbuf, [] self.replay(waitbuf) if self.state!='parse': self.inbuf.write(data) return data = self.inbuf.getvalue() + data self.inbuf.close() ... | 15,559 |
def flush (self): self._debug(HURT_ME_PLENTY, "flush") # flushing in wait state raises a filter exception if self.state=='wait': raise FilterWait("HtmlParser[%d]: waiting for data"%self.level) self.parser.flush() | def flush (self): # flushing in wait state raises a filter exception if self.state=='wait': raise FilterWait("HtmlParser[%d]: waiting for data"%self.level) self.parser.flush() | 15,560 |
def replay (self, waitbuf): """call the handler functions again with buffer data""" self._debug(NIGHTMARE, "replay", waitbuf) for item in waitbuf: if item[0]==DATA: self._data(item[1]) elif item[0]==STARTTAG: self.startElement(item[1], item[2]) elif item[0]==ENDTAG: self.endElement(item[1]) elif item[0]==COMMENT: self.... | def replay (self, waitbuf): """call the handler functions again with buffer data""" for item in waitbuf: if item[0]==DATA: self._data(item[1]) elif item[0]==STARTTAG: self.startElement(item[1], item[2]) elif item[0]==ENDTAG: self.endElement(item[1]) elif item[0]==COMMENT: self.comment(item[1]) | 15,561 |
def cdata (self, data): """character data""" self._debug(NIGHTMARE, "cdata", `data`) return self._data(data) | def cdata (self, data): """character data""" return self._data(data) | 15,562 |
def characters (self, data): """characters""" self._debug(NIGHTMARE, "characters", `data`) return self._data(data) | def characters (self, data): """characters""" return self._data(data) | 15,563 |
def comment (self, data): """a comment; accept only non-empty comments""" self._debug(NIGHTMARE, "comment", `data`) item = [COMMENT, data] if self.state=='wait': return self.waitbuf.append(item) if self.comments and data: self.buf.append(item) | def comment (self, data): """a comment; accept only non-empty comments""" item = [COMMENT, data] if self.state=='wait': return self.waitbuf.append(item) if self.comments and data: self.buf.append(item) | 15,564 |
def doctype (self, data): self._debug(NIGHTMARE, "doctype", `data`) return self._data("<!DOCTYPE%s>"%data) | def doctype (self, data): return self._data("<!DOCTYPE%s>"%data) | 15,565 |
def pi (self, data): self._debug(NIGHTMARE, "pi", `data`) return self._data("<?%s?>"%data) | def pi (self, data): return self._data("<?%s?>"%data) | 15,566 |
def startElement (self, tag, attrs): """We get a new start tag. New rules could be appended to the pending rules. No rules can be removed from the list.""" # default data self._debug(NIGHTMARE, "startElement", `tag`) tag = check_spelling(tag, self.url) item = [STARTTAG, tag, attrs] if self.state=='wait': return self.wa... | def startElement (self, tag, attrs): """We get a new start tag. New rules could be appended to the pending rules. No rules can be removed from the list.""" # default data tag = check_spelling(tag, self.url) item = [STARTTAG, tag, attrs] if self.state=='wait': return self.waitbuf.append(item) rulelist = [] filtered = 0... | 15,567 |
def startElement (self, tag, attrs): """We get a new start tag. New rules could be appended to the pending rules. No rules can be removed from the list.""" # default data self._debug(NIGHTMARE, "startElement", `tag`) tag = check_spelling(tag, self.url) item = [STARTTAG, tag, attrs] if self.state=='wait': return self.wa... | def startElement (self, tag, attrs): """We get a new start tag. New rules could be appended to the pending rules. No rules can be removed from the list.""" # default data self._debug(NIGHTMARE, "startElement", `tag`) tag = check_spelling(tag, self.url) item = [STARTTAG, tag, attrs] if self.state=='wait': return self.wa... | 15,568 |
def startElement (self, tag, attrs): """We get a new start tag. New rules could be appended to the pending rules. No rules can be removed from the list.""" # default data self._debug(NIGHTMARE, "startElement", `tag`) tag = check_spelling(tag, self.url) item = [STARTTAG, tag, attrs] if self.state=='wait': return self.wa... | def startElement (self, tag, attrs): """We get a new start tag. New rules could be appended to the pending rules. No rules can be removed from the list.""" # default data self._debug(NIGHTMARE, "startElement", `tag`) tag = check_spelling(tag, self.url) item = [STARTTAG, tag, attrs] if self.state=='wait': return self.wa... | 15,569 |
def endElement (self, tag): """We know the following: if a rule matches, it must be the one on the top of the stack. So we look only at the top rule. | def endElement (self, tag): """We know the following: if a rule matches, it must be the one on the top of the stack. So we look only at the top rule. | 15,570 |
def jsStartElement (self, tag, attrs): """Check popups for onmouseout and onmouseover. Inline extern javascript sources""" changed = 0 self.js_src = None self.js_output = 0 self.js_popup = 0 for name in ('onmouseover', 'onmouseout'): if attrs.has_key(name) and self.jsPopup(attrs, name): self._debug(NIGHTMARE, "JS: del"... | def jsStartElement (self, tag, attrs): """Check popups for onmouseout and onmouseover. Inline extern javascript sources""" changed = 0 self.js_src = None self.js_output = 0 self.js_popup = 0 for name in ('onmouseover', 'onmouseout'): if attrs.has_key(name) and self.jsPopup(attrs, name): del attrs[name] changed = 1 if ... | 15,571 |
def jsPopup (self, attrs, name): """check if attrs[name] javascript opens a popup window""" self._debug(NIGHTMARE, "JS: jsPopup") val = resolve_html_entities(attrs[name]) if not val: return self.js_env.attachListener(self) try: self.js_env.executeScriptAsFunction(val, 0.0) except jslib.error, msg: pass self.js_env.deta... | def jsPopup (self, attrs, name): """check if attrs[name] javascript opens a popup window""" val = resolve_html_entities(attrs[name]) if not val: return self.js_env.attachListener(self) try: self.js_env.executeScriptAsFunction(val, 0.0) except jslib.error, msg: pass self.js_env.detachListener(self) res = self.js_popup ... | 15,572 |
def jsForm (self, name, action, target): """when hitting a (named) form, notify the JS engine about that""" if not name: return self._debug(HURT_ME_PLENTY, "jsForm", `name`, `action`, `target`) self.js_env.addForm(name, action, target) | def jsForm (self, name, action, target): """when hitting a (named) form, notify the JS engine about that""" if not name: return self.js_env.addForm(name, action, target) | 15,573 |
def jsScriptData (self, data, url, ver): """Callback for loading <script src=""> data in the background If downloading is finished, data is None""" assert self.state=='wait' if data is None: if not self.js_script: print >> sys.stderr, "HtmlParser[%d]: empty JS src"%self.level, url else: self.buf.append([STARTTAG, "scri... | def jsScriptData (self, data, url, ver): """Callback for loading <script src=""> data in the background If downloading is finished, data is None""" assert self.state=='wait' if data is None: if not self.js_script: print >> sys.stderr, "HtmlParser[%d]: empty JS src"%self.level, url else: self.buf.append([STARTTAG, "scri... | 15,574 |
def jsScriptData (self, data, url, ver): """Callback for loading <script src=""> data in the background If downloading is finished, data is None""" assert self.state=='wait' if data is None: if not self.js_script: print >> sys.stderr, "HtmlParser[%d]: empty JS src"%self.level, url else: self.buf.append([STARTTAG, "scri... | def jsScriptData (self, data, url, ver): """Callback for loading <script src=""> data in the background If downloading is finished, data is None""" assert self.state=='wait' if data is None: if not self.js_script: print >> sys.stderr, "HtmlParser[%d]: empty JS src"%self.level, url else: self.buf.append([STARTTAG, "scri... | 15,575 |
def jsScriptSrc (self, url, language): """Start a background download for <script src=""> tags""" assert self.state=='parse' ver = 0.0 if language: mo = re.search(r'(?i)javascript(?P<num>\d\.\d)', language) if mo: ver = float(mo.group('num')) if self.base_url: url = urlparse.urljoin(self.base_url, url) else: url = urlp... | def jsScriptSrc (self, url, language): """Start a background download for <script src=""> tags""" assert self.state=='parse' ver = 0.0 if language: mo = re.search(r'(?i)javascript(?P<num>\d\.\d)', language) if mo: ver = float(mo.group('num')) if self.base_url: url = urlparse.urljoin(self.base_url, url) else: url = urlp... | 15,576 |
def jsScript (self, script, ver, item): """execute given script with javascript version ver""" self._debug(NIGHTMARE, "JS: jsScript", ver, `script`) assert self.state == 'parse' assert len(self.buf) >= 2 self.js_output = 0 self.js_env.attachListener(self) # start recursive html filter (used by jsProcessData) self.js_ht... | def jsScript (self, script, ver, item): """execute given script with javascript version ver""" assert self.state == 'parse' assert len(self.buf) >= 2 self.js_output = 0 self.js_env.attachListener(self) # start recursive html filter (used by jsProcessData) self.js_html = FilterHtmlParser(self.rules, self.pics, self.url... | 15,577 |
def jsEndScript (self, item): self._debug(NIGHTMARE, "JS: endScript") assert len(self.buf) >= 2 if self.js_output: try: self.js_html.feed('') self.js_html.flush() except FilterWait: self.state = 'wait' self.waited = 'True' make_timer(0.1, lambda : self.jsEndScript(item)) return self.js_html._debugbuf() assert not self.... | def jsEndScript (self, item): assert len(self.buf) >= 2 if self.js_output: try: self.js_html.feed('') self.js_html.flush() except FilterWait: self.state = 'wait' self.waited = 'True' make_timer(0.1, lambda : self.jsEndScript(item)) return self.js_html._debugbuf() assert not self.js_html.inbuf.getvalue() assert not sel... | 15,578 |
def jsEndScript (self, item): self._debug(NIGHTMARE, "JS: endScript") assert len(self.buf) >= 2 if self.js_output: try: self.js_html.feed('') self.js_html.flush() except FilterWait: self.state = 'wait' self.waited = 'True' make_timer(0.1, lambda : self.jsEndScript(item)) return self.js_html._debugbuf() assert not self.... | def jsEndScript (self, item): self._debug(NIGHTMARE, "JS: endScript") assert len(self.buf) >= 2 if self.js_output: try: self.js_html.feed('') self.js_html.flush() except FilterWait: self.state = 'wait' self.waited = 'True' make_timer(0.1, lambda : self.jsEndScript(item)) return self.js_html._debugbuf() assert not self.... | 15,579 |
def applyfilter (filterstage, data, fun, attrs): """ Apply all filters which are registered in the given filter stage. For different filter stages we have different data objects. Look at the filter examples. One can prevent all filtering with the 'nofilter' attribute, or deactivate single filter modules with 'nofilter-... | def applyfilter (filterstage, data, fun, attrs): """ Apply all filters which are registered in the given filter stage. For different filter stages we have different data objects. Look at the filter examples. One can prevent all filtering with the 'nofilter' attribute, or deactivate single filter modules with 'nofilter-... | 15,580 |
def stopwatch (pidfile, watchfile): """stop webcleaner and the monitor""" if not os.path.exists(watchfile): return i18n._("Watcher was not running (no lock file found)"), 1 msg, status = _stop(watchfile) msg2, status2 = stop(pidfile) return "%s\n%s"%(msg,msg2), status2 | def stopwatch (pidfile, watchfile): """stop webcleaner and the monitor""" if not os.path.exists(watchfile): return i18n._("Watcher was not running (no lock file found)"), 1 msg1, status1 = _stop(watchfile) msg2, status2 = stop(pidfile) return "%s\n%s"%(msg,msg2), status2 | 15,581 |
def stopwatch (pidfile, watchfile): """stop webcleaner and the monitor""" if not os.path.exists(watchfile): return i18n._("Watcher was not running (no lock file found)"), 1 msg, status = _stop(watchfile) msg2, status2 = stop(pidfile) return "%s\n%s"%(msg,msg2), status2 | def stopwatch (pidfile, watchfile): """stop webcleaner and the monitor""" if not os.path.exists(watchfile): return i18n._("Watcher was not running (no lock file found)"), 1 msg, status = _stop(watchfile) msg2, status2 = stop(pidfile) if msg1: if msg2: msg = "%s\n%s"%(msg1,msg2) else: msg = msg1 else: msg = msg2 if stat... | 15,582 |
def toxml (self): """Rule data as XML for storing""" s = u"""<?xml version="1.0" encoding="%s"?> | def toxml (self): """Rule data as XML for storing""" s = u"""<?xml version="1.0" encoding="%s"?> | 15,583 |
def process_headers (self): i = self.recv_buffer.find('\r\n\r\n') if i >= 0: # Two newlines ends headers i += 4 # Skip over newline terminator # the first 2 chars are the newline of request data = self.read(i)[2:] self.headers = rfc822.Message(StringIO(data)) #debug(HURT_ME_PLENTY, "Proxy: C/Headers", `self.headers.hea... | def process_headers (self): i = self.recv_buffer.find('\r\n\r\n') if i >= 0: # Two newlines ends headers i += 4 # Skip over newline terminator # the first 2 chars are the newline of request data = self.read(i)[2:] self.headers = rfc822.Message(StringIO(data)) #debug(HURT_ME_PLENTY, "Proxy: C/Headers", `self.headers.hea... | 15,584 |
def process_headers (self): i = self.recv_buffer.find('\r\n\r\n') if i >= 0: # Two newlines ends headers i += 4 # Skip over newline terminator # the first 2 chars are the newline of request data = self.read(i)[2:] self.headers = rfc822.Message(StringIO(data)) #debug(HURT_ME_PLENTY, "Proxy: C/Headers", `self.headers.hea... | def process_headers (self): i = self.recv_buffer.find('\r\n\r\n') if i >= 0: # Two newlines ends headers i += 4 # Skip over newline terminator # the first 2 chars are the newline of request data = self.read(i)[2:] self.headers = rfc822.Message(StringIO(data)) #debug(HURT_ME_PLENTY, "Proxy: C/Headers", `self.headers.hea... | 15,585 |
def fix_configdata (): """ Fix install and config paths in the config file. """ name = "_webcleaner2_configdata.py" conffile = os.path.join(sys.prefix, "Lib", "site-packages", name) lines = [] for line in file(conffile): if line.startswith("install_") or \ line.startswith("config_") or \ line.startswith("template_"): l... | def fix_configdata (): """ Fix install and config paths in the config file. """ name = "_webcleaner2_configdata.py" conffile = os.path.join(sys.prefix, "Lib", "site-packages", name) lines = [] for line in file(conffile): if line.startswith("install_") or \ line.startswith("config_") or \ line.startswith("template_"): l... | 15,586 |
def do_quit (event=None): root.destroy() | def do_quit (event=None): root.destroy() | 15,587 |
def purge_tempfiles (): """ Ask if user wants to purge local config files. """ import wc files = glob.glob(os.path.join(wc.ConfigDir, "local_*.zap")) if not files: return import tkMessageBox answer = tkMessageBox.askyesno(_("%s config purge") % wc.AppName, _("""There are local filter rules in the configuration director... | def purge_tempfiles (): """ Ask if user wants to purge local config files. """ import wc files = glob.glob(os.path.join(wc.ConfigDir, "local_*.zap")) if not files: return import tkMessageBox answer = tkMessageBox.askyesno(_("%s config purge") % wc.AppName, _("""There are local filter rules in the configuration director... | 15,588 |
def process_headers (self): """look for headers and process them if found""" # Headers are terminated by a blank line .. now in the regexp, # we want to say it's either a newline at the beginning of # the document, or it's a lot of headers followed by two newlines. # The cleaner alternative would be to read one line at... | def process_headers (self): """look for headers and process them if found""" # Headers are terminated by a blank line .. now in the regexp, # we want to say it's either a newline at the beginning of # the document, or it's a lot of headers followed by two newlines. # The cleaner alternative would be to read one line at... | 15,589 |
def check_connect (self, addr): """ Check if the connection is etablished. See also http://cr.yp.to/docs/connect.html and connect(2) manpage. """ wc.log.debug(wc.LOG_PROXY, '%s check connect', self) self.connect_checks += 1 if self.connect_checks >= 50: wc.log.info(wc.LOG_PROXY, '%s connect timed out', self) self.handl... | def check_connect (self, addr): """ Check if the connection is etablished. See also http://cr.yp.to/docs/connect.html and connect(2) manpage. """ wc.log.debug(wc.LOG_PROXY, '%s check connect', self) self.connect_checks += 1 if self.connect_checks >= 50: wc.log.info(wc.LOG_PROXY, '%s connect timed out', self) self.handl... | 15,590 |
def process_headers (self): # Headers are terminated by a blank line .. now in the regexp, # we want to say it's either a newline at the beginning of # the document, or it's a lot of headers followed by two newlines. # The cleaner alternative would be to read one line at a time # until we get to a blank line... m = re.... | def process_headers (self): # Headers are terminated by a blank line .. now in the regexp, # we want to say it's either a newline at the beginning of # the document, or it's a lot of headers followed by two newlines. # The cleaner alternative would be to read one line at a time # until we get to a blank line... m = re.... | 15,591 |
def socket_send (sock, data): """ Send data to socket. """ sock.sendall(data) debug("Socket sent %r" % data) | def socket_send (sock, data): """ Send data to socket. """ sock.sendall(data) debug("Socket sent %d bytes: %r" % (len(data), data)) | 15,592 |
def socket_read (sock): """ Read data from socket until no more data is available. """ data = "" while wc.proxy.readable_socket(sock): s = sock.recv(8192) if not s: break data += s debug("Socket read %r" % data) return data | def socket_read (sock): """ Read data from socket until no more data is available. """ data = "" while wc.proxy.readable_socket(sock): s = sock.recv(8192) if not s: break data += s debug("Socket read %d bytes: %r" % (len(data), data)) return data | 15,593 |
def socketfile_read (sock): """ Read data from socket until no more data is available. """ data = "" while wc.proxy.readable_socket(sock): s = sock.read(1) if not s: break data += s debug("Socket file read %r" % data) return data + sock._rbuf | def socketfile_read (sock): """ Read data from socket until no more data is available. """ data = "" while wc.proxy.readable_socket(sock): s = sock.read(1) if not s: break data += s debug("Socket file read %d bytes: %r" % (len(data), data)) return data + sock._rbuf | 15,594 |
def create_conf_file(self, directory, data=[]): data.insert(0, "# this file is automatically created by setup.py") filename = os.path.join(directory, self.config_file) # add metadata metanames = dir(self.metadata) + \ ['fullname', 'contact', 'contact_email'] for name in metanames: method = "get_" + name cmd = "%s = %s"... | def create_conf_file(self, directory, data=[]): data.insert(0, "# this file is automatically created by setup.py") filename = os.path.join(directory, self.config_file) # add metadata metanames = dir(self.metadata) + \ ['fullname', 'contact', 'contact_email'] for name in metanames: method = "get_" + name cmd = "%s = %s"... | 15,595 |
def write (self): """ Write pickled cache to disk. """ fp = file(self.filename, 'wb') pickle.dump(self.cache, fp, 1) fp.close() | def write (self): """ Write pickled cache to disk. """ fp = file(self.filename, 'wb') pickle.dump(self.cache, fp, 1) fp.close() | 15,596 |
def check_headers (self): """add missing content-type and/or encoding headers if needed""" # 304 Not Modified does not send any type or encoding info, # because this info was cached if self.statuscode == '304': return # check content-type against our own guess i = self.document.find('?') if i>0: document = self.documen... | def if self.headers.get('Content-Type', '').lower().startswith('application/x-httpd-php'): self.headers['Content-Type'] = 'text/html' check_headers if self.headers.get('Content-Type', '').lower().startswith('application/x-httpd-php'): self.headers['Content-Type'] = 'text/html' (self): if self.headers.get('Content-Type'... | 15,597 |
def create_socket (family, socktype, proto=0): """ Create a socket with given family and type. If SSL context is given an SSL socket is created. """ sock = socket.socket(family, socktype, proto=proto) # store family, type and proto in the object sock.family = family sock.socktype = socktype sock.proto = proto # XXX dis... | def sock = wc.objproxy.Proxy(sock) create_socket sock = wc.objproxy.Proxy(sock) (family, sock = wc.objproxy.Proxy(sock) socktype, sock = wc.objproxy.Proxy(sock) proto=0): sock = wc.objproxy.Proxy(sock) """ sock = wc.objproxy.Proxy(sock) Create sock = wc.objproxy.Proxy(sock) a sock = wc.objproxy.Proxy(sock) sock... | 15,598 |
def accept (self): """ Accept a new connection on the socket. | def accept (self): """ Accept a new connection on the socket. | 15,599 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.