rem stringlengths 0 322k | add stringlengths 0 2.05M | context stringlengths 8 228k |
|---|---|---|
if self.pointer == i: clr.append('selected') | y = i + 1 clr = deque(base_clr) | def draw(self): base_clr = deque() base_clr.append('in_taskview') lst = self.get_list() |
descr = obj.get_description() self.addstr(y, 0, descr, self.wid) self.color_at(y, 0, self.wid, clr) | if self.pointer == i: clr.append('selected') | def draw(self): base_clr = deque() base_clr.append('in_taskview') lst = self.get_list() |
else: if self.hei > 1: self.addstr(1, 0, "No task in the queue.") self.color_at(1, 0, self.wid, base_clr, 'error') | descr = obj.get_description() self.addstr(y, 0, descr, self.wid) self.color_at(y, 0, self.wid, clr) | def draw(self): base_clr = deque() base_clr.append('in_taskview') lst = self.get_list() |
self.color_reset() | else: if self.hei > 1: self.addstr(1, 0, "No task in the queue.") self.color_at(1, 0, self.wid, base_clr, 'error') self.color_reset() | def draw(self): base_clr = deque() base_clr.append('in_taskview') lst = self.get_list() |
alias(e=quit) | alias(e=edit) | def alias(**kw): for key, value in kw.items(): by_name[key] = value |
@depends_on(app_editor, Applications.app_self) | def app_editor(self, c): try: default_editor = os.environ['EDITOR'] except KeyError: pass else: parts = default_editor.split() exe_name = os.path.basename(parts[0]) if exe_name in get_executables(): return tuple(parts) + tuple(c) | |
sys.stdout.write("\033]2;" + hostname + \ self.fm.env.pwd.path + "\007") | sys.stdout.write("\033]2;" + hostname + cwd + "\007") | def draw(self): """Erase the window, then draw all objects in the container""" self.win.touchwin() DisplayableContainer.draw(self) if self.settings.update_title: hostname = str(socket.gethostname()) try: cwd = self.fm.env.pwd.path except: cwd = ' - ranger' sys.stdout.write("\033]2;" + hostname + \ self.fm.env.pwd.path + "\007") self.win.refresh() |
import os | def load_content_if_outdated(self, *a, **k): """ Load the contents of the directory if it's outdated or not done yet """ | |
cached_mtime = self.stat.st_mtime | cached_mtime = self.load_content_mtime | def load_content_if_outdated(self, *a, **k): """ Load the contents of the directory if it's outdated or not done yet """ |
return True | return False | def has_preview(self): if not self.fm.settings.preview_files: return False if self.is_socket or self.is_fifo or self.is_device: return False if not self.accessible: return False if self.image or self.container: return True if PREVIEW_WHITELIST.search(self.basename): return True if PREVIEW_BLACKLIST.search(self.basename): return False if self.path == '/dev/core' or self.path == '/proc/kcore': return False if self.is_binary(): return False return True |
if self.fm.settings.preview_script: return True | def has_preview(self): if self.fm.settings.preview_script: return True if not self.fm.settings.preview_files: return False if self.is_socket or self.is_fifo or self.is_device: return False if not self.accessible: return False if self.image or self.container: return False if PREVIEW_WHITELIST.search(self.basename): return True if PREVIEW_BLACKLIST.search(self.basename): return False if self.path == '/dev/core' or self.path == '/proc/kcore': return False if self.is_binary(): return False return True | |
self.notify("%s complete" % path) | def on_after(signal): self.notify("%s complete" % path) exit = signal.process.poll() content = signal.loader.stdout_buffer content += signal.process.stdout.read() if exit == 0: data[(width, height)] = content elif exit == 3: data[(-1, height)] = content elif exit == 4: data[(width, -1)] = content elif exit == 5: data[(-1, -1)] = content else: data[(-1, -1)] = None # XXX if self.env.cf.path == path: self.ui.browser.pager.need_redraw = True self.ui.browser.need_redraw = True data['loading'] = False | |
result = self.settings.collapse_preview and self.preview and \ not self.columns[-1].has_preview() and self.stretch_ratios if result: return True if self.columns[-1].target: target = self.columns[-1].target | if not self.settings.collapse_preview or not self.preview \ or not self.stretch_ratios: return False result = not self.columns[-1].has_preview() target = self.columns[-1].target if not result and target and target.is_file: | def _collapse(self): # Should the last column be cut off? (Because there is no preview) result = self.settings.collapse_preview and self.preview and \ not self.columns[-1].has_preview() and self.stretch_ratios if result: return True if self.columns[-1].target: target = self.columns[-1].target try: result = not self.fm.previews[target.realpath]['foundpreview'] except: return self.old_collapse self.old_collapse = result return result |
for child in reversed(rows): if not child.has_preview(): right_end = child.x - 1 else: break if right_end < left_start: right_end = self.wid - 1 | if not self.pager.visible: for child in reversed(rows): if not child.has_preview(): right_end = child.x - 1 else: break if right_end < left_start: right_end = self.wid - 1 | def _draw_borders(self): win = self.win self.color('in_browser', 'border') |
max(1, self.wid - left)) | max(1, self.wid - left - pad)) | def resize(self, y, x, hei, wid): """Resize all the columns according to the given ratio""" DisplayableContainer.resize(self, y, x, hei, wid) borders = self.settings.draw_borders pad = 1 if borders else 0 left = pad |
if self.end_hook(process): | if self.end_hook: | def generate(self): self.process = process = Popen(self.args, stdout=open(os.devnull, 'w'), stderr=PIPE) if self.begin_hook: self.begin_hook(process) while process.poll() is None: rd, _, __ = select.select( [process.stderr], [], [], 0.05) if rd: self.notify(process.stderr.readline(), bad=True) sleep(0.02) yield if self.end_hook(process): self.end_hook(process) |
and cls.allow_abbrev] | and cls.allow_abbrev \ or cmd == name] | def get_command(name, abbrev=True): if abbrev: lst = [cls for cmd, cls in by_name.items() \ if cmd.startswith(name) \ and cls.allow_abbrev] if len(lst) == 0: raise KeyError if len(lst) == 1: return lst[0] raise ValueError("Ambiguous command") else: try: return by_name[name] except KeyError: return None |
i = absolute | if absolute < 0: i = absolute + len(self.lines) else: i = absolute | def move(self, relative=0, absolute=None, pages=False): i = self.scroll_begin if isinstance(absolute, int): i = absolute |
if chr(key) in digits and ANYKEY not in tree: | if key in digitlist and ANYKEY not in tree: | def _do_eval_quantifier(self, key): if self.eval_command: tree = self.tree_pointer else: tree = self.dir_tree_pointer if chr(key) in digits and ANYKEY not in tree: attr = self.eval_command and 'quant' or 'direction_quant' if getattr(self, attr) is None: setattr(self, attr, 0) setattr(self, attr, getattr(self, attr) * 10 + key - 48) else: self.eval_quantifier = False return None return True |
chr(key) in digits or self.direction_keys._tree[key] | key in digitlist or self.direction_keys._tree[key] | def _do_eval_command(self, key): assert isinstance(self.tree_pointer, dict), self.tree_pointer try: self.tree_pointer = self.tree_pointer[key] except TypeError: self.failure = True return None except KeyError: try: chr(key) in digits or self.direction_keys._tree[key] self.tree_pointer = self.tree_pointer[DIRKEY] except KeyError: try: self.tree_pointer = self.tree_pointer[ANYKEY] except KeyError: self.failure = True return None else: self.matches.append(key) assert isinstance(self.tree_pointer, (Binding, dict)) self._try_to_finish() else: assert isinstance(self.tree_pointer, (Binding, dict)) self.eval_command = False self.eval_quantifier = True self.dir_tree_pointer = self.direction_keys._tree else: if isinstance(self.tree_pointer, dict): try: self.command = self.tree_pointer[PASSIVE_ACTION] except (KeyError, TypeError): self.command = None self._try_to_finish() |
kw['stdout'] = sys.stderr | kw['stdout'] = sys.stdout | def run(self): """ Run the application in the way specified by the options. |
log("freeing {0}".format(self)) | def __del__(self): log("freeing {0}".format(self)) for handler in self.handlers: self.settings.signal_unbind(handler) | |
for key, val in special_keys.items(): | for key, val in tuple(special_keys.items()): | def parse_keybinding(obj): """ Translate a keybinding to a sequence of integers Example: lol<CR> => (ord('l'), ord('o'), ord('l'), ord('\n')) => (108, 111, 108, 10) x<A-Left> => (120, (27, curses.KEY_LEFT)) """ assert isinstance(obj, (tuple, int, str)) if isinstance(obj, tuple): for char in obj: yield char elif isinstance(obj, int): yield obj elif isinstance(obj, str): in_brackets = False bracket_content = None for char in obj: if in_brackets: if char == '>': in_brackets = False string = ''.join(bracket_content).lower() try: keys = special_keys[string] for key in keys: yield key except KeyError: yield ord('<') for c in bracket_content: yield ord(c) yield ord('>') except TypeError: yield keys # it was no tuple, just an int else: bracket_content.append(char) else: if char == '<': in_brackets = True bracket_content = [] else: yield ord(char) if in_brackets: yield ord('<') for c in bracket_content: yield ord(c) |
default_editor = os.environ['EDITOR'] parts = default_editor.split() exe_name = os.path.basename(parts[0]) | try: default_editor = os.environ['EDITOR'] except KeyError: pass else: parts = default_editor.split() exe_name = os.path.basename(parts[0]) if exe_name in self.fm.executables: return tuple(parts) + tuple(c) | def app_editor(self, c): default_editor = os.environ['EDITOR'] parts = default_editor.split() exe_name = os.path.basename(parts[0]) |
if exe_name in self.fm.executables: return tuple(parts) + tuple(c) else: return self.either(c, 'vim', 'emacs', 'nano') | return self.either(c, 'vim', 'emacs', 'nano') | def app_editor(self, c): default_editor = os.environ['EDITOR'] parts = default_editor.split() exe_name = os.path.basename(parts[0]) |
pos = self.wid - 3 | pos = self.wid - 1 | def click(self, event): """Handle a MouseEvent""" direction = event.mouse_wheel_direction() if direction: self.fm.tab_move(direction) self.need_redraw = True return True |
def open_console(self, mode=':', string='', prompt=None): | def open_console(self, mode=cmode.COMMAND, string='', prompt=None): | def open_console(self, mode=':', string='', prompt=None): """Open the console if the current UI supports that""" if hasattr(self.ui, 'open_console'): self.ui.open_console(mode, string, prompt=prompt) |
map('zd', fm.toggle_boolean_option('directories_first')) | map('zd', fm.toggle_boolean_option('sort_directories_first')) | def move(arg): arg.wdg.move(narg=arg.n, **arg.direction) |
map('<home>', wdg.move(right=0, absolute=True)) map('<end>', wdg.move(right=-1, absolute=True)) | map('<home>', '<C-A>', wdg.move(right=0, absolute=True)) map('<end>', '<C-E>', wdg.move(right=-1, absolute=True)) | def ctrl_c(arg): try: item = arg.fm.loader.queue[0] except: arg.fm.notify("Type Q or :quit<Enter> to exit Ranger") else: arg.fm.notify("Aborting: " + item.get_description()) arg.fm.loader.remove(index=0) |
map('<backspace>', wdg.delete(-1)) map('<delete>', wdg.delete(0)) | map('<backspace>', '<C-H>', wdg.delete(-1)) map('<delete>', '<C-D>', wdg.delete(0)) | def ctrl_c(arg): try: item = arg.fm.loader.queue[0] except: arg.fm.notify("Type Q or :quit<Enter> to exit Ranger") else: arg.fm.notify("Aborting: " + item.get_description()) arg.fm.loader.remove(index=0) |
obj = CommandLoader(args=['mv'] + mv_flags + + [f.path for f in copied_files] | obj = CommandLoader(args=['mv'] + mv_flags \ + [f.path for f in copied_files] \ | def refresh(_): cwd = self.env.get_directory(original_path) cwd.load_content() |
obj = CommandLoader(args=['cp', '-f'] + cp_flags | obj = CommandLoader(args=['cp', '-f'] + cp_flags \ | def refresh(_): cwd = self.env.get_directory(original_path) cwd.load_content() |
obj = CommandLoader(args=['cp'] + cp_flags + [f.path for f in copied_files] | obj = CommandLoader(args=['cp'] + cp_flags \ + [f.path for f in copied_files] \ | def refresh(_): cwd = self.env.get_directory(original_path) cwd.load_content() |
if ranger.debug: | if ranger.arg.debug: | def notify(self, text, duration=4, bad=False): if isinstance(text, Exception): if ranger.debug: raise bad = True text = str(text) self.log.appendleft(text) if hasattr(self.ui, 'notify'): self.ui.notify(text, duration=duration, bad=bad) |
minY = goodY.min() yRange = maxY - minY | def plotHistogram(coaddName, chiSqOrder): coadd = pyfits.open(coaddName) coaddData = coadd[0].data # undo normalization coaddData *= float(chiSqOrder) # get rid of nans and infs goodData = numpy.extract(numpy.isfinite(coaddData.flat), coaddData.flat) goodData = numpy.extract(goodData < 50, goodData) hist, binEdges = numpy.histogram(goodData, bins=NBins) hist = numpy.array(hist, dtype=float) hist /= hist.sum() if UseLogForY: dataY = numpy.log10(hist) else: dataY = hist dataX = binEdges[0:-1] if UseSqrtForX: plotDataX = numpy.sqrt(dataX) else: plotDataX = dataX # plot histogram: log10(frequency) vs. value pyplot.plot(plotDataX, dataY, drawstyle="steps") if UseLogForY: pyplot.ylabel('log10 frequency') else: pyplot.ylabel('frequency') if UseSqrtForX: pyplot.xlabel('sqrt of sum of (counts/noise)^2') else: pyplot.xlabel('sum of (counts/noise)^2') # plot chiSq probability distribution chiSqX = dataX chiSqDist = numpy.power(chiSqX, (chiSqOrder / 2.0) - 1) * numpy.exp(-chiSqX / 2.0) chiSqDist /= chiSqDist.sum() if UseLogForY: chiSqDistY = numpy.log10(chiSqDist) else: chiSqDistY = chiSqDist pyplot.plot(plotDataX, chiSqDistY) # set plot limits goodY = numpy.extract(numpy.isfinite(dataY), dataY) maxY = goodY.max() minY = goodY.min() yRange = maxY - minY # plot out to where Y falls to 1% of max value maxYInd = goodY.argmax() yEndVal = minY + (yRange * 0.01) smallYIndices = numpy.where(goodY < yEndVal)[0] endInd = numpy.extract(smallYIndices > maxYInd, smallYIndices)[0] pyplot.xlim((0, plotDataX[endInd])) yMargin = yRange * 0.05 pyplot.ylim((minY, maxY + yMargin)) pyplot.show() | |
yEndVal = minY + (yRange * 0.01) smallYIndices = numpy.where(goodY < yEndVal)[0] endInd = numpy.extract(smallYIndices > maxYInd, smallYIndices)[0] | tailMinY = goodY[maxYInd:].min() yRange = maxY - tailMinY yEndVal = tailMinY + (yRange * 0.01) endInd = numpy.where(goodY[maxYInd:] <= yEndVal)[0][0] + maxYInd endInd = len(goodY)-1 | def plotHistogram(coaddName, chiSqOrder): coadd = pyfits.open(coaddName) coaddData = coadd[0].data # undo normalization coaddData *= float(chiSqOrder) # get rid of nans and infs goodData = numpy.extract(numpy.isfinite(coaddData.flat), coaddData.flat) goodData = numpy.extract(goodData < 50, goodData) hist, binEdges = numpy.histogram(goodData, bins=NBins) hist = numpy.array(hist, dtype=float) hist /= hist.sum() if UseLogForY: dataY = numpy.log10(hist) else: dataY = hist dataX = binEdges[0:-1] if UseSqrtForX: plotDataX = numpy.sqrt(dataX) else: plotDataX = dataX # plot histogram: log10(frequency) vs. value pyplot.plot(plotDataX, dataY, drawstyle="steps") if UseLogForY: pyplot.ylabel('log10 frequency') else: pyplot.ylabel('frequency') if UseSqrtForX: pyplot.xlabel('sqrt of sum of (counts/noise)^2') else: pyplot.xlabel('sum of (counts/noise)^2') # plot chiSq probability distribution chiSqX = dataX chiSqDist = numpy.power(chiSqX, (chiSqOrder / 2.0) - 1) * numpy.exp(-chiSqX / 2.0) chiSqDist /= chiSqDist.sum() if UseLogForY: chiSqDistY = numpy.log10(chiSqDist) else: chiSqDistY = chiSqDist pyplot.plot(plotDataX, chiSqDistY) # set plot limits goodY = numpy.extract(numpy.isfinite(dataY), dataY) maxY = goodY.max() minY = goodY.min() yRange = maxY - minY # plot out to where Y falls to 1% of max value maxYInd = goodY.argmax() yEndVal = minY + (yRange * 0.01) smallYIndices = numpy.where(goodY < yEndVal)[0] endInd = numpy.extract(smallYIndices > maxYInd, smallYIndices)[0] pyplot.xlim((0, plotDataX[endInd])) yMargin = yRange * 0.05 pyplot.ylim((minY, maxY + yMargin)) pyplot.show() |
BaseHTTPServer.HTTPServer.__init__(self, (host, port), RecordHandler) | try: BaseHTTPServer.HTTPServer.__init__(self, (host, port), RecordHandler) except Exception, e: logging.critical('Could not start HTTPServer on port %d: %s', port, e) return | def __init__( self, http_archive_filename, use_deterministic_script, real_dns_lookup, host='localhost', port=80): self.use_deterministic_script = use_deterministic_script self.archive_filename = http_archive_filename self.real_http_request = RealHttpRequest(real_dns_lookup) |
BaseHTTPServer.HTTPServer.__init__(self, (host, port), ReplayHandler) | try: BaseHTTPServer.HTTPServer.__init__(self, (host, port), ReplayHandler) except Exception, e: logging.critical('Could not start HTTPServer on port %d: %s', port, e) return | def __init__(self, http_archive_filename, use_deterministic_script, host='localhost', port=80): self.use_deterministic_script = use_deterministic_script self.http_archive = httparchive.HttpArchive.Create(http_archive_filename) logging.info('Loaded %d responses from %s', len(self.http_archive), http_archive_filename) BaseHTTPServer.HTTPServer.__init__(self, (host, port), ReplayHandler) logging.info('Replaying on %s:%s...', host, port) |
query.filter("platform =", int(self.request.get("platform_filter"))) if self.request.get("version_vilter"): query.filter("version =", int(self.request.get("version_vilter"))) | query.filter("platform =", self.request.get("platform_filter")) if self.request.get("version_filter"): query.filter("version =", self.request.get("version_filter")) | def do_set_search(self): query = models.TestSet.all(); query.order("-date") |
self._ipfw([ 'pipe', dns_pipe, 'config', 'bw', '0', 'delay', delay_ms, 'plr', packet_loss_rate ]) self._ipfw(['add', self.pipe_set, 'pipe', dns_pipe, 'udp', 'from', 'any', 'to', '127.0.0.1', 'dst-port', '53']) | def set_traffic_shaping(self, up_bandwidth = '0', down_bandwidth = '0', delay_ms = '0', packet_loss_rate = '0'): """Start shaping traffic. | |
def __init__(self, host='127.0.0.1', port=53, platform_settings=platformsettings.get_platform_settings()): | def __init__(self, host='127.0.0.1', port=53, platform_settings=None): if not platformsettings: platformsettings.get_platform_settings() | def __init__(self, host='127.0.0.1', port=53, platform_settings=platformsettings.get_platform_settings()): self.host = host self.platform_settings = platform_settings try: SocketServer.ThreadingUDPServer.__init__( self, (host, port), UdpDnsHandler) except socket.error, (error_number, msg): if error_number == errno.EACCES: raise PermissionDenied raise logging.info('Started DNS server on (%s:%s)...', host, port) platform_settings.set_primary_dns(host) |
self.platform_settings = platform_settings | self.restore_primary_dns = platform_settings.restore_primary_dns | def __init__(self, host='127.0.0.1', port=53, platform_settings=platformsettings.get_platform_settings()): self.host = host self.platform_settings = platform_settings try: SocketServer.ThreadingUDPServer.__init__( self, (host, port), UdpDnsHandler) except socket.error, (error_number, msg): if error_number == errno.EACCES: raise PermissionDenied raise logging.info('Started DNS server on (%s:%s)...', host, port) platform_settings.set_primary_dns(host) |
self.connection.close() | self.close_connection = 1 | def send_archived_http_response(self, response): try: # We need to set the server name before we start the response. # Take a scan through the response headers here use_chunked = False has_content_length = False server_name = "WebPageReplay" for header, value in response.headers: if header == "server": server_name = value if header == "transfer-encoding": use_chunked = True if header == "content-length": has_content_length = True self.server_version = server_name self.sys_version = "" |
logging.critical("Traffic Shaping Exception ", e) | logging.critical("Traffic Shaping Exception: %s ", e) | def set_traffic_shaping(self, up_bandwidth = '0', down_bandwidth = '0', delay_ms = '0', packet_loss_rate = '0'): """Start shaping traffic. |
logging.basicConfig(level=log_level) | logging.basicConfig(level=log_level, format='%(asctime)s %(levelname)s %(message)s') | def format_description(self, description): if description: return description + '\n' else: return '' |
def __call_(self, request, headers): | def __call__(self, request, headers): | def __call_(self, request, headers): logging.debug('RealHttpRequest: %s %s', request.host, request.path) host_ip = self._real_dns_lookup(request.host) connection = httplib.HTTPConnection(host_ip) connection.request( request.command, request.path, request.request_body, headers) response = connection.getresponse() connection.close() return response |
'queue', queue_size | def __enter__(self): if self.is_traffic_shaping: self.platformsettings.ipfw(['delete', self._RULE_SET]) if (self.up_bandwidth == '0' and self.down_bandwidth == '0' and self.delay_ms == '0' and self.packet_loss_rate == '0'): return | |
'dst-port', '53' | 'dst-port', '53', | def __enter__(self): if self.is_traffic_shaping: self.platformsettings.ipfw(['delete', self._RULE_SET]) if (self.up_bandwidth == '0' and self.down_bandwidth == '0' and self.delay_ms == '0' and self.packet_loss_rate == '0'): return |
'mask', 'src-port', '0xffff' | 'mask', 'src-port', '0xffff', | def __enter__(self): if self.is_traffic_shaping: self.platformsettings.ipfw(['delete', self._RULE_SET]) if (self.up_bandwidth == '0' and self.down_bandwidth == '0' and self.delay_ms == '0' and self.packet_loss_rate == '0'): return |
'pipe', self._UPLOAD_PIPE, | 'queue', self._UPLOAD_QUEUE, | def __enter__(self): if self.is_traffic_shaping: self.platformsettings.ipfw(['delete', self._RULE_SET]) if (self.up_bandwidth == '0' and self.down_bandwidth == '0' and self.delay_ms == '0' and self.packet_loss_rate == '0'): return |
'dst-port', '80' | 'dst-port', '80', | def __enter__(self): if self.is_traffic_shaping: self.platformsettings.ipfw(['delete', self._RULE_SET]) if (self.up_bandwidth == '0' and self.down_bandwidth == '0' and self.delay_ms == '0' and self.packet_loss_rate == '0'): return |
'mask', 'dst-port', '0xffff' | 'mask', 'dst-port', '0xffff', | def __enter__(self): if self.is_traffic_shaping: self.platformsettings.ipfw(['delete', self._RULE_SET]) if (self.up_bandwidth == '0' and self.down_bandwidth == '0' and self.delay_ms == '0' and self.packet_loss_rate == '0'): return |
'pipe', self._DOWNLOAD_PIPE, | 'queue', self._DOWNLOAD_QUEUE, | def __enter__(self): if self.is_traffic_shaping: self.platformsettings.ipfw(['delete', self._RULE_SET]) if (self.up_bandwidth == '0' and self.down_bandwidth == '0' and self.delay_ms == '0' and self.packet_loss_rate == '0'): return |
'src-port', '80' | 'src-port', '80', | def __enter__(self): if self.is_traffic_shaping: self.platformsettings.ipfw(['delete', self._RULE_SET]) if (self.up_bandwidth == '0' and self.down_bandwidth == '0' and self.delay_ms == '0' and self.packet_loss_rate == '0'): return |
"-p", str(self.config["packet_loss_rate"] / 100), | "-p", str(self.config["packet_loss_rate"] / 100.0), | def StartProxy(self): logging.debug("Starting Web-Page-Replay") log_level = "info" if self.log_level: log_level = self.log_level cmdline = [ replay_path, "-x", # Disables DNS intercepting "-d", str(self.config["download_bandwidth_kbps"]) + "KBit/s", "-u", str(self.config["upload_bandwidth_kbps"]) + "KBit/s", "-m", str(self.config["round_trip_time_ms"]), "-p", str(self.config["packet_loss_rate"] / 100), "-l", log_level, runner_cfg.replay_data_archive ] logging.debug("Starting replay proxy: %s", str(cmdline)) self.proxy_process = subprocess.Popen(cmdline) |
json_error.error = format % args self.response.out.write(json.dumps(json_error)) | json_error['error'] = format % args self.response.out.write(json.encode(json_error)) | def send_json_error(self, format, *args): """Send a fatal request error to the error log and json output.""" logging.error(format, *args) json_error = {} json_error.error = format % args self.response.out.write(json.dumps(json_error)) |
json_output.obj = test_set json_output.summaries = test_set.summaries self.response.out.write(json.dumps(json_output)) | json_output['obj'] = test_set json_output['summaries'] = [s for s in test_set.summaries] self.response.out.write(json.encode(json_output)) | def do_set(self): """Lookup a specific TestSet.""" set_id = self.request.get("id") if not set_id: self.send_json_error("Bad request, no id param") return test_set = models.TestSet.get(db.Key(set_id)) if not test_set: self.send_json_error("Could not find id: ", id) return |
json_output.obj = test_summary | json_output['obj'] = test_summary | def do_summary(self): """ Lookup a specific TestSummary""" set_id = self.request.get("id") if not set_id: self.send_json_error("Bad request, no id param") return test_summary = models.TestSummary.get(db.Key(set_id)) if not test_summary: self.send_json_error("Could not find id: ", id) return |
json_output.results = test_results self.response.out.write(json.dumps(json_output)) | json_output['results'] = [r for r in test_results] self.response.out.write(json.encode(json_output)) | def do_summary(self): """ Lookup a specific TestSummary""" set_id = self.request.get("id") if not set_id: self.send_json_error("Bad request, no id param") return test_summary = models.TestSummary.get(db.Key(set_id)) if not test_summary: self.send_json_error("Could not find id: ", id) return |
if headers.get('content-type', '').startswith('text/'): | content_type = headers.get('content-type', '') if (content_type.startswith('text/') or content_type == 'application/x-javascript'): | def remove_header(self, key): for i, (k, v) in enumerate(self.headers): if key == k: self.headers.pop(i) return |
'plr', self.packet_loss_rate | 'plr', self.packet_loss_rate, 'queue', queue_size | def __enter__(self): if self.is_traffic_shaping: self.platformsettings.ipfw(['delete', self.pipe_set]) if (self.up_bandwidth == '0' and self.down_bandwidth == '0' and self.delay_ms == '0' and self.packet_loss_rate == '0'): return try: upload_pipe = '1' # The IPFW pipe for upload rules. download_pipe = '2' # The IPFW pipe for download rules. dns_pipe = '3' # The IPFW pipe for DNS. |
'queue', '4000000' | 'queue', queue_size | def __enter__(self): if self.is_traffic_shaping: self.platformsettings.ipfw(['delete', self.pipe_set]) if (self.up_bandwidth == '0' and self.down_bandwidth == '0' and self.delay_ms == '0' and self.packet_loss_rate == '0'): return try: upload_pipe = '1' # The IPFW pipe for upload rules. download_pipe = '2' # The IPFW pipe for download rules. dns_pipe = '3' # The IPFW pipe for DNS. |
self._ipfw([ | self.platformsettings.ipfw([ | def __enter__(self): if self.is_traffic_shaping: self.platformsettings.ipfw(['delete', self.pipe_set]) if (self.up_bandwidth == '0' and self.down_bandwidth == '0' and self.delay_ms == '0' and self.packet_loss_rate == '0'): return try: upload_pipe = '1' # The IPFW pipe for upload rules. download_pipe = '2' # The IPFW pipe for download rules. dns_pipe = '3' # The IPFW pipe for DNS. |
option_parser.add_option('-n', '--no-deterministic_script', default=True, | harness_group.add_option('-n', '--no-deterministic_script', default=True, | def format_description(self, description): if description: return description + '\n' else: return '' |
option_parser.add_option('-P', '--no-dns_private_passthrough', default=True, | harness_group.add_option('-P', '--no-dns_private_passthrough', default=True, | def format_description(self, description): if description: return description + '\n' else: return '' |
option_parser.add_option('-x', '--no-dns_forwarding', default=True, | harness_group.add_option('-x', '--no-dns_forwarding', default=True, | def format_description(self, description): if description: return description + '\n' else: return '' |
delay_ms = str(int(delay_ms) / 2) | self.delay_ms = str(int(self.delay_ms) / 2) | def __enter__(self): if self.is_traffic_shaping: self.platformsettings.ipfw(['delete', self.pipe_set]) if (self.up_bandwidth == '0' and self.down_bandwidth == '0' and self.delay_ms == '0' and self.packet_loss_rate == '0'): return try: upload_pipe = '1' # The IPFW pipe for upload rules. download_pipe = '2' # The IPFW pipe for download rules. dns_pipe = '3' # The IPFW pipe for DNS. |
if shape_dns: | if self.shape_dns: | def __enter__(self): if self.is_traffic_shaping: self.platformsettings.ipfw(['delete', self.pipe_set]) if (self.up_bandwidth == '0' and self.down_bandwidth == '0' and self.delay_ms == '0' and self.packet_loss_rate == '0'): return try: upload_pipe = '1' # The IPFW pipe for upload rules. download_pipe = '2' # The IPFW pipe for download rules. dns_pipe = '3' # The IPFW pipe for DNS. |
obj.read_bytes_kb = int(request.get('read_bytes_kb')) obj.write_bytes_kb = int(request.get('write_bytes_kb')) | obj.read_bytes_kb = int(float(request.get('read_bytes_kb'))) obj.write_bytes_kb = int(float(request.get('write_bytes_kb'))) | def ApplyStatisticsData(request, obj): """Applies statistics uploaded via the request to the object.""" obj.using_spdy = bool(request.get('using_spdy')=="CHECKED") obj.start_load_time = int(request.get('start_load_time')) obj.commit_load_time = int(request.get('commit_load_time')) obj.doc_load_time = int(request.get('doc_load_time')) obj.paint_time = int(request.get('paint_time')) obj.total_time = int(request.get('total_time')) obj.last_load_time = int(request.get('last_load_time')) obj.num_requests = int(request.get('num_requests')) obj.num_connects = int(request.get('num_connects')) obj.num_sessions = int(request.get('num_sessions')) obj.read_bytes_kb = int(request.get('read_bytes_kb')) obj.write_bytes_kb = int(request.get('write_bytes_kb')) |
d = self.process_rs(owner, content, rs) | d = self.process_rs(rs) | def collect(self, vs=None, rs=None): """ Collect data. |
d.addCallback(lambda x: len(x) > 0 and x[0][0] or None) | d.addCallback(lambda x: len(x) > 0 and (x[0][0] + 1) or None) | def age(self, ctx, lb=None, vs=None, rs=None, sorry=False): """ Return the age of the given resource. |
query that should return URL (without prefix) of matching results. The URL will be prefixed by API version. | query that should return URL of matching results. | def childFactory(self, ctx, name): """ Dispatch the search to the generic search handler. """ return SearchGenericResource(self.dbpool, name) |
Run the search query and prefix the results with API URL. | Run the search query. | def search(self, ctx): """ Run the search query and prefix the results with API URL. """ api = IApiVersion(ctx) d = self.dbpool.runQueryInPast(ctx, self.query(), {'term': self.term}) d.addCallback(lambda results: ["/api/%s/%s" % (api, y[0]) for y in results]) return d |
api = IApiVersion(ctx) | def search(self, ctx): """ Run the search query and prefix the results with API URL. """ api = IApiVersion(ctx) d = self.dbpool.runQueryInPast(ctx, self.query(), {'term': self.term}) d.addCallback(lambda results: ["/api/%s/%s" % (api, y[0]) for y in results]) return d | |
self.query(), {'term': self.term}) d.addCallback(lambda results: ["/api/%s/%s" % (api, y[0]) for y in results]) | self.query(), {'term': self.term}) d.addCallback(lambda results: [y[0] for y in results]) | def search(self, ctx): """ Run the search query and prefix the results with API URL. """ api = IApiVersion(ctx) d = self.dbpool.runQueryInPast(ctx, self.query(), {'term': self.term}) d.addCallback(lambda results: ["/api/%s/%s" % (api, y[0]) for y in results]) return d |
SELECT 'loadbalancer/' || name || '/' | SELECT '/loadbalancer/' || name || '/' | def query(self): return """ |
SELECT 'loadbalancer/' || lb || '/virtualserver/' || vs || '/' | SELECT '/loadbalancer/' || lb || '/virtualserver/' || vs || '/' | def query(self): return """ |
SELECT 'loadbalancer/' || lb || '/virtualserver/' || vs || '/realserver/' || rs || '/' | SELECT '/loadbalancer/' || lb || '/virtualserver/' || vs || '/realserver/' || rs || '/' | def query(self): return """ |
mo = re.match(r"[rb](\d+)", rs) | mo = re.match(r"([rb])(\d+)", rs) | def parse(self, vs=None, rs=None): """ Parse vs and rs into v, s, g, r. """ if vs is not None: mo = re.match(r"v(\d+)s(\d+)g(\d+)", vs) if not mo: raise ValueError("%r is not a valid virtual server" % vs) v, s, g = int(mo.group(1)), int(mo.group(2)), int(mo.group(3)) if rs is not None: mo = re.match(r"[rb](\d+)", rs) if not mo: raise ValueError("%r is not a valid real server" % rs) r = int(mo.group(1)) return v, s, g, r return v, s, g, None return None, None, None, None |
r = int(mo.group(1)) return v, s, g, r return v, s, g, None return None, None, None, None | r = int(mo.group(2)) return v, s, g, r, mo.group(1) == "b" return v, s, g, None, None return None, None, None, None, None | def parse(self, vs=None, rs=None): """ Parse vs and rs into v, s, g, r. """ if vs is not None: mo = re.match(r"v(\d+)s(\d+)g(\d+)", vs) if not mo: raise ValueError("%r is not a valid virtual server" % vs) v, s, g = int(mo.group(1)), int(mo.group(2)), int(mo.group(3)) if rs is not None: mo = re.match(r"[rb](\d+)", rs) if not mo: raise ValueError("%r is not a valid real server" % rs) r = int(mo.group(1)) return v, s, g, r return v, s, g, None return None, None, None, None |
v, s, g, r = self.parse(vs, rs) | v, s, g, r, backup = self.parse(vs, rs) | def collect(self, vs=None, rs=None): """ Collect data for an Alteon """ v, s, g, r = self.parse(vs, rs) if v is not None: if r is not None: # Collect data to refresh a specific real server d = self.process_rs(v, s, g, r) else: # Collect data to refresh a virtual server d = self.process_vs(v, s, g) else: # Otherwise, collect everything d = self.process_all() return d |
d = self.process_rs(v, s, g, r) | d = self.process_rs(v, s, g, r, backup) | def collect(self, vs=None, rs=None): """ Collect data for an Alteon """ v, s, g, r = self.parse(vs, rs) if v is not None: if r is not None: # Collect data to refresh a specific real server d = self.process_rs(v, s, g, r) else: # Collect data to refresh a virtual server d = self.process_vs(v, s, g) else: # Otherwise, collect everything d = self.process_all() return d |
v, s, g, r = self.parse(vs, rs) | v, s, g, r, backup = self.parse(vs, rs) if backup is True: yield {} return | def actions(self, vs=None, rs=None): """ List possible actions. |
v, s, g, r = self.parse(vs, rs) | v, s, g, r, backup = self.parse(vs, rs) if backup is True: yield None return | def execute(self, action, actionargs=None, vs=None, rs=None): """ Execute an action. |
g = defer.waitForDeferred( | gr = defer.waitForDeferred( | def process_vs(self, v, s, g): """ Process data for a given virtual server when no real server is provided |
yield g g.getResult() | yield gr gr.getResult() | def process_vs(self, v, s, g): """ Process data for a given virtual server when no real server is provided |
mo = re.match(r"(.*)|(.*)", vs) | mo = re.match(r"(.*)\|(.*)", vs) | def parse(self, vs=None, rs=None): """ Parse vs and rs into owner, content, rs """ if vs is not None: mo = re.match(r"(.*)|(.*)", vs) if not mo: raise ValueError("%r is not a valid virtual server" % vs) owner, content = mo.groups() return owner, content, rs return None, None, None |
d = self.process_rs(rs) | d = self.process_rs(owner, content, rs, None) | def collect(self, vs=None, rs=None): """ Collect data. |
rs = defer.waitForDeferred(self.process_rs(service)) | rs = defer.waitForDeferred(self.process_rs(owner, content, service)) | def process_vs(self, owner, content): """ Process data for a given virtual server when no real server is provided |
rs = defer.waitForDeferred(self.process_rs(service, backup)) | rs = defer.waitForDeferred(self.process_rs(owner, content, service, backup)) | def process_vs(self, owner, content): """ Process data for a given virtual server when no real server is provided |
def process_rs(self, service, backup=False): | def process_rs(self, owner, content, service, backup=False): | def process_rs(self, service, backup=False): """ Process data for a given virtual server and real server. |
vip = "mark%d:%d" % tuple(self.cache(('virtualServerFwMark', v), ('virtualServerPort', v))) | vip = "mark%d:0" % self.cache(('virtualServerFwMark', v)) | def process_vs(self, v): """ Process data for a given virtual server when no real server is provided |
def nogateway(x): | def nogateway(): | def nogateway(x): request.setResponseCode(504) # Gateway timeout return "No gateway available" |
d.addBoth(lambda x: x is not None and collector.collect(vs, rs) and x or x) | d.addBoth(lambda x: refresh(x)) | def execute_and_refresh(collector): d = collector.execute(action, actionargs, vs, rs) # We refresh only if the result is not None (action has # been executed) We don't alter the original result. Don't # refresh a whole load balancer. if vs is not None or rs is not None: d.addBoth(lambda x: x is not None and collector.collect(vs, rs) and x or x) return d |
for rstate in state[1:]: | for rstate in states[1:]: | def aggregate_state(states): if not states: return "ok" state = states[0] for rstate in state[1:]: if rstate == "ok": if cur == "disabled": state = "ok" continue if cur == "down": state = "degraded" continue continue if rstate == "disabled": continue if rstate == "down": if cur == "ok": state = "degraded" continue if cur == "disabled": state = "down" continue continue return state |
if cur == "disabled": | if state == "disabled": | def aggregate_state(states): if not states: return "ok" state = states[0] for rstate in state[1:]: if rstate == "ok": if cur == "disabled": state = "ok" continue if cur == "down": state = "degraded" continue continue if rstate == "disabled": continue if rstate == "down": if cur == "ok": state = "degraded" continue if cur == "disabled": state = "down" continue continue return state |
if cur == "down": | if state == "down": | def aggregate_state(states): if not states: return "ok" state = states[0] for rstate in state[1:]: if rstate == "ok": if cur == "disabled": state = "ok" continue if cur == "down": state = "degraded" continue continue if rstate == "disabled": continue if rstate == "down": if cur == "ok": state = "degraded" continue if cur == "disabled": state = "down" continue continue return state |
if cur == "ok": | if state == "ok": | def aggregate_state(states): if not states: return "ok" state = states[0] for rstate in state[1:]: if rstate == "ok": if cur == "disabled": state = "ok" continue if cur == "down": state = "degraded" continue continue if rstate == "disabled": continue if rstate == "down": if cur == "ok": state = "degraded" continue if cur == "disabled": state = "down" continue continue return state |
fc_name, fc_date = get_flowcell_info(fc_dir) | def main(config_file, fc_dir, run_info_yaml=None): work_dir = os.getcwd() with open(config_file) as in_handle: config = yaml.load(in_handle) if run_info_yaml: with open(run_info_yaml) as in_handle: run_details = yaml.load(in_handle) run_info = dict(details=run_details, run_id="") else: galaxy_api = GalaxyApiAccess(config['galaxy_url'], config['galaxy_api_key']) run_info = galaxy_api.run_details(fc_name) fc_name, fc_date = get_flowcell_info(fc_dir) run_items = _add_multiplex_to_control(run_info["details"]) fastq_dir = get_fastq_dir(fc_dir) align_dir = os.path.join(work_dir, "alignments") # process each flowcell lane pool = (Pool(config["algorithm"]["num_cores"]) if config["algorithm"]["num_cores"] > 1 else None) map_fn = pool.map if pool else map try: map_fn(_process_lane_wrapper, ((i, fastq_dir, fc_name, fc_date, align_dir, config, config_file) for i in run_items)) except: if pool: pool.terminate() raise # process samples, potentially multiplexed across multiple lanes sample_files, sample_fastq, sample_info = organize_samples(align_dir, fastq_dir, work_dir, fc_name, fc_date, run_items) try: map_fn(_process_sample_wrapper, ((name, sample_fastq[name], sample_info[name], bam_files, work_dir, config, config_file) for name, bam_files in sample_files)) except: if pool: pool.terminate() raise write_metrics(run_info, work_dir, fc_dir, fc_name, fc_date, fastq_dir) | |
os.makedirs(align_dir) | try: os.makedirs(align_dir) except OSError: pass assert os.path.exists(align_dir) | def do_alignment(fastq1, fastq2, align_ref, sam_ref, lane_name, sample_name, align_dir, config, config_file): """Align to the provided reference genome, returning an aligned SAM file. """ aligner_to_use = config["algorithm"]["aligner"] if not os.path.exists(align_dir): os.makedirs(align_dir) print lane_name, "Aligning with", aligner_to_use if aligner_to_use == "bowtie": sam_file = bowtie_to_sam(fastq1, fastq2, align_ref, lane_name, align_dir, config) elif aligner_to_use == "bwa": sam_file = bwa_align_to_sam(fastq1, fastq2, align_ref, lane_name, align_dir, config) elif aligner_to_use == "maq": sam_file = maq_align_to_sam(fastq1, fastq2, align_ref, lane_name, sample_name, align_dir, config_file) else: raise ValueError("Do not recognize aligner: %s" % aligner_to_use) print lane_name, "Converting to sorted BAM file" sam_to_sort_bam(sam_file, sam_ref, fastq1, fastq2, sample_name, lane_name, config_file) |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.