bugged stringlengths 4 228k | fixed stringlengths 0 96.3M | __index_level_0__ int64 0 481k |
|---|---|---|
def perform_query (self, query, changeset, tickets, wiki, page=0): keywords = query.split(' ') | def perform_query (self, query, changeset, tickets, wiki, page=0): keywords = query.split(' ') | 26,100 |
def _render_diff(self, req, repos, chgset, diff_options): """Raw Unified Diff version""" req.send_response(200) req.send_header('Content-Type', 'text/plain;charset=utf-8') req.send_header('Content-Disposition', 'filename=Changeset%s.diff' % req.args.get('rev')) req.end_headers() | def _render_diff(self, req, repos, chgset, diff_options): """Raw Unified Diff version""" req.send_response(200) req.send_header('Content-Type', 'text/plain;charset=utf-8') req.send_header('Content-Disposition', 'inline;' 'filename=Changeset%s.diff' % chgset.rev) req.end_headers() | 26,101 |
def _render_zip(self, req, repos, chgset): """ZIP archive with all the added and/or modified files.""" req.send_response(200) req.send_header('Content-Type', 'application/zip') req.send_header('Content-Disposition', 'filename=Changeset%s.zip' % chgset.rev) req.end_headers() | def _render_zip(self, req, repos, chgset): """ZIP archive with all the added and/or modified files.""" req.send_response(200) req.send_header('Content-Type', 'application/zip') req.send_header('Content-Disposition', 'attachment;' 'filename=Changeset%s.zip' % chgset.rev) req.end_headers() | 26,102 |
def populate_hdf(hdf, env, req=None): from trac import __version__ from time import gmtime, localtime, strftime hdf['trac'] = { 'version': __version__, 'time': strftime('%c', localtime()), 'time.gmt': strftime('%a, %d %b %Y %H:%M:%S GMT', gmtime()) } hdf['trac.href'] = { 'wiki': env.href.wiki(), 'browser': env.href.bro... | def populate_hdf(hdf, env, req=None): from trac import __version__ from time import gmtime, localtime, strftime hdf['trac'] = { 'version': __version__, 'time': strftime('%c', localtime()), 'time.gmt': strftime('%a, %d %b %Y %H:%M:%S GMT', gmtime()) } hdf['trac.href'] = { 'wiki': env.href.wiki(), 'browser': env.href.bro... | 26,103 |
def process(self, req, text, in_paragraph=False): if self.error: text = system_message(Markup('Error: Failed to load processor ' '<code>%s</code>', self.name), self.error) else: text = self.processor(req, text) if in_paragraph: content_for_span = None interrupt_paragraph = False if isinstance(text, Element): tagname = ... | def process(self, req, text, in_paragraph=False): if self.error: text = system_message(Markup('Error: Failed to load processor ' '<code>%s</code>', self.name), self.error) else: text = self.processor(req, text) if in_paragraph: content_for_span = None interrupt_paragraph = False if isinstance(text, Element): tagname = ... | 26,104 |
def process(self, req, text, in_paragraph=False): if self.error: text = system_message(Markup('Error: Failed to load processor ' '<code>%s</code>', self.name), self.error) else: text = self.processor(req, text) if in_paragraph: content_for_span = None interrupt_paragraph = False if isinstance(text, Element): tagname = ... | def process(self, req, text, in_paragraph=False): if self.error: text = system_message(Markup('Error: Failed to load processor ' '<code>%s</code>', self.name), self.error) else: text = self.processor(req, text) if in_paragraph: content_for_span = None interrupt_paragraph = False if isinstance(text, Element): tagname = ... | 26,105 |
def simple_tag_handler(self, open_tag, close_tag): """Generic handler for simple binary style tags""" if self.tag_open_p((open_tag, close_tag)): return self.close_tag(close_tag) else: self.open_tag(open_tag, close_tag) return open_tag | def simple_tag_handler(self, match, open_tag, close_tag): """Generic handler for simple binary style tags""" if self.tag_open_p((open_tag, close_tag)): return self.close_tag(close_tag) else: self.open_tag(open_tag, close_tag) return open_tag | 26,106 |
def _bold_formatter(self, match, fullmatch): return self.simple_tag_handler('<strong>', '</strong>') | def _bold_formatter(self, match, fullmatch): return self.simple_tag_handler('<strong>', '</strong>') | 26,107 |
def _italic_formatter(self, match, fullmatch): return self.simple_tag_handler('<i>', '</i>') | def _italic_formatter(self, match, fullmatch): return self.simple_tag_handler('<i>', '</i>') | 26,108 |
def _underline_formatter(self, match, fullmatch): if match[0] == '!': return match[1:] else: return self.simple_tag_handler('<span class="underline">', '</span>') | def _underline_formatter(self, match, fullmatch): if match[0] == '!': return match[1:] else: return self.simple_tag_handler('<span class="underline">', '</span>') | 26,109 |
def _strike_formatter(self, match, fullmatch): if match[0] == '!': return match[1:] else: return self.simple_tag_handler('<del>', '</del>') | def _strike_formatter(self, match, fullmatch): if match[0] == '!': return match[1:] else: return self.simple_tag_handler('<del>', '</del>') | 26,110 |
def _subscript_formatter(self, match, fullmatch): if match[0] == '!': return match[1:] else: return self.simple_tag_handler('<sub>', '</sub>') | def _subscript_formatter(self, match, fullmatch): if match[0] == '!': return match[1:] else: return self.simple_tag_handler('<sub>', '</sub>') | 26,111 |
def _superscript_formatter(self, match, fullmatch): if match[0] == '!': return match[1:] else: return self.simple_tag_handler('<sup>', '</sup>') | def _superscript_formatter(self, match, fullmatch): if match[0] == '!': return match[1:] else: return self.simple_tag_handler('<sup>', '</sup>') | 26,112 |
def save(self): """Write the configuration options to the primary file.""" if not self.filename: return | def save(self): """Write the configuration options to the primary file.""" if not self.filename: return | 26,113 |
def has_site_option(self, section, name): return self.site_parser.has_option(section, name) | def has_site_option(self, section, name): return self.site_parser.has_option(section, name) | 26,114 |
def __init__(self, config, name): self.config = config self.name = name self.overriden = {} | def __init__(self, config, name): self.config = config self.name = name self.overriden = {} | 26,115 |
def set(self, name, value): """Change a configuration value. These changes are not persistent unless saved with `save()`. """ if not self.config.parser.has_section(self.name): self.config.parser.add_section(self.name) if value is None: self.overriden[name] = True value = '' else: value = to_unicode(value).encode('utf-... | def set(self, name, value): """Change a configuration value. These changes are not persistent unless saved with `save()`. """ if not self.config.parser.has_section(self.name): self.config.parser.add_section(self.name) if value is None: self.overridden[name] = True value = '' else: value = to_unicode(value).encode('utf... | 26,116 |
def _get_changes(env, repos, revs, full=None, req=None, format=None): db = env.get_db_cnx() changes = {} for rev in revs: changeset = repos.get_changeset(rev) message = changeset.message files = None if format == 'changelog': files = [change[0] for change in changeset.get_changes()] elif message: if not full: message =... | def _get_changes(env, repos, revs, full=None, req=None, format=None): db = env.get_db_cnx() changes = {} for rev in revs: changeset = repos.get_changeset(rev) message = changeset.message files = None if format == 'changelog': files = [change[0] for change in changeset.get_changes()] elif message: if not full: message =... | 26,117 |
def _render_confirm(self, req, attachment): perm_map = {'ticket': 'TICKET_ADMIN', 'wiki': 'WIKI_DELETE'} req.perm.assert_permission(perm_map[attachment.parent_type]) | def _render_confirm(self, req, attachment): perm_map = {'ticket': 'TICKET_ADMIN', 'wiki': 'WIKI_DELETE'} req.perm.assert_permission(perm_map[attachment.parent_type]) | 26,118 |
def __init__(self, path, user=None, password=None, host=None, port=None, params={}): import MySQLdb | def __init__(self, path, user=None, password=None, host=None, port=None, params={}): import MySQLdb | 26,119 |
def promote_session(self, sid): """ Promotes an anonymous session to an authenticated session, if there is no preexisting session data for that user name. """ assert self.req.authname != 'anonymous', \ 'Cannot promote session of anonymous user' | def promote_session(self, sid): """ Promotes an anonymous session to an authenticated session, if there is no preexisting session data for that user name. """ assert self.req.authname != 'anonymous', \ 'Cannot promote session of anonymous user' | 26,120 |
def render (self): if self.args.has_key('create'): self.perm.assert_permission(perm.TICKET_CREATE) self.create_ticket() | def render (self): if self.args.has_key('create'): self.create_ticket() | 26,121 |
def writeline(self, text): match = header_re.search(text) if match: self.hdf.setValue('%s.name.old' % self.prefix, match.group(1)) self.hdf.setValue('%s.name.new' % self.prefix, match.group(2)) return self.count = self.count + 1 if self.count < 3: return match = line_re.search(text) if match: pfx = '%s.changes.%d' % (s... | def writeline(self, text): match = header_re.search(text) if match: self.hdf.setValue('%s.name.old' % self.prefix, match.group(1)) self.hdf.setValue('%s.name.new' % self.prefix, match.group(2)) return self.count = self.count + 1 if self.count < 3: return match = line_re.search(text) if match: self.changeno += 1 pfx = '... | 26,122 |
def writeline(self, text): match = header_re.search(text) if match: self.hdf.setValue('%s.name.old' % self.prefix, match.group(1)) self.hdf.setValue('%s.name.new' % self.prefix, match.group(2)) return self.count = self.count + 1 if self.count < 3: return match = line_re.search(text) if match: pfx = '%s.changes.%d' % (s... | def writeline(self, text): match = header_re.search(text) if match: self.hdf.setValue('%s.name.old' % self.prefix, match.group(1)) self.hdf.setValue('%s.name.new' % self.prefix, match.group(2)) return self.count = self.count + 1 if self.count < 3: return match = line_re.search(text) if match: pfx = '%s.changes.%d' % (s... | 26,123 |
def render(self): self.perm.assert_permission(perm.TICKET_VIEW) | def render(self): self.perm.assert_permission(perm.TICKET_VIEW) | 26,124 |
def render(self): self.perm.assert_permission(perm.TICKET_VIEW) | def render(self): self.perm.assert_permission(perm.TICKET_VIEW) | 26,125 |
def process(self, req, text, in_paragraph=False): if self.error: return system_message(Markup('Error: Failed to load processor ' '<code>%s</code>', self.name), self.error) text = self.processor(req, text) if in_paragraph: content_for_span = None interrupt_paragraph = False if isinstance(text, Element): tagname = text.t... | def process(self, req, text, in_paragraph=False): if self.error: return system_message(Markup('Error: Failed to load processor ' '<code>%s</code>', self.name), self.error) text = self.processor(req, text) if in_paragraph: content_for_span = None interrupt_paragraph = False if isinstance(text, Element): tagname = text.t... | 26,126 |
def validate_ticket(req, ticket): """Validate a ticket after it's been populated from user input. Must return a list of `(field, message)` tuples, one for each problem detected. `field` can be `None` to indicate an overall problem with the ticket. Therefore, a return value of `[]` means everything is OK.""" | def validate_ticket(req, ticket): """Validate a ticket after it's been populated from user input. Must return a list of `(field, message)` tuples, one for each problem detected. `field` can be `None` to indicate an overall problem with the ticket. Therefore, a return value of `[]` means everything is OK.""" | 26,127 |
def send_file(self, path, mimetype=None): """ Send a local file to the browser. This method includes the "Last-Modified", "Content-Type" and "Content-Length" headers in the response, corresponding to the file attributes. It also checks the last modification time of the local file against the "If-Modified-Since" provide... | def send_file(self, path, mimetype=None): """ Send a local file to the browser. This method includes the "Last-Modified", "Content-Type" and "Content-Length" headers in the response, corresponding to the file attributes. It also checks the last modification time of the local file against the "If-Modified-Since" provide... | 26,128 |
def _render_view(self, req, db, page): req.perm.assert_permission('WIKI_VIEW') | def _render_view(self, req, db, page): req.perm.assert_permission('WIKI_VIEW') | 26,129 |
def format(self, text, out): if not text: return self.out = out self._open_tags = [] | def format(self, text, out): if not text: return self.out = out self._open_tags = [] | 26,130 |
def _heading_formatter(self, match, fullmatch): match = match.strip() self.close_table() self.close_paragraph() self.close_indentation() self.close_list() | def _heading_formatter(self, match, fullmatch): match = match.strip() self.close_table() self.close_paragraph() self.close_indentation() self.close_list() | 26,131 |
def _render_diff(self, req, pagename, version): # Stores the diff-style in the session if it has been changed, and adds # diff-style related item to the HDF self.perm.assert_permission(perm.WIKI_VIEW) | def _render_diff(self, req, pagename, version): # Stores the diff-style in the session if it has been changed, and adds # diff-style related item to the HDF self.perm.assert_permission(perm.WIKI_VIEW) | 26,132 |
def get_environment(req, mpr): global env_cache, env_cache_lock options = req.get_options() if not options.has_key('TracEnv') and not options.has_key('TracEnvParentDir'): raise EnvironmentError, \ 'Missing PythonOption "TracEnv" or "TracEnvParentDir". Trac '\ 'requires one of these options to locate the Trac environme... | def get_environment(req, mpr): global env_cache, env_cache_lock options = req.get_options() if not options.has_key('TracEnv') and not options.has_key('TracEnvParentDir'): raise EnvironmentError, \ 'Missing PythonOption "TracEnv" or "TracEnvParentDir". Trac '\ 'requires one of these options to locate the Trac environme... | 26,133 |
def display_html(self, req, query): req.hdf['title'] = 'Custom Query' add_stylesheet(req, 'css/report.css') | def display_html(self, req, query): req.hdf['title'] = 'Custom Query' add_stylesheet(req, 'css/report.css') | 26,134 |
def display_html(self, req, query): req.hdf['title'] = 'Custom Query' add_stylesheet(req, 'css/report.css') | def display_html(self, req, query): req.hdf['title'] = 'Custom Query' add_stylesheet(req, 'css/report.css') | 26,135 |
def get_changes(self): pool = Pool(self.pool) tmp = Pool(pool) root = fs.revision_root(self.fs_ptr, self.rev, pool()) editor = repos.RevisionChangeCollector(self.fs_ptr, self.rev, pool()) e_ptr, e_baton = delta.make_editor(editor, pool()) repos.svn_repos_replay(root, e_ptr, e_baton, pool()) | def get_changes(self): pool = Pool(self.pool) tmp = Pool(pool) root = fs.revision_root(self.fs_ptr, self.rev, pool()) editor = repos.RevisionChangeCollector(self.fs_ptr, self.rev, pool()) e_ptr, e_baton = delta.make_editor(editor, pool()) repos.svn_repos_replay(root, e_ptr, e_baton, pool()) | 26,136 |
def get_changes(self): pool = Pool(self.pool) tmp = Pool(pool) root = fs.revision_root(self.fs_ptr, self.rev, pool()) editor = repos.RevisionChangeCollector(self.fs_ptr, self.rev, pool()) e_ptr, e_baton = delta.make_editor(editor, pool()) repos.svn_repos_replay(root, e_ptr, e_baton, pool()) | def get_changes(self): pool = Pool(self.pool) tmp = Pool(pool) root = fs.revision_root(self.fs_ptr, self.rev, pool()) editor = repos.RevisionChangeCollector(self.fs_ptr, self.rev, pool()) e_ptr, e_baton = delta.make_editor(editor, pool()) repos.svn_repos_replay(root, e_ptr, e_baton, pool()) | 26,137 |
def _render_view(self, req, attachment): perm_map = {'ticket': 'TICKET_VIEW', 'wiki': 'WIKI_VIEW'} req.perm.assert_permission(perm_map[attachment.parent_type]) | def _render_view(self, req, attachment): perm_map = {'ticket': 'TICKET_VIEW', 'wiki': 'WIKI_VIEW'} req.perm.assert_permission(perm_map[attachment.parent_type]) | 26,138 |
def load(self, filename): fd = open(filename, 'r') for line in fd: u, h = line.strip().split(':') if '$' in h or self.crypt: self.hash[u] = h else: print >>sys.stderr, 'Warning: cannot parse password for ' \ 'user "%s" without the "crypt" module' % u | def load(self, filename): fd = open(filename, 'r') for line in fd: line = line.strip() if not line: continue try: u, h = line.split(':') except ValueError: print >>sys.stderr, 'Warning: invalid password line in %s: %s' \ % (filename, line) continue if '$' in h or self.crypt: self.hash[u] = h else: print >>sys.stderr, '... | 26,139 |
def load_htdigest(self, filename, realm): """Load account information from apache style htdigest files, only users from the specified realm are used """ fd = open(filename, 'r') for line in fd.readlines(): u, r, a1 = line.strip().split(':') if r == realm: self.hash[u] = a1 if self.hash == {}: print >> sys.stderr, "Warn... | def load_htdigest(self, filename, realm): """Load account information from apache style htdigest files, only users from the specified realm are used """ fd = open(filename, 'r') for line in fd.readlines(): line = line.strip() if not line: continue try: u, r, a1 = line.split(':') except ValueError: print >>sys.stderr, '... | 26,140 |
def get_templates_dirs(): """Return a list of directories containing the provided ClearSilver templates. """ | def get_templates_dirs(): """Return a list of directories containing the provided ClearSilver templates. """ | 26,141 |
def get_templates_dirs(): """Return a list of directories containing the provided ClearSilver templates. """ | def get_templates_dirs(): """Return a list of directories containing the provided ClearSilver templates. """ | 26,142 |
def populate_hdf(self, req, handler): """Add chrome-related data to the HDF.""" | def populate_hdf(self, req, handler): """Add chrome-related data to the HDF.""" | 26,143 |
def process(self, req, text, inline=False): if self.error: return system_message('Error: Failed to load processor <code>%s</code>' % self.name, self.error) text = self.processor(req, text, self.env) if inline: code_block_start = re.compile('^<div class="code-block">') code_block_end = re.compile('</div>$') text, nr = c... | def process(self, req, text, inline=False): if self.error: return system_message('Error: Failed to load processor <code>%s</code>' % self.name, self.error) text = self.processor(req, text, self.env) if inline: code_block_start = re.compile('^<div class="code-block">') code_block_end = re.compile('</div>$') text, nr = c... | 26,144 |
def _lhref_formatter(self, match, fullmatch): ns = fullmatch.group('lns') target = fullmatch.group('ltgt') label = fullmatch.group('label') if ns in self.link_resolvers: return self._link_resolvers[ns](self, ns, target, label) elif target[:2] == '//': return self._make_ext_link(ns+':'+target, label) else: return match | def _lhref_formatter(self, match, fullmatch): ns = fullmatch.group('lns') target = fullmatch.group('ltgt') label = fullmatch.group('label') or target if ns in self.link_resolvers: return self._link_resolvers[ns](self, ns, target, label) elif target[:2] == '//': return self._make_ext_link(ns+':'+target, label) else: ret... | 26,145 |
def query(self, constraints, order=None, desc=0): href = href_join(self.base, 'query') params = [] for field in constraints: values = constraints[field] if type(values) != list: values = [values] for value in values: params.append(field + '=' + urllib.quote(value)) if order: params.append('order=' + urllib.quote(order)... | def query(self, constraints, order=None, desc=0): href = href_join(self.base, 'query') params = [] for field in constraints.keys(): values = constraints[field] if type(values) != list: values = [values] for value in values: params.append(field + '=' + urllib.quote(value)) if order: params.append('order=' + urllib.quote... | 26,146 |
def get_custom_fields(env): cfg = env.config.options('ticket-custom') if not cfg: return [] names = [] items = {} for k, v in cfg: items[k] = v if '.' not in k: names.append(k) fields = [] for name in names: field = { 'name': name, 'type': items[name], 'order': items.get(name + '.order', '0'), 'label': items.get(name +... | def get_custom_fields(env): cfg = env.config.options('ticket-custom') if not cfg: return [] names = [] items = {} for k, v in cfg: items[k] = v if '.' not in k: names.append(k) fields = [] for name in names: field = { 'name': name, 'type': items[name], 'order': items.get(name + '.order', '0'), 'label': items.get(name +... | 26,147 |
def insert_custom_fields(env, hdf, vals = {}): fields = get_custom_fields(env) i = 0 for f in fields: name = f['name'] val = vals.get('custom_' + name, f['value']) pfx = 'ticket.custom.%i' % i hdf['%s.name' % pfx] = f['name'] hdf['%s.type' % pfx] = f['type'] hdf['%s.label' % pfx] = f['label'] or f['name'] hdf['%s.value... | def insert_custom_fields(env, hdf, vals = {}): fields = get_custom_fields(env) i = 0 for f in fields: name = f['name'] val = vals.get('custom_' + name, f['value']) pfx = 'ticket.custom.%d' % i hdf['%s.name' % pfx] = f['name'] hdf['%s.type' % pfx] = f['type'] hdf['%s.label' % pfx] = f['label'] or f['name'] hdf['%s.value... | 26,148 |
def insert_custom_fields(env, hdf, vals = {}): fields = get_custom_fields(env) i = 0 for f in fields: name = f['name'] val = vals.get('custom_' + name, f['value']) pfx = 'ticket.custom.%i' % i hdf['%s.name' % pfx] = f['name'] hdf['%s.type' % pfx] = f['type'] hdf['%s.label' % pfx] = f['label'] or f['name'] hdf['%s.value... | def insert_custom_fields(env, hdf, vals = {}): fields = get_custom_fields(env) i = 0 for f in fields: name = f['name'] val = vals.get('custom_' + name, f['value']) pfx = 'ticket.custom.%i' % i hdf['%s.name' % pfx] = f['name'] hdf['%s.type' % pfx] = f['type'] hdf['%s.label' % pfx] = f['label'] or f['name'] hdf['%s.value... | 26,149 |
def delete(self, version=None, db=None): assert self.exists, 'Cannot delete non-existent page' if not db: db = self.env.get_db_cnx() handle_ta = True else: handle_ta = False | def delete(self, version=None, db=None): assert self.exists, 'Cannot delete non-existent page' if not db: db = self.env.get_db_cnx() handle_ta = True else: handle_ta = False | 26,150 |
def real_main(): import sync import Href import perm import auth from util import redirect path_info = os.getenv('PATH_INFO') remote_addr = os.getenv('REMOTE_ADDR') remote_user = os.getenv('REMOTE_USER') http_cookie = os.getenv('HTTP_COOKIE') http_referer = os.getenv('HTTP_REFERER') cgi_location = os.getenv('SCRIPT_NA... | def real_main(): import sync import Href import perm import auth from util import redirect path_info = os.getenv('PATH_INFO') remote_addr = os.getenv('REMOTE_ADDR') remote_user = os.getenv('REMOTE_USER') http_cookie = os.getenv('HTTP_COOKIE') http_referer = os.getenv('HTTP_REFERER') cgi_location = os.getenv('SCRIPT_NA... | 26,151 |
def test_get_change_extent(self): self.assertEqual((3, 0), diff2._get_change_extent('xxx', 'xxx')) self.assertEqual((0, 0), diff2._get_change_extent('', 'xxx')) self.assertEqual((0, 0), diff2._get_change_extent('xxx', '')) self.assertEqual((0, 0), diff2._get_change_extent('xxx', 'yyy')) self.assertEqual((1, -1), diff2.... | def test_get_change_extent(self): self.assertEqual((3, 0), diff2._get_change_extent('xxx', 'xxx')) self.assertEqual((0, 0), diff2._get_change_extent('', 'xxx')) self.assertEqual((0, 0), diff2._get_change_extent('xxx', '')) self.assertEqual((0, 0), diff2._get_change_extent('xxx', 'yyy')) self.assertEqual((1, -1), diff2.... | 26,152 |
def test_insert_blank_line(self): opcodes = diff2._get_opcodes(['A', 'B'], ['A', 'B', ''], ignore_blank_lines=0) self.assertEqual(('equal', 0, 2, 0, 2), opcodes.next()) self.assertEqual(('insert', 2, 2, 2, 3), opcodes.next()) | def test_insert_blank_line(self): opcodes = Diff._get_opcodes(['A', 'B'], ['A', 'B', ''], ignore_blank_lines=0) self.assertEqual(('equal', 0, 2, 0, 2), opcodes.next()) self.assertEqual(('insert', 2, 2, 2, 3), opcodes.next()) | 26,153 |
def test_insert_blank_line(self): opcodes = diff2._get_opcodes(['A', 'B'], ['A', 'B', ''], ignore_blank_lines=0) self.assertEqual(('equal', 0, 2, 0, 2), opcodes.next()) self.assertEqual(('insert', 2, 2, 2, 3), opcodes.next()) | def test_insert_blank_line(self): opcodes = Diff._get_opcodes(['A', 'B'], ['A', 'B', ''], ignore_blank_lines=0) self.assertEqual(('equal', 0, 2, 0, 2), opcodes.next()) self.assertEqual(('insert', 2, 2, 2, 3), opcodes.next()) | 26,154 |
def test_insert_blank_line(self): opcodes = diff2._get_opcodes(['A', 'B'], ['A', 'B', ''], ignore_blank_lines=0) self.assertEqual(('equal', 0, 2, 0, 2), opcodes.next()) self.assertEqual(('insert', 2, 2, 2, 3), opcodes.next()) | def test_insert_blank_line(self): opcodes = diff2._get_opcodes(['A', 'B'], ['A', 'B', ''], ignore_blank_lines=0) self.assertEqual(('equal', 0, 2, 0, 2), opcodes.next()) self.assertEqual(('insert', 2, 2, 2, 3), opcodes.next()) | 26,155 |
def test_delete_blank_line(self): opcodes = diff2._get_opcodes(['A', 'B', ''], ['A', 'B'], ignore_blank_lines=0) self.assertEqual(('equal', 0, 2, 0, 2), opcodes.next()) self.assertEqual(('delete', 2, 3, 2, 2), opcodes.next()) | def test_delete_blank_line(self): opcodes = Diff._get_opcodes(['A', 'B', ''], ['A', 'B'], ignore_blank_lines=0) self.assertEqual(('equal', 0, 2, 0, 2), opcodes.next()) self.assertEqual(('delete', 2, 3, 2, 2), opcodes.next()) | 26,156 |
def test_delete_blank_line(self): opcodes = diff2._get_opcodes(['A', 'B', ''], ['A', 'B'], ignore_blank_lines=0) self.assertEqual(('equal', 0, 2, 0, 2), opcodes.next()) self.assertEqual(('delete', 2, 3, 2, 2), opcodes.next()) | def test_delete_blank_line(self): opcodes = Diff._get_opcodes(['A', 'B', ''], ['A', 'B'], ignore_blank_lines=0) self.assertEqual(('equal', 0, 2, 0, 2), opcodes.next()) self.assertEqual(('delete', 2, 3, 2, 2), opcodes.next()) | 26,157 |
def test_delete_blank_line(self): opcodes = diff2._get_opcodes(['A', 'B', ''], ['A', 'B'], ignore_blank_lines=0) self.assertEqual(('equal', 0, 2, 0, 2), opcodes.next()) self.assertEqual(('delete', 2, 3, 2, 2), opcodes.next()) | def test_delete_blank_line(self): opcodes = diff2._get_opcodes(['A', 'B', ''], ['A', 'B'], ignore_blank_lines=0) self.assertEqual(('equal', 0, 2, 0, 2), opcodes.next()) self.assertEqual(('delete', 2, 3, 2, 2), opcodes.next()) | 26,158 |
def test_space_changes(self): opcodes = diff2._get_opcodes(['A', 'B b'], ['A', 'B b'], ignore_space_changes=0) self.assertEqual(('equal', 0, 1, 0, 1), opcodes.next()) self.assertEqual(('replace', 1, 2, 1, 2), opcodes.next()) | def test_space_changes(self): opcodes = Diff._get_opcodes(['A', 'B b'], ['A', 'B b'], ignore_space_changes=0) self.assertEqual(('equal', 0, 1, 0, 1), opcodes.next()) self.assertEqual(('replace', 1, 2, 1, 2), opcodes.next()) | 26,159 |
def test_space_changes(self): opcodes = diff2._get_opcodes(['A', 'B b'], ['A', 'B b'], ignore_space_changes=0) self.assertEqual(('equal', 0, 1, 0, 1), opcodes.next()) self.assertEqual(('replace', 1, 2, 1, 2), opcodes.next()) | def test_space_changes(self): opcodes = Diff._get_opcodes(['A', 'B b'], ['A', 'B b'], ignore_space_changes=0) self.assertEqual(('equal', 0, 1, 0, 1), opcodes.next()) self.assertEqual(('replace', 1, 2, 1, 2), opcodes.next()) | 26,160 |
def test_case_changes(self): opcodes = diff2._get_opcodes(['A', 'B b'], ['A', 'B B'], ignore_case=0) self.assertEqual(('equal', 0, 1, 0, 1), opcodes.next()) self.assertEqual(('replace', 1, 2, 1, 2), opcodes.next()) | def test_case_changes(self): opcodes = Diff._get_opcodes(['A', 'B b'], ['A', 'B B'], ignore_case=0) self.assertEqual(('equal', 0, 1, 0, 1), opcodes.next()) self.assertEqual(('replace', 1, 2, 1, 2), opcodes.next()) | 26,161 |
def test_case_changes(self): opcodes = diff2._get_opcodes(['A', 'B b'], ['A', 'B B'], ignore_case=0) self.assertEqual(('equal', 0, 1, 0, 1), opcodes.next()) self.assertEqual(('replace', 1, 2, 1, 2), opcodes.next()) | def test_case_changes(self): opcodes = Diff._get_opcodes(['A', 'B b'], ['A', 'B B'], ignore_case=0) self.assertEqual(('equal', 0, 1, 0, 1), opcodes.next()) self.assertEqual(('replace', 1, 2, 1, 2), opcodes.next()) | 26,162 |
def test_space_and_case_changes(self): opcodes = diff2._get_opcodes(['A', 'B b'], ['A', 'B B'], ignore_case=0, ignore_space_changes=0) self.assertEqual(('equal', 0, 1, 0, 1), opcodes.next()) self.assertEqual(('replace', 1, 2, 1, 2), opcodes.next()) | def test_space_and_case_changes(self): opcodes = Diff._get_opcodes(['A', 'B b'], ['A', 'B B'], ignore_case=0, ignore_space_changes=0) self.assertEqual(('equal', 0, 1, 0, 1), opcodes.next()) self.assertEqual(('replace', 1, 2, 1, 2), opcodes.next()) | 26,163 |
def test_space_and_case_changes(self): opcodes = diff2._get_opcodes(['A', 'B b'], ['A', 'B B'], ignore_case=0, ignore_space_changes=0) self.assertEqual(('equal', 0, 1, 0, 1), opcodes.next()) self.assertEqual(('replace', 1, 2, 1, 2), opcodes.next()) | def test_space_and_case_changes(self): opcodes = Diff._get_opcodes(['A', 'B b'], ['A', 'B B'], ignore_case=0, ignore_space_changes=0) self.assertEqual(('equal', 0, 1, 0, 1), opcodes.next()) self.assertEqual(('replace', 1, 2, 1, 2), opcodes.next()) | 26,164 |
def test_grouped_opcodes_context1(self): opcodes = diff2._get_opcodes(['A', 'B', 'C', 'D', 'E', 'F', 'G', 'H'], ['A', 'B', 'C', 'd', 'e', 'f', 'G', 'H']) groups = diff2._group_opcodes(opcodes, n=1) group = groups.next() self.assertEqual(('equal', 2, 3, 2, 3), group[0]) self.assertEqual(('replace', 3, 6, 3, 6), group[1]... | def test_grouped_opcodes_context1(self): opcodes = Diff._get_opcodes(['A', 'B', 'C', 'D', 'E', 'F', 'G', 'H'], ['A', 'B', 'C', 'd', 'e', 'f', 'G', 'H']) groups = diff2._group_opcodes(opcodes, n=1) group = groups.next() self.assertEqual(('equal', 2, 3, 2, 3), group[0]) self.assertEqual(('replace', 3, 6, 3, 6), group[1])... | 26,165 |
def test_grouped_opcodes_context1(self): opcodes = diff2._get_opcodes(['A', 'B', 'C', 'D', 'E', 'F', 'G', 'H'], ['A', 'B', 'C', 'd', 'e', 'f', 'G', 'H']) groups = diff2._group_opcodes(opcodes, n=1) group = groups.next() self.assertEqual(('equal', 2, 3, 2, 3), group[0]) self.assertEqual(('replace', 3, 6, 3, 6), group[1]... | def test_grouped_opcodes_context1(self): opcodes = diff2._get_opcodes(['A', 'B', 'C', 'D', 'E', 'F', 'G', 'H'], ['A', 'B', 'C', 'd', 'e', 'f', 'G', 'H']) groups = Diff._group_opcodes(opcodes, n=1) group = groups.next() self.assertEqual(('equal', 2, 3, 2, 3), group[0]) self.assertEqual(('replace', 3, 6, 3, 6), group[1])... | 26,166 |
def wrap(t, cols=75, initial_indent='', subsequent_indent=''): try: import textwrap return '\n'.join(textwrap.wrap(t, replace_whitespace=0, width=cols, break_long_words=0, initial_indent=initial_indent, subsequent_indent=subsequent_indent)) except ImportError: return t | def wrap(t, cols=75, initial_indent='', subsequent_indent=''): try: import textwrap t = t.replace('\r\n', '\n').replace('\r', '\n') wrapper = textwrap.TextWrapper(cols, replace_whitespace=0, break_long_words=0, initial_indent=initial_indent, subsequent_indent=subsequent_indent)) except ImportError: return t | 26,167 |
def wrap(t, cols=75, initial_indent='', subsequent_indent=''): try: import textwrap return '\n'.join(textwrap.wrap(t, replace_whitespace=0, width=cols, break_long_words=0, initial_indent=initial_indent, subsequent_indent=subsequent_indent)) except ImportError: return t | def wrap(t, cols=75, initial_indent='', subsequent_indent=''): try: import textwrap return '\n'.join(textwrap.wrap(t, replace_whitespace=0, width=cols, break_long_words=0, initial_indent=initial_indent, subsequent_indent=subsequent_indent) wrappedLines = [] for line in t.split('\n'): wrappedLines += wrapper.wrap(line.r... | 26,168 |
def notify(self, tktid, newticket=1, modtime=0): | def notify(self, tktid, newticket=1, modtime=0): | 26,169 |
def notify(self, tktid, newticket=1, modtime=0): | def notify(self, tktid, newticket=1, modtime=0): | 26,170 |
def wiki_page_added(page): """ Called whenever a new Wiki page is added. """ | def wiki_page_added(page): """ Called whenever a new Wiki page is added. """ | 26,171 |
def wiki_page_changed(page, version, t, comment, author, ipnr): """ Called when a page has been modified. """ | def wiki_page_changed(page, version, t, comment, author, ipnr): """ Called when a page has been modified. """ | 26,172 |
def wiki_page_deleted(page): """ Called when a page has been deleted. """ | def wiki_page_deleted(page): """ Called when a page has been deleted. """ | 26,173 |
def wiki_page_deleted(page): """ Called when a page has been deleted. """ | def wiki_page_deleted(page): """ Called when a page has been deleted. """ | 26,174 |
def get_macros(): """ Return an iterable that provides the names of the provided macros. """ | def get_macros(): """ Return an iterable that provides the names of the provided macros. """ | 26,175 |
def get_macro_description(name): """ Return a plain text description of the macro with the specified name. """ | def get_macro_description(name): """Return a plain text description of the macro with the specified name. """ | 26,176 |
def render_macro(req, name, content): """ Return the HTML output of the macro. """ | def render_macro(req, name, content): """ Return the HTML output of the macro. """ | 26,177 |
def get_wiki_syntax(): """ Return an iterable that provides additional wiki syntax. """ | def get_wiki_syntax(): """ Return an iterable that provides additional wiki syntax. """ | 26,178 |
def get_link_resolvers(): """ Return an iterable over (namespace, formatter) tuples. """ | def get_link_resolvers(): """ Return an iterable over (namespace, formatter) tuples. """ | 26,179 |
def get_link_resolvers(): """ Return an iterable over (namespace, formatter) tuples. """ | def get_link_resolvers(): """ Return an iterable over (namespace, formatter) tuples. """ | 26,180 |
def __init__(self): self._pages = None | def __init__(self): self._pages = None | 26,181 |
def _load_pages(self): self._pages = {} db = self.env.get_db_cnx() cursor = db.cursor() cursor.execute("SELECT DISTINCT name FROM wiki") for (name,) in cursor: self._pages[name] = True | def _load_pages(self): self._pages = {} db = self.env.get_db_cnx() cursor = db.cursor() cursor.execute("SELECT DISTINCT name FROM wiki") for (name,) in cursor: self._pages[name] = True | 26,182 |
def get_pages(self, prefix=None): if self._pages is None: self._load_pages() for page in self._pages.keys(): if not prefix or page.startswith(prefix): yield page | def get_pages(self, prefix=None): """Iterate over the names of existing Wiki pages. If the `prefix` parameter is given, only names that start with that prefix are included. """ self._update_index() for page in self._index.keys(): if not prefix or page.startswith(prefix): yield page | 26,183 |
def has_page(self, pagename): if self._pages is None: self._load_pages() return pagename in self._pages.keys() | def has_page(self, pagename): if self._pages is None: self._load_pages() return pagename in self._pages.keys() | 26,184 |
def wiki_page_deleted(self, page): if self.has_page(page.name): self.log.debug('Removing page %s from index' % page.name) del self._pages[page.name] | def wiki_page_deleted(self, page): if self.has_page(page.name): self.log.debug('Removing page %s from index' % page.name) del self._pages[page.name] | 26,185 |
def get_wiki_syntax(self): yield (r"!?(^|(?<=[^A-Za-z]))[A-Z][a-z]+(?:[A-Z][a-z]*[a-z/])+(?:#[A-Za-z0-9]+)?(?=\Z|\s|[.,;:!?\)}\]])", lambda x, y, z: self._format_link(x, 'wiki', y, y)) | def get_wiki_syntax(self): yield (r"!?(^|(?<=[^A-Za-z]))[A-Z][a-z]+(?:[A-Z][a-z]*[a-z/])+(?:#[A-Za-z0-9]+)?(?=\Z|\s|[.,;:!?\)}\]])", lambda x, y, z: self._format_link(x, 'wiki', y, y)) | 26,186 |
def _format_link(self, formatter, ns, page, label): anchor = '' if page.find('#') != -1: anchor = page[page.find('#'):] page = page[:page.find('#')] page = urllib.unquote(page) label = urllib.unquote(label) | def_format_link(self,formatter,ns,page,label):anchor=''ifpage.find('#')!=-1:anchor=page[page.find('#'):]page=page[:page.find('#')]page=urllib.unquote(page)label=urllib.unquote(label) | 26,187 |
def get_attachments_hdf(self, cnx, type, id, hdf, prefix): from Wiki import wiki_to_oneliner files = self.get_attachments(cnx, type, id) idx = 0 for file in files: p = '%s.%d' % (prefix, idx) hdf.setValue(p + '.name', file['filename']) hdf.setValue(p + '.descr', wiki_to_oneliner(file['description'], hdf, self,self.get_... | def get_attachments_hdf(self, cnx, type, id, hdf, prefix): from Wiki import wiki_to_oneliner files = self.get_attachments(cnx, type, id) idx = 0 for file in files: p = '%s.%d' % (prefix, idx) hdf.setValue(p + '.name', file['filename']) hdf.setValue(p + '.descr', wiki_to_oneliner(file['description'], hdf, self,self.get_... | 26,188 |
def display(self): self.env.log.debug("Displaying file: %s mime-type: %s" % (self.filename, self.mime_type)) # We don't have to guess if the charset is specified in the # svn:mime-type property ctpos = self.mime_type.find('charset=') if ctpos >= 0: charset = self.mime_type[ctpos + 8:] self.env.log.debug("Charset %s se... | def display(self): self.env.log.debug("Displaying file: %s mime-type: %s" % (self.filename, self.mime_type)) # We don't have to guess if the charset is specified in the # svn:mime-type property ctpos = self.mime_type.find('charset=') if ctpos >= 0: charset = self.mime_type[ctpos + 8:] self.env.log.debug("Charset %s se... | 26,189 |
def _history(self, path, start, end, limit=None): scoped_path = posixpath.join(self.scope[1:], path) return _get_history(scoped_path, self.authz, self.fs_ptr, self.pool, start, end, limit) | def _history(self, path, start, end, limit=None, pool=None): scoped_path = posixpath.join(self.scope[1:], path) return _get_history(scoped_path, self.authz, self.fs_ptr, self.pool, start, end, limit) | 26,190 |
def _history(self, path, start, end, limit=None): scoped_path = posixpath.join(self.scope[1:], path) return _get_history(scoped_path, self.authz, self.fs_ptr, self.pool, start, end, limit) | def _history(self, path, start, end, limit=None): scoped_path = posixpath.join(self.scope[1:], path) return _get_history(scoped_path, self.authz, self.fs_ptr, self.pool, start, end, limit) | 26,191 |
def previous_rev(self, rev, path=''): rev = self.normalize_rev(rev) if rev > 1: # don't use oldest here, as it's too expensive try: for _, prev in self._history(path, 0, rev-1, limit=1): return prev except (SystemError, # "null arg to internal routine" in 1.2.x core.SubversionException): # in 1.3.x pass return None | def previous_rev(self, rev, path=''): rev = self.normalize_rev(rev) if rev > 1: # don't use oldest here, as it's too expensive try: for _, prev in self._history(path, 0, rev-1, limit=1): return prev except (SystemError, # "null arg to internal routine" in 1.2.x core.SubversionException): # in 1.3.x pass return None | 26,192 |
def get_path_history(self, path, rev=None, limit=None): path = self.normalize_path(path) rev = self.normalize_rev(rev) expect_deletion = False subpool = Pool(self.pool) while rev: subpool.clear() if self.has_node(path, rev, subpool): if expect_deletion: # it was missing, now it's there again: # rev+1 must be a delete ... | def get_path_history(self, path, rev=None, limit=None): path = self.normalize_path(path) rev = self.normalize_rev(rev) expect_deletion = False subpool = Pool(self.pool) while rev: subpool.clear() if self.has_node(path, rev, subpool): if expect_deletion: # it was missing, now it's there again: # rev+1 must be a delete ... | 26,193 |
def get_path_history(self, path, rev=None, limit=None): path = self.normalize_path(path) rev = self.normalize_rev(rev) expect_deletion = False subpool = Pool(self.pool) while rev: subpool.clear() if self.has_node(path, rev, subpool): if expect_deletion: # it was missing, now it's there again: # rev+1 must be a delete ... | def get_path_history(self, path, rev=None, limit=None): path = self.normalize_path(path) rev = self.normalize_rev(rev) expect_deletion = False subpool = Pool(self.pool) while rev: subpool.clear() if self.has_node(path, rev, subpool): if expect_deletion: # it was missing, now it's there again: # rev+1 must be a delete ... | 26,194 |
def get_path_history(self, path, rev=None, limit=None): path = self.normalize_path(path) rev = self.normalize_rev(rev) expect_deletion = False subpool = Pool(self.pool) while rev: subpool.clear() if self.has_node(path, rev, subpool): if expect_deletion: # it was missing, now it's there again: # rev+1 must be a delete ... | def get_path_history(self, path, rev=None, limit=None): path = self.normalize_path(path) rev = self.normalize_rev(rev) expect_deletion = False subpool = Pool(self.pool) while rev: subpool.clear() if self.has_node(path, rev, subpool): if expect_deletion: # it was missing, now it's there again: # rev+1 must be a delete ... | 26,195 |
def process_request(self, req): req.perm.assert_permission('TICKET_CREATE') | def process_request(self, req): req.perm.assert_permission('TICKET_CREATE') | 26,196 |
def print_diff (self, old_path, new_path, pool): options = ['-u'] options.append('-L') options.append("%s\t(revision %d)" % (old_path, self.rev-1)) options.append('-L') options.append("%s\t(revision %d)" % (new_path, self.rev)) | defpobj.close() if sys.platform[:3] != "win" and sys.platform != "os2emx": os.waitpid(-1, 0) print_diffpobj.close() if sys.platform[:3] != "win" and sys.platform != "os2emx": os.waitpid(-1, 0) (self,pobj.close() if sys.platform[:3] != "win" and sys.platform != "os2emx": os.waitpid(-1, 0) old_path,pobj.close() if sys.pl... | 26,197 |
def format_props(self): tkt = self.ticket fields = [f for f in tkt.fields if f['type'] != 'textarea' and f['name'] not in ('summary', 'cc')] t = self.modtime or tkt.time_changed width = [0, 0, 0, 0] for i, f in enum([f['name'] for f in fields]): if not tkt.values.has_key(f): continue fval = tkt[f] if fval.find('\n') > ... | def format_props(self): tkt = self.ticket fields = [f for f in tkt.fields if f['type'] != 'textarea' and f['name'] not in ('summary', 'cc')] t = self.modtime or tkt.time_changed width = [0, 0, 0, 0] for i, f in enum([f['name'] for f in fields]): if not tkt.values.has_key(f): continue fval = tkt[f] if fval.find('\n') > ... | 26,198 |
def format_props(self): tkt = self.ticket fields = [f for f in tkt.fields if f['type'] != 'textarea' and f['name'] not in ('summary', 'cc')] t = self.modtime or tkt.time_changed width = [0, 0, 0, 0] for i, f in enum([f['name'] for f in fields]): if not tkt.values.has_key(f): continue fval = tkt[f] if fval.find('\n') > ... | def format_props(self): tkt = self.ticket fields = [f for f in tkt.fields if f['type'] != 'textarea' and f['name'] not in ('summary', 'cc')] t = self.modtime or tkt.time_changed width = [0, 0, 0, 0] for i, f in enum([f['name'] for f in fields]): if not tkt.values.has_key(f): continue fval = tkt[f] if fval.find('\n') > ... | 26,199 |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.