source_code
stringlengths 45
7.24k
| function_name
stringlengths 1
111
| messages
list |
|---|---|---|
def x(s, tokenize=True, encoding='utf-8', **kwargs):
tags = []
for sentence in parse(s, tokenize, True, False, False, False, encoding,
**kwargs).split():
for token in sentence:
tags.append((token[0], token[1]))
return tags
|
tag
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(s, tokenize=True, encoding='utf-8', **kwargs):\n tags = []\n for sentence in parse(s, tokenize, True, False, False, False, encoding,\n **kwargs).split():\n for token in sentence:\n tags.append((token[0], token[1]))\n return tags\n",
"role": "user"
},
{
"content": "tag",
"role": "assistant"
}
] |
def x(s, top=10, **kwargs):
return parser.find_keywords(s, top=top, frequency=parser.frequency)
|
keywords
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(s, top=10, **kwargs):\n return parser.find_keywords(s, top=top, frequency=parser.frequency)\n",
"role": "user"
},
{
"content": "keywords",
"role": "assistant"
}
] |
def x(text, tokenizer=None):
_tokenizer = tokenizer if tokenizer is not None else NLTKPunktTokenizer()
return SentenceTokenizer(tokenizer=_tokenizer).itokenize(text)
|
sent_tokenize
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(text, tokenizer=None):\n _tokenizer = tokenizer if tokenizer is not None else NLTKPunktTokenizer()\n return SentenceTokenizer(tokenizer=_tokenizer).itokenize(text)\n",
"role": "user"
},
{
"content": "sent_tokenize",
"role": "assistant"
}
] |
def x(text, tokenizer=None, include_punc=True, *args, **kwargs):
_tokenizer = tokenizer if tokenizer is not None else NLTKPunktTokenizer()
words = chain.from_iterable(WordTokenizer(tokenizer=_tokenizer).
itokenize(sentence, include_punc, *args, **kwargs) for sentence in
sent_tokenize(text, tokenizer=_tokenizer))
return words
|
word_tokenize
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(text, tokenizer=None, include_punc=True, *args, **kwargs):\n _tokenizer = tokenizer if tokenizer is not None else NLTKPunktTokenizer()\n words = chain.from_iterable(WordTokenizer(tokenizer=_tokenizer).\n itokenize(sentence, include_punc, *args, **kwargs) for sentence in\n sent_tokenize(text, tokenizer=_tokenizer))\n return words\n",
"role": "user"
},
{
"content": "word_tokenize",
"role": "assistant"
}
] |
def x(self, text, include_punc=True, nested=False):
self.tokens = [w for w in (self.word_tokenize(s, include_punc) for s in
self.sent_tokenize(text))]
if nested:
return self.tokens
else:
return list(chain.from_iterable(self.tokens))
|
tokenize
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, text, include_punc=True, nested=False):\n self.tokens = [w for w in (self.word_tokenize(s, include_punc) for s in\n self.sent_tokenize(text))]\n if nested:\n return self.tokens\n else:\n return list(chain.from_iterable(self.tokens))\n",
"role": "user"
},
{
"content": "tokenize",
"role": "assistant"
}
] |
def x(self, text, **kwargs):
sentences = self.sent_tok.tokenize(text, realign_boundaries=kwargs.get(
'realign_boundaries', True))
return sentences
|
sent_tokenize
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, text, **kwargs):\n sentences = self.sent_tok.tokenize(text, realign_boundaries=kwargs.get(\n 'realign_boundaries', True))\n return sentences\n",
"role": "user"
},
{
"content": "sent_tokenize",
"role": "assistant"
}
] |
def x(self, text, include_punc=True):
if text.strip() == '':
return []
_tokens = self.word_tok.tokenize(text)
if len(_tokens) == 1:
if _tokens[0] in PUNCTUATION:
if include_punc:
return _tokens
else:
return []
if include_punc:
return _tokens
else:
words = [(word if word.startswith("'") else strip_punc(word, all=
False)) for word in _tokens if strip_punc(word, all=False)]
return list(words)
|
word_tokenize
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, text, include_punc=True):\n if text.strip() == '':\n return []\n _tokens = self.word_tok.tokenize(text)\n if len(_tokens) == 1:\n if _tokens[0] in PUNCTUATION:\n if include_punc:\n return _tokens\n else:\n return []\n if include_punc:\n return _tokens\n else:\n words = [(word if word.startswith(\"'\") else strip_punc(word, all=\n False)) for word in _tokens if strip_punc(word, all=False)]\n return list(words)\n",
"role": "user"
},
{
"content": "word_tokenize",
"role": "assistant"
}
] |
def x(self, text, **kwargs):
sentences = find_sentences(text, punctuation=kwargs.get('punctuation',
PUNCTUATION), abbreviations=kwargs.get('abbreviations',
ABBREVIATIONS_DE), replace=kwargs.get('replace', replacements),
linebreak='\\n{2,}')
return sentences
|
sent_tokenize
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, text, **kwargs):\n sentences = find_sentences(text, punctuation=kwargs.get('punctuation',\n PUNCTUATION), abbreviations=kwargs.get('abbreviations',\n ABBREVIATIONS_DE), replace=kwargs.get('replace', replacements),\n linebreak='\\\\n{2,}')\n return sentences\n",
"role": "user"
},
{
"content": "sent_tokenize",
"role": "assistant"
}
] |
def x(self, text, include_punc=True, **kwargs):
return self.tokenizer.word_tokenize(text, include_punc, **kwargs)
|
tokenize
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, text, include_punc=True, **kwargs):\n return self.tokenizer.word_tokenize(text, include_punc, **kwargs)\n",
"role": "user"
},
{
"content": "tokenize",
"role": "assistant"
}
] |
def x(self, text):
if text.strip() == '':
return ''
elif text.strip() in PUNCTUATION:
_sym = text.strip()
if _sym in tuple('.?!'):
_tag = '.'
else:
_tag = _sym
if self.lemmata:
return '{0}/{1}/O/O/{0}'.format(_sym, _tag)
else:
return '{0}/{1}/O/O'.format(_sym, _tag)
if self.tokenize:
_tokenized = ' '.join(self.tokenizer.tokenize(text))
else:
_tokenized = text
_parsed = pattern_parse(_tokenized, tokenize=False, tags=self.tags,
chunks=self.chunks, relations=self.relations, lemmata=self.lemmata,
encoding=self.encoding, tagset=self.tagset)
if self.pprint:
_parsed = pattern_pprint(_parsed)
return _parsed
|
parse
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, text):\n if text.strip() == '':\n return ''\n elif text.strip() in PUNCTUATION:\n _sym = text.strip()\n if _sym in tuple('.?!'):\n _tag = '.'\n else:\n _tag = _sym\n if self.lemmata:\n return '{0}/{1}/O/O/{0}'.format(_sym, _tag)\n else:\n return '{0}/{1}/O/O'.format(_sym, _tag)\n if self.tokenize:\n _tokenized = ' '.join(self.tokenizer.tokenize(text))\n else:\n _tokenized = text\n _parsed = pattern_parse(_tokenized, tokenize=False, tags=self.tags,\n chunks=self.chunks, relations=self.relations, lemmata=self.lemmata,\n encoding=self.encoding, tagset=self.tagset)\n if self.pprint:\n _parsed = pattern_pprint(_parsed)\n return _parsed\n",
"role": "user"
},
{
"content": "parse",
"role": "assistant"
}
] |
def x(self, text):
_extracted = []
if text.strip() == '':
return _extracted
parsed_sentences = self._parse_text(text)
for s in parsed_sentences:
tokens = s.split()
new_np = []
for t in tokens:
w, tag, phrase, role = t.split('/')
if 'NP' in phrase and not self._is_verb(w, tag):
if len(new_np) > 0 and w.lower() in START_NEW_NP:
_extracted.append(' '.join(new_np))
new_np = [w]
else:
new_np.append(w.lower() if tokens[0].startswith(w) and
not tag.startswith('N') else w)
else:
if len(new_np) > 0:
_extracted.append(' '.join(new_np))
new_np = []
return self._filter_extracted(_extracted)
|
extract
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, text):\n _extracted = []\n if text.strip() == '':\n return _extracted\n parsed_sentences = self._parse_text(text)\n for s in parsed_sentences:\n tokens = s.split()\n new_np = []\n for t in tokens:\n w, tag, phrase, role = t.split('/')\n if 'NP' in phrase and not self._is_verb(w, tag):\n if len(new_np) > 0 and w.lower() in START_NEW_NP:\n _extracted.append(' '.join(new_np))\n new_np = [w]\n else:\n new_np.append(w.lower() if tokens[0].startswith(w) and \n not tag.startswith('N') else w)\n else:\n if len(new_np) > 0:\n _extracted.append(' '.join(new_np))\n new_np = []\n return self._filter_extracted(_extracted)\n",
"role": "user"
},
{
"content": "extract",
"role": "assistant"
}
] |
def x(self, extracted_list):
_filtered = []
for np in extracted_list:
_np = np.split()
if _np[0] in INSIGNIFICANT:
_np.pop(0)
try:
if _np[-1] in INSIGNIFICANT:
_np.pop(-1)
if _np[0] in INSIGNIFICANT:
_np.pop(0)
except IndexError:
_np = []
if len(_np) > 0:
_filtered.append(' '.join(_np))
return _filtered
|
filter_extracted
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, extracted_list):\n _filtered = []\n for np in extracted_list:\n _np = np.split()\n if _np[0] in INSIGNIFICANT:\n _np.pop(0)\n try:\n if _np[-1] in INSIGNIFICANT:\n _np.pop(-1)\n if _np[0] in INSIGNIFICANT:\n _np.pop(0)\n except IndexError:\n _np = []\n if len(_np) > 0:\n _filtered.append(' '.join(_np))\n return _filtered\n",
"role": "user"
},
{
"content": "filter_extracted",
"role": "assistant"
}
] |
def x(self, text):
if isinstance(self.tokenizer, PatternTokenizer):
parsed_text = pattern_parse(text, tokenize=True, lemmata=False)
else:
_tokenized = []
_sentences = sent_tokenize(text, tokenizer=self.tokenizer)
for s in _sentences:
_tokenized.append(' '.join(self.tokenizer.tokenize(s)))
parsed_text = pattern_parse(_tokenized, tokenize=False, lemmata=False)
return parsed_text.split('\n')
|
parse_text
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, text):\n if isinstance(self.tokenizer, PatternTokenizer):\n parsed_text = pattern_parse(text, tokenize=True, lemmata=False)\n else:\n _tokenized = []\n _sentences = sent_tokenize(text, tokenizer=self.tokenizer)\n for s in _sentences:\n _tokenized.append(' '.join(self.tokenizer.tokenize(s)))\n parsed_text = pattern_parse(_tokenized, tokenize=False, lemmata=False)\n return parsed_text.split('\\n')\n",
"role": "user"
},
{
"content": "parse_text",
"role": "assistant"
}
] |
def x(self, sentence, tokenize=True):
if sentence.strip() == '':
return []
elif sentence.strip() in PUNCTUATION:
if self.include_punc:
_sym = sentence.strip()
if _sym in tuple('.?!'):
_tag = '.'
else:
_tag = _sym
return [(_sym, _tag)]
else:
return []
if tokenize:
_tokenized = ' '.join(self.tokenizer.tokenize(sentence))
sentence = _tokenized
_tagged = pattern_tag(sentence, tokenize=False, encoding=self.encoding,
tagset=self.tagset)
if self.include_punc:
return _tagged
else:
_tagged = [(word, t) for word, t in _tagged if not
PUNCTUATION_REGEX.match(unicode(t))]
return _tagged
|
tag
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, sentence, tokenize=True):\n if sentence.strip() == '':\n return []\n elif sentence.strip() in PUNCTUATION:\n if self.include_punc:\n _sym = sentence.strip()\n if _sym in tuple('.?!'):\n _tag = '.'\n else:\n _tag = _sym\n return [(_sym, _tag)]\n else:\n return []\n if tokenize:\n _tokenized = ' '.join(self.tokenizer.tokenize(sentence))\n sentence = _tokenized\n _tagged = pattern_tag(sentence, tokenize=False, encoding=self.encoding,\n tagset=self.tagset)\n if self.include_punc:\n return _tagged\n else:\n _tagged = [(word, t) for word, t in _tagged if not\n PUNCTUATION_REGEX.match(unicode(t))]\n return _tagged\n",
"role": "user"
},
{
"content": "tag",
"role": "assistant"
}
] |
def x(v, encoding='utf-8'):
if isinstance(encoding, basestring):
encoding = ((encoding,),) + (('windows-1252',), ('utf-8', 'ignore'))
if isinstance(v, binary_type):
for e in encoding:
try:
return v.decode(*e)
except:
pass
return v
return unicode(v)
|
decode_string
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(v, encoding='utf-8'):\n if isinstance(encoding, basestring):\n encoding = ((encoding,),) + (('windows-1252',), ('utf-8', 'ignore'))\n if isinstance(v, binary_type):\n for e in encoding:\n try:\n return v.decode(*e)\n except:\n pass\n return v\n return unicode(v)\n",
"role": "user"
},
{
"content": "decode_string",
"role": "assistant"
}
] |
def x(v, encoding='utf-8'):
if isinstance(encoding, basestring):
encoding = ((encoding,),) + (('windows-1252',), ('utf-8', 'ignore'))
if isinstance(v, unicode):
for e in encoding:
try:
return v.encode(*e)
except:
pass
return v
return str(v)
|
encode_string
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(v, encoding='utf-8'):\n if isinstance(encoding, basestring):\n encoding = ((encoding,),) + (('windows-1252',), ('utf-8', 'ignore'))\n if isinstance(v, unicode):\n for e in encoding:\n try:\n return v.encode(*e)\n except:\n pass\n return v\n return str(v)\n",
"role": "user"
},
{
"content": "encode_string",
"role": "assistant"
}
] |
def x(cmd, mode=os.F_OK | os.X_OK, path=None):
def x(fn, mode):
return os.path.exists(fn) and os.access(fn, mode
) and not os.path.isdir(fn)
if os.path.dirname(cmd):
if _access_check(cmd, mode):
return cmd
return None
if path is None:
path = os.environ.get('PATH', os.defpath)
if not path:
return None
path = path.split(os.pathsep)
if sys.platform == 'win32':
if not os.curdir in path:
path.insert(0, os.curdir)
pathext = os.environ.get('PATHEXT', '').split(os.pathsep)
if any([cmd.lower().endswith(ext.lower()) for ext in pathext]):
files = [cmd]
else:
files = [(cmd + ext) for ext in pathext]
else:
files = [cmd]
seen = set()
for dir in path:
normdir = os.path.normcase(dir)
if normdir not in seen:
seen.add(normdir)
for thefile in files:
name = os.path.join(dir, thefile)
if _access_check(name, mode):
return name
return None
|
shutil_which
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(cmd, mode=os.F_OK | os.X_OK, path=None):\n\n def x(fn, mode):\n return os.path.exists(fn) and os.access(fn, mode\n ) and not os.path.isdir(fn)\n if os.path.dirname(cmd):\n if _access_check(cmd, mode):\n return cmd\n return None\n if path is None:\n path = os.environ.get('PATH', os.defpath)\n if not path:\n return None\n path = path.split(os.pathsep)\n if sys.platform == 'win32':\n if not os.curdir in path:\n path.insert(0, os.curdir)\n pathext = os.environ.get('PATHEXT', '').split(os.pathsep)\n if any([cmd.lower().endswith(ext.lower()) for ext in pathext]):\n files = [cmd]\n else:\n files = [(cmd + ext) for ext in pathext]\n else:\n files = [cmd]\n seen = set()\n for dir in path:\n normdir = os.path.normcase(dir)\n if normdir not in seen:\n seen.add(normdir)\n for thefile in files:\n name = os.path.join(dir, thefile)\n if _access_check(name, mode):\n return name\n return None\n",
"role": "user"
},
{
"content": "shutil_which",
"role": "assistant"
}
] |
def x(self, from_lang=None, to='de'):
if from_lang is None:
from_lang = self.translator.detect(self.string)
return self.translator.translate(self.string, from_lang=from_lang,
to_lang=to)
|
translate
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, from_lang=None, to='de'):\n if from_lang is None:\n from_lang = self.translator.detect(self.string)\n return self.translator.translate(self.string, from_lang=from_lang,\n to_lang=to)\n",
"role": "user"
},
{
"content": "translate",
"role": "assistant"
}
] |
def x(self):
_lemmatizer = PatternParserLemmatizer(tokenizer=NLTKPunktTokenizer())
_raw = ' '.join(self) + '.'
_lemmas = _lemmatizer.lemmatize(_raw)
return self.__class__([Word(l, t) for l, t in _lemmas])
|
lemmatize
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n _lemmatizer = PatternParserLemmatizer(tokenizer=NLTKPunktTokenizer())\n _raw = ' '.join(self) + '.'\n _lemmas = _lemmatizer.lemmatize(_raw)\n return self.__class__([Word(l, t) for l, t in _lemmas])\n",
"role": "user"
},
{
"content": "lemmatize",
"role": "assistant"
}
] |
def x(self, tokenizer=None):
t = tokenizer if tokenizer is not None else self.tokenizer
return WordList(t.tokenize(self.raw))
|
tokenize
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, tokenizer=None):\n t = tokenizer if tokenizer is not None else self.tokenizer\n return WordList(t.tokenize(self.raw))\n",
"role": "user"
},
{
"content": "tokenize",
"role": "assistant"
}
] |
def x(self):
return WordList([phrase.strip() for phrase in self.np_extractor.extract
(self.raw) if len(phrase.split()) > 1])
|
noun_phrases
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n return WordList([phrase.strip() for phrase in self.np_extractor.extract\n (self.raw) if len(phrase.split()) > 1])\n",
"role": "user"
},
{
"content": "noun_phrases",
"role": "assistant"
}
] |
def x(self):
return [(Word(word, pos_tag=t), unicode(t)) for word, t in self.
pos_tagger.tag(self.raw)]
|
pos_tags
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n return [(Word(word, pos_tag=t), unicode(t)) for word, t in self.\n pos_tagger.tag(self.raw)]\n",
"role": "user"
},
{
"content": "pos_tags",
"role": "assistant"
}
] |
def x(self):
counts = defaultdict(int)
stripped_words = [lowerstrip(word) for word in self.words]
for word in stripped_words:
counts[word] += 1
return counts
|
word_counts
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n counts = defaultdict(int)\n stripped_words = [lowerstrip(word) for word in self.words]\n for word in stripped_words:\n counts[word] += 1\n return counts\n",
"role": "user"
},
{
"content": "word_counts",
"role": "assistant"
}
] |
def x(self):
return {'raw': self.raw, 'start_index': self.start_index, 'end_index':
self.end_index, 'stripped': self.stripped, 'noun_phrases': self.
noun_phrases, 'polarity': self.polarity, 'subjectivity': self.
subjectivity}
|
dict
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n return {'raw': self.raw, 'start_index': self.start_index, 'end_index':\n self.end_index, 'stripped': self.stripped, 'noun_phrases': self.\n noun_phrases, 'polarity': self.polarity, 'subjectivity': self.\n subjectivity}\n",
"role": "user"
},
{
"content": "dict",
"role": "assistant"
}
] |
def x(self):
return WordList(word_tokenize(self.raw, self.tokenizer, include_punc=False)
)
|
words
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n return WordList(word_tokenize(self.raw, self.tokenizer, include_punc=False)\n )\n",
"role": "user"
},
{
"content": "words",
"role": "assistant"
}
] |
def x(self):
_RETURN_TYPE = namedtuple('Sentiment', ['polarity', 'subjectivity'])
_polarity = 0
_subjectivity = 0
for s in self.sentences:
_polarity += s.polarity
_subjectivity += s.subjectivity
try:
polarity = _polarity / len(self.sentences)
except ZeroDivisionError:
polarity = 0.0
try:
subjectivity = _subjectivity / len(self.sentences)
except ZeroDivisionError:
subjectivity = 0.0
return _RETURN_TYPE(polarity, subjectivity)
|
sentiment
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n _RETURN_TYPE = namedtuple('Sentiment', ['polarity', 'subjectivity'])\n _polarity = 0\n _subjectivity = 0\n for s in self.sentences:\n _polarity += s.polarity\n _subjectivity += s.subjectivity\n try:\n polarity = _polarity / len(self.sentences)\n except ZeroDivisionError:\n polarity = 0.0\n try:\n subjectivity = _subjectivity / len(self.sentences)\n except ZeroDivisionError:\n subjectivity = 0.0\n return _RETURN_TYPE(polarity, subjectivity)\n",
"role": "user"
},
{
"content": "sentiment",
"role": "assistant"
}
] |
def x(self, *args, **kwargs):
return json.dumps(self.serialized, *args, **kwargs)
|
to_json
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, *args, **kwargs):\n return json.dumps(self.serialized, *args, **kwargs)\n",
"role": "user"
},
{
"content": "to_json",
"role": "assistant"
}
] |
def x(self):
sentence_objects = []
sentences = sent_tokenize(self.raw, tokenizer=self.tokenizer)
char_index = 0
for sent in sentences:
try:
start_index = self.raw.index(sent, char_index)
char_index += len(sent)
end_index = start_index + len(sent)
except ValueError:
start_index = None
end_index = None
s = Sentence(sent, start_index=start_index, end_index=end_index,
tokenizer=self.tokenizer, np_extractor=self.np_extractor,
pos_tagger=self.pos_tagger, analyzer=self.analyzer, parser=self
.parser, classifier=self.classifier)
sentence_objects.append(s)
return sentence_objects
|
create_sentence_objects
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n sentence_objects = []\n sentences = sent_tokenize(self.raw, tokenizer=self.tokenizer)\n char_index = 0\n for sent in sentences:\n try:\n start_index = self.raw.index(sent, char_index)\n char_index += len(sent)\n end_index = start_index + len(sent)\n except ValueError:\n start_index = None\n end_index = None\n s = Sentence(sent, start_index=start_index, end_index=end_index,\n tokenizer=self.tokenizer, np_extractor=self.np_extractor,\n pos_tagger=self.pos_tagger, analyzer=self.analyzer, parser=self\n .parser, classifier=self.classifier)\n sentence_objects.append(s)\n return sentence_objects\n",
"role": "user"
},
{
"content": "create_sentence_objects",
"role": "assistant"
}
] |
def x(string, n=3, punctuation=PUNCTUATION, continuous=False):
def x(s, punctuation=set(punctuation)):
return [w for w in s if (isinstance(w, Word) and w.string or w) not in
punctuation]
if n <= 0:
return []
if isinstance(string, basestring):
s = [strip_punctuation(s.split(' ')) for s in tokenize(string)]
if isinstance(string, Sentence):
s = [strip_punctuation(string)]
if isinstance(string, Text):
s = [strip_punctuation(s) for s in string]
if continuous:
s = [sum(s, [])]
g = []
for s in s:
g.extend([tuple(s[i:i + n]) for i in range(len(s) - n + 1)])
return g
|
ngrams
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(string, n=3, punctuation=PUNCTUATION, continuous=False):\n\n def x(s, punctuation=set(punctuation)):\n return [w for w in s if (isinstance(w, Word) and w.string or w) not in\n punctuation]\n if n <= 0:\n return []\n if isinstance(string, basestring):\n s = [strip_punctuation(s.split(' ')) for s in tokenize(string)]\n if isinstance(string, Sentence):\n s = [strip_punctuation(string)]\n if isinstance(string, Text):\n s = [strip_punctuation(s) for s in string]\n if continuous:\n s = [sum(s, [])]\n g = []\n for s in s:\n g.extend([tuple(s[i:i + n]) for i in range(len(s) - n + 1)])\n return g\n",
"role": "user"
},
{
"content": "ngrams",
"role": "assistant"
}
] |
def x(s, n=3):
if n == 0:
return s[0:0]
return re.sub('((.)\\2{%s,})' % (n - 1), lambda m: m.group(1)[0] * n, s)
|
deflood
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(s, n=3):\n if n == 0:\n return s[0:0]\n return re.sub('((.)\\\\2{%s,})' % (n - 1), lambda m: m.group(1)[0] * n, s)\n",
"role": "user"
},
{
"content": "deflood",
"role": "assistant"
}
] |
def x(string, token=[WORD, POS, CHUNK, PNP], column=4):
if isinstance(string, basestring):
print('\n\n'.join([table(sentence, fill=column) for sentence in
Text(string, token)]))
if isinstance(string, Text):
print('\n\n'.join([table(sentence, fill=column) for sentence in
string]))
if isinstance(string, Sentence):
print(table(string, fill=column))
|
pprint
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(string, token=[WORD, POS, CHUNK, PNP], column=4):\n if isinstance(string, basestring):\n print('\\n\\n'.join([table(sentence, fill=column) for sentence in\n Text(string, token)]))\n if isinstance(string, Text):\n print('\\n\\n'.join([table(sentence, fill=column) for sentence in\n string]))\n if isinstance(string, Sentence):\n print(table(string, fill=column))\n",
"role": "user"
},
{
"content": "pprint",
"role": "assistant"
}
] |
def x(path, encoding='utf-8', comment=';;;'):
if path:
if isinstance(path, basestring) and os.path.exists(path):
if PY2:
f = codecs.open(path, 'r', encoding='utf-8')
else:
f = open(path, 'r', encoding='utf-8')
elif isinstance(path, basestring):
f = path.splitlines()
else:
f = path
for i, line in enumerate(f):
line = line.strip(codecs.BOM_UTF8) if i == 0 and isinstance(line,
binary_type) else line
line = line.strip()
line = decode_utf8(line, encoding)
if not line or comment and line.startswith(comment):
continue
yield line
return
|
read
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(path, encoding='utf-8', comment=';;;'):\n if path:\n if isinstance(path, basestring) and os.path.exists(path):\n if PY2:\n f = codecs.open(path, 'r', encoding='utf-8')\n else:\n f = open(path, 'r', encoding='utf-8')\n elif isinstance(path, basestring):\n f = path.splitlines()\n else:\n f = path\n for i, line in enumerate(f):\n line = line.strip(codecs.BOM_UTF8) if i == 0 and isinstance(line,\n binary_type) else line\n line = line.strip()\n line = decode_utf8(line, encoding)\n if not line or comment and line.startswith(comment):\n continue\n yield line\n return\n",
"role": "user"
},
{
"content": "read",
"role": "assistant"
}
] |
def x(token, tag):
if tag.startswith(('NNP-', 'NNPS-')):
return token, '%s-%s' % (NOUN, tag.split('-')[-1])
if tag in ('NN', 'NNS', 'NNP', 'NNPS', 'NP'):
return token, NOUN
if tag in ('MD', 'VB', 'VBD', 'VBG', 'VBN', 'VBP', 'VBZ'):
return token, VERB
if tag in ('JJ', 'JJR', 'JJS'):
return token, ADJ
if tag in ('RB', 'RBR', 'RBS', 'WRB'):
return token, ADV
if tag in ('PRP', 'PRP$', 'WP', 'WP$'):
return token, PRON
if tag in ('DT', 'PDT', 'WDT', 'EX'):
return token, DET
if tag in ('IN',):
return token, PREP
if tag in ('CD',):
return token, NUM
if tag in ('CC',):
return token, CONJ
if tag in ('UH',):
return token, INTJ
if tag in ('POS', 'RP', 'TO'):
return token, PRT
if tag in ('SYM', 'LS', '.', '!', '?', ',', ':', '(', ')', '"', '#', '$'):
return token, PUNC
return token, X
|
penntreebank2universal
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(token, tag):\n if tag.startswith(('NNP-', 'NNPS-')):\n return token, '%s-%s' % (NOUN, tag.split('-')[-1])\n if tag in ('NN', 'NNS', 'NNP', 'NNPS', 'NP'):\n return token, NOUN\n if tag in ('MD', 'VB', 'VBD', 'VBG', 'VBN', 'VBP', 'VBZ'):\n return token, VERB\n if tag in ('JJ', 'JJR', 'JJS'):\n return token, ADJ\n if tag in ('RB', 'RBR', 'RBS', 'WRB'):\n return token, ADV\n if tag in ('PRP', 'PRP$', 'WP', 'WP$'):\n return token, PRON\n if tag in ('DT', 'PDT', 'WDT', 'EX'):\n return token, DET\n if tag in ('IN',):\n return token, PREP\n if tag in ('CD',):\n return token, NUM\n if tag in ('CC',):\n return token, CONJ\n if tag in ('UH',):\n return token, INTJ\n if tag in ('POS', 'RP', 'TO'):\n return token, PRT\n if tag in ('SYM', 'LS', '.', '!', '?', ',', ':', '(', ')', '\"', '#', '$'):\n return token, PUNC\n return token, X\n",
"role": "user"
},
{
"content": "penntreebank2universal",
"role": "assistant"
}
] |
def x(string, punctuation=PUNCTUATION, abbreviations=ABBREVIATIONS, replace
=replacements, linebreak='\\n{2,}'):
punctuation = tuple(punctuation.replace('.', ''))
for a, b in replace.items():
string = re.sub(a, b, string)
if isinstance(string, unicode):
string = string.replace(u'“', u' “ ')
string = string.replace(u'”', u' ” ')
string = string.replace(u'‘', u' ‘ ')
string = string.replace(u'’', u' ’ ')
string = re.sub('\r\n', '\n', string)
string = re.sub(linebreak, ' %s ' % EOS, string)
string = re.sub('\\s+', ' ', string)
tokens = []
for t in TOKEN.findall(string + ' '):
if len(t) > 0:
tail = []
while t.startswith(punctuation) and not t in replace:
if t.startswith(punctuation):
tokens.append(t[0])
t = t[1:]
while t.endswith(punctuation + ('.',)) and not t in replace:
if t.endswith(punctuation):
tail.append(t[-1])
t = t[:-1]
if t.endswith('...'):
tail.append('...')
t = t[:-3].rstrip('.')
if t.endswith('.'):
if t in abbreviations or RE_ABBR1.match(t
) is not None or RE_ABBR2.match(t
) is not None or RE_ABBR3.match(t) is not None:
break
else:
tail.append(t[-1])
t = t[:-1]
if t != '':
tokens.append(t)
tokens.extend(reversed(tail))
sentences, i, j = [[]], 0, 0
while j < len(tokens):
if tokens[j] in ('...', '.', '!', '?', EOS):
while j < len(tokens) and tokens[j] in ("'", '"', u'”', u'’',
'...', '.', '!', '?', ')', EOS):
if tokens[j] in ("'", '"') and sentences[-1].count(tokens[j]
) % 2 == 0:
break
j += 1
sentences[-1].extend(t for t in tokens[i:j] if t != EOS)
sentences.append([])
i = j
j += 1
sentences[-1].extend(tokens[i:j])
sentences = (' '.join(s) for s in sentences if len(s) > 0)
sentences = (RE_SARCASM.sub('(!)', s) for s in sentences)
sentences = [RE_EMOTICONS.sub(lambda m: m.group(1).replace(' ', '') + m
.group(2), s) for s in sentences]
return sentences
|
find_tokens
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(string, punctuation=PUNCTUATION, abbreviations=ABBREVIATIONS, replace\n =replacements, linebreak='\\\\n{2,}'):\n punctuation = tuple(punctuation.replace('.', ''))\n for a, b in replace.items():\n string = re.sub(a, b, string)\n if isinstance(string, unicode):\n string = string.replace(u'“', u' “ ')\n string = string.replace(u'”', u' ” ')\n string = string.replace(u'‘', u' ‘ ')\n string = string.replace(u'’', u' ’ ')\n string = re.sub('\\r\\n', '\\n', string)\n string = re.sub(linebreak, ' %s ' % EOS, string)\n string = re.sub('\\\\s+', ' ', string)\n tokens = []\n for t in TOKEN.findall(string + ' '):\n if len(t) > 0:\n tail = []\n while t.startswith(punctuation) and not t in replace:\n if t.startswith(punctuation):\n tokens.append(t[0])\n t = t[1:]\n while t.endswith(punctuation + ('.',)) and not t in replace:\n if t.endswith(punctuation):\n tail.append(t[-1])\n t = t[:-1]\n if t.endswith('...'):\n tail.append('...')\n t = t[:-3].rstrip('.')\n if t.endswith('.'):\n if t in abbreviations or RE_ABBR1.match(t\n ) is not None or RE_ABBR2.match(t\n ) is not None or RE_ABBR3.match(t) is not None:\n break\n else:\n tail.append(t[-1])\n t = t[:-1]\n if t != '':\n tokens.append(t)\n tokens.extend(reversed(tail))\n sentences, i, j = [[]], 0, 0\n while j < len(tokens):\n if tokens[j] in ('...', '.', '!', '?', EOS):\n while j < len(tokens) and tokens[j] in (\"'\", '\"', u'”', u'’',\n '...', '.', '!', '?', ')', EOS):\n if tokens[j] in (\"'\", '\"') and sentences[-1].count(tokens[j]\n ) % 2 == 0:\n break\n j += 1\n sentences[-1].extend(t for t in tokens[i:j] if t != EOS)\n sentences.append([])\n i = j\n j += 1\n sentences[-1].extend(tokens[i:j])\n sentences = (' '.join(s) for s in sentences if len(s) > 0)\n sentences = (RE_SARCASM.sub('(!)', s) for s in sentences)\n sentences = [RE_EMOTICONS.sub(lambda m: m.group(1).replace(' ', '') + m\n .group(2), s) for s in sentences]\n return sentences\n",
"role": "user"
},
{
"content": "find_tokens",
"role": "assistant"
}
] |
def x(token, tag='NN'):
if isinstance(token, (list, tuple)):
token, tag = token
if token.endswith('ing'):
tag = 'VBG'
if token.endswith('ly'):
tag = 'RB'
if token.endswith('s') and not token.endswith(('is', 'ous', 'ss')):
tag = 'NNS'
if token.endswith(('able', 'al', 'ful', 'ible', 'ient', 'ish', 'ive',
'less', 'tic', 'ous')) or '-' in token:
tag = 'JJ'
if token.endswith('ed'):
tag = 'VBN'
if token.endswith(('ate', 'ify', 'ise', 'ize')):
tag = 'VBP'
return [token, tag]
|
suffix_rules
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(token, tag='NN'):\n if isinstance(token, (list, tuple)):\n token, tag = token\n if token.endswith('ing'):\n tag = 'VBG'\n if token.endswith('ly'):\n tag = 'RB'\n if token.endswith('s') and not token.endswith(('is', 'ous', 'ss')):\n tag = 'NNS'\n if token.endswith(('able', 'al', 'ful', 'ible', 'ient', 'ish', 'ive',\n 'less', 'tic', 'ous')) or '-' in token:\n tag = 'JJ'\n if token.endswith('ed'):\n tag = 'VBN'\n if token.endswith(('ate', 'ify', 'ise', 'ize')):\n tag = 'VBP'\n return [token, tag]\n",
"role": "user"
},
{
"content": "suffix_rules",
"role": "assistant"
}
] |
def x(tokens, lexicon={}, model=None, morphology=None, context=None,
entities=None, default=('NN', 'NNP', 'CD'), language='en', map=None, **
kwargs):
tagged = []
for i, token in enumerate(tokens):
tagged.append([token, lexicon.get(token, i == 0 and lexicon.get(
token.lower()) or None)])
for i, (token, tag) in enumerate(tagged):
prev, next = (None, None), (None, None)
if i > 0:
prev = tagged[i - 1]
if i < len(tagged) - 1:
next = tagged[i + 1]
if tag is None or token in (model is not None and model.unknown or ()):
if model is not None:
tagged[i] = model.apply([token, None], prev, next)
elif token.istitle() and language != 'de':
tagged[i] = [token, default[1]]
elif CD.match(token) is not None:
tagged[i] = [token, default[2]]
elif morphology is not None:
tagged[i] = morphology.apply([token, default[0]], prev, next)
elif language == 'en':
tagged[i] = _suffix_rules([token, default[0]])
else:
tagged[i] = [token, default[0]]
if context is not None and model is None:
tagged = context.apply(tagged)
if entities is not None:
tagged = entities.apply(tagged)
if map is not None:
tagged = [(list(map(token, tag)) or [token, default[0]]) for token,
tag in tagged]
return tagged
|
find_tags
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(tokens, lexicon={}, model=None, morphology=None, context=None,\n entities=None, default=('NN', 'NNP', 'CD'), language='en', map=None, **\n kwargs):\n tagged = []\n for i, token in enumerate(tokens):\n tagged.append([token, lexicon.get(token, i == 0 and lexicon.get(\n token.lower()) or None)])\n for i, (token, tag) in enumerate(tagged):\n prev, next = (None, None), (None, None)\n if i > 0:\n prev = tagged[i - 1]\n if i < len(tagged) - 1:\n next = tagged[i + 1]\n if tag is None or token in (model is not None and model.unknown or ()):\n if model is not None:\n tagged[i] = model.apply([token, None], prev, next)\n elif token.istitle() and language != 'de':\n tagged[i] = [token, default[1]]\n elif CD.match(token) is not None:\n tagged[i] = [token, default[2]]\n elif morphology is not None:\n tagged[i] = morphology.apply([token, default[0]], prev, next)\n elif language == 'en':\n tagged[i] = _suffix_rules([token, default[0]])\n else:\n tagged[i] = [token, default[0]]\n if context is not None and model is None:\n tagged = context.apply(tagged)\n if entities is not None:\n tagged = entities.apply(tagged)\n if map is not None:\n tagged = [(list(map(token, tag)) or [token, default[0]]) for token,\n tag in tagged]\n return tagged\n",
"role": "user"
},
{
"content": "find_tags",
"role": "assistant"
}
] |
def x(tagged, language='en'):
chunked = [x for x in tagged]
tags = ''.join('%s%s' % (tag, SEPARATOR) for token, tag in tagged)
for tag, rule in CHUNKS[int(language in ('ca', 'es', 'pt', 'fr', 'it',
'pt', 'ro'))]:
for m in rule.finditer(tags):
i = m.start()
j = tags[:i].count(SEPARATOR)
n = m.group(0).count(SEPARATOR)
for k in range(j, j + n):
if len(chunked[k]) == 3:
continue
if len(chunked[k]) < 3:
if k == j and chunked[k][1] in ('CC', 'CJ', ','):
j += 1
elif k == j:
chunked[k].append('B-' + tag)
else:
chunked[k].append('I-' + tag)
for chink in filter(lambda x: len(x) < 3, chunked):
chink.append('O')
for i, (word, tag, chunk) in enumerate(chunked):
if tag.startswith('RB') and chunk == 'B-NP':
if i < len(chunked) - 1 and not chunked[i + 1][1].startswith('JJ'):
chunked[i + 0][2] = 'B-ADVP'
chunked[i + 1][2] = 'B-NP'
if i < len(chunked) - 1 and chunked[i + 1][1] in ('CC', 'CJ', ','):
chunked[i + 1][2] = 'O'
if i < len(chunked) - 2 and chunked[i + 1][2] == 'O':
chunked[i + 2][2] = 'B-NP'
return chunked
|
find_chunks
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(tagged, language='en'):\n chunked = [x for x in tagged]\n tags = ''.join('%s%s' % (tag, SEPARATOR) for token, tag in tagged)\n for tag, rule in CHUNKS[int(language in ('ca', 'es', 'pt', 'fr', 'it',\n 'pt', 'ro'))]:\n for m in rule.finditer(tags):\n i = m.start()\n j = tags[:i].count(SEPARATOR)\n n = m.group(0).count(SEPARATOR)\n for k in range(j, j + n):\n if len(chunked[k]) == 3:\n continue\n if len(chunked[k]) < 3:\n if k == j and chunked[k][1] in ('CC', 'CJ', ','):\n j += 1\n elif k == j:\n chunked[k].append('B-' + tag)\n else:\n chunked[k].append('I-' + tag)\n for chink in filter(lambda x: len(x) < 3, chunked):\n chink.append('O')\n for i, (word, tag, chunk) in enumerate(chunked):\n if tag.startswith('RB') and chunk == 'B-NP':\n if i < len(chunked) - 1 and not chunked[i + 1][1].startswith('JJ'):\n chunked[i + 0][2] = 'B-ADVP'\n chunked[i + 1][2] = 'B-NP'\n if i < len(chunked) - 1 and chunked[i + 1][1] in ('CC', 'CJ', ','):\n chunked[i + 1][2] = 'O'\n if i < len(chunked) - 2 and chunked[i + 1][2] == 'O':\n chunked[i + 2][2] = 'B-NP'\n return chunked\n",
"role": "user"
},
{
"content": "find_chunks",
"role": "assistant"
}
] |
def x(chunked):
for ch in chunked:
ch.append('O')
for i, chunk in enumerate(chunked):
if chunk[2].endswith('PP') and chunk[-1] == 'O':
if i < len(chunked) - 1 and (chunked[i + 1][2].endswith(('NP',
'PP')) or chunked[i + 1][1] in ('VBG', 'VBN')):
chunk[-1] = 'B-PNP'
pp = True
for ch in chunked[i + 1:]:
if not (ch[2].endswith(('NP', 'PP')) or ch[1] in ('VBG',
'VBN')):
break
if ch[2].endswith('PP') and pp:
ch[-1] = 'I-PNP'
if not ch[2].endswith('PP'):
ch[-1] = 'I-PNP'
pp = False
return chunked
|
find_prepositions
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(chunked):\n for ch in chunked:\n ch.append('O')\n for i, chunk in enumerate(chunked):\n if chunk[2].endswith('PP') and chunk[-1] == 'O':\n if i < len(chunked) - 1 and (chunked[i + 1][2].endswith(('NP',\n 'PP')) or chunked[i + 1][1] in ('VBG', 'VBN')):\n chunk[-1] = 'B-PNP'\n pp = True\n for ch in chunked[i + 1:]:\n if not (ch[2].endswith(('NP', 'PP')) or ch[1] in ('VBG',\n 'VBN')):\n break\n if ch[2].endswith('PP') and pp:\n ch[-1] = 'I-PNP'\n if not ch[2].endswith('PP'):\n ch[-1] = 'I-PNP'\n pp = False\n return chunked\n",
"role": "user"
},
{
"content": "find_prepositions",
"role": "assistant"
}
] |
def x(chunked):
tag = lambda token: token[2].split('-')[-1]
chunks = []
for token in chunked:
if len(chunks) == 0 or token[2].startswith('B-') or tag(token) != tag(
chunks[-1][-1]):
chunks.append([])
chunks[-1].append(token + ['O'])
id = 0
for i, chunk in enumerate(chunks):
if tag(chunk[-1]) == 'VP' and i > 0 and tag(chunks[i - 1][-1]) == 'NP':
if chunk[-1][-1] == 'O':
id += 1
for token in chunk:
token[-1] = 'VP-' + str(id)
for token in chunks[i - 1]:
token[-1] += '*NP-SBJ-' + str(id)
token[-1] = token[-1].lstrip('O-*')
if tag(chunk[-1]) == 'VP' and i < len(chunks) - 1 and tag(chunks[i +
1][-1]) == 'NP':
if chunk[-1][-1] == 'O':
id += 1
for token in chunk:
token[-1] = 'VP-' + str(id)
for token in chunks[i + 1]:
token[-1] = '*NP-OBJ-' + str(id)
token[-1] = token[-1].lstrip('O-*')
for i, chunk in enumerate(chunks):
if 0 < i < len(chunks) - 1 and len(chunk) == 1 and chunk[-1][-1
] == 'O':
t0, t1, t2 = chunks[i - 1][-1], chunks[i][0], chunks[i + 1][0]
if tag(t1) == 'PP' and t2[1] in ('DT', 'PR', 'PRP$'):
if t0[0] in BE and t1[0] in ('in', 'at'):
t1[-1] = 'PP-LOC'
if t0[0] in GO and t1[0] in ('to', 'towards'):
t1[-1] = 'PP-DIR'
related = []
[related.extend(chunk) for chunk in chunks]
return related
|
find_relations
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(chunked):\n tag = lambda token: token[2].split('-')[-1]\n chunks = []\n for token in chunked:\n if len(chunks) == 0 or token[2].startswith('B-') or tag(token) != tag(\n chunks[-1][-1]):\n chunks.append([])\n chunks[-1].append(token + ['O'])\n id = 0\n for i, chunk in enumerate(chunks):\n if tag(chunk[-1]) == 'VP' and i > 0 and tag(chunks[i - 1][-1]) == 'NP':\n if chunk[-1][-1] == 'O':\n id += 1\n for token in chunk:\n token[-1] = 'VP-' + str(id)\n for token in chunks[i - 1]:\n token[-1] += '*NP-SBJ-' + str(id)\n token[-1] = token[-1].lstrip('O-*')\n if tag(chunk[-1]) == 'VP' and i < len(chunks) - 1 and tag(chunks[i +\n 1][-1]) == 'NP':\n if chunk[-1][-1] == 'O':\n id += 1\n for token in chunk:\n token[-1] = 'VP-' + str(id)\n for token in chunks[i + 1]:\n token[-1] = '*NP-OBJ-' + str(id)\n token[-1] = token[-1].lstrip('O-*')\n for i, chunk in enumerate(chunks):\n if 0 < i < len(chunks) - 1 and len(chunk) == 1 and chunk[-1][-1\n ] == 'O':\n t0, t1, t2 = chunks[i - 1][-1], chunks[i][0], chunks[i + 1][0]\n if tag(t1) == 'PP' and t2[1] in ('DT', 'PR', 'PRP$'):\n if t0[0] in BE and t1[0] in ('in', 'at'):\n t1[-1] = 'PP-LOC'\n if t0[0] in GO and t1[0] in ('to', 'towards'):\n t1[-1] = 'PP-DIR'\n related = []\n [related.extend(chunk) for chunk in chunks]\n return related\n",
"role": "user"
},
{
"content": "find_relations",
"role": "assistant"
}
] |
def x(string, parser, top=10, frequency={}, **kwargs):
lemmata = kwargs.pop('lemmata', kwargs.pop('stem', True))
chunks = []
wordcount = 0
for sentence in parser.parse(string, chunks=True, lemmata=lemmata).split():
for w in sentence:
if w[2] == 'B-NP':
chunks.append([w])
wordcount += 1
elif w[2] == 'I-NP' and w[1][:3] == chunks[-1][-1][1][:3] == 'NNP':
chunks[-1][-1][+0] += ' ' + w[+0]
chunks[-1][-1][-1] += ' ' + w[-1]
elif w[2] == 'I-NP':
chunks[-1].append(w)
wordcount += 1
m = {}
for i, chunk in enumerate(chunks):
head = True
if parser.language not in ('ca', 'es', 'pt', 'fr', 'it', 'pt', 'ro'):
chunk = list(reversed(chunk))
for w in chunk:
if w[1].startswith('NN'):
if lemmata:
k = w[-1]
else:
k = w[0].lower()
if not k in m:
m[k] = [0.0, set(), 1.0, 1.0, 1.0]
m[k][0] += 1 / float(wordcount)
m[k][1].add(' '.join(map(lambda x: x[0], chunk)).lower())
m[k][2] += 1 if float(i) / len(chunks) <= 0.25 else 0
m[k][3] += 1 if w[3] == 'O' else 0
m[k][4] += 1 if head else 0
head = False
for k in m:
if frequency:
df = frequency.get(k, 0.0)
df = max(df, 1e-10)
df = log(1.0 / df, 2.71828)
else:
df = 1.0
m[k][0] = max(1e-10, m[k][0] * df)
m[k][1] = 1 + float(len(m[k][1]))
hmean = lambda a: len(a) / sum(1.0 / x for x in a)
m = [(hmean(m[k]), k) for k in m]
m = sorted(m, key=lambda x: x[1])
m = sorted(m, key=lambda x: x[0], reverse=True)
m = [k for score, k in m]
return m[:top]
|
find_keywords
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(string, parser, top=10, frequency={}, **kwargs):\n lemmata = kwargs.pop('lemmata', kwargs.pop('stem', True))\n chunks = []\n wordcount = 0\n for sentence in parser.parse(string, chunks=True, lemmata=lemmata).split():\n for w in sentence:\n if w[2] == 'B-NP':\n chunks.append([w])\n wordcount += 1\n elif w[2] == 'I-NP' and w[1][:3] == chunks[-1][-1][1][:3] == 'NNP':\n chunks[-1][-1][+0] += ' ' + w[+0]\n chunks[-1][-1][-1] += ' ' + w[-1]\n elif w[2] == 'I-NP':\n chunks[-1].append(w)\n wordcount += 1\n m = {}\n for i, chunk in enumerate(chunks):\n head = True\n if parser.language not in ('ca', 'es', 'pt', 'fr', 'it', 'pt', 'ro'):\n chunk = list(reversed(chunk))\n for w in chunk:\n if w[1].startswith('NN'):\n if lemmata:\n k = w[-1]\n else:\n k = w[0].lower()\n if not k in m:\n m[k] = [0.0, set(), 1.0, 1.0, 1.0]\n m[k][0] += 1 / float(wordcount)\n m[k][1].add(' '.join(map(lambda x: x[0], chunk)).lower())\n m[k][2] += 1 if float(i) / len(chunks) <= 0.25 else 0\n m[k][3] += 1 if w[3] == 'O' else 0\n m[k][4] += 1 if head else 0\n head = False\n for k in m:\n if frequency:\n df = frequency.get(k, 0.0)\n df = max(df, 1e-10)\n df = log(1.0 / df, 2.71828)\n else:\n df = 1.0\n m[k][0] = max(1e-10, m[k][0] * df)\n m[k][1] = 1 + float(len(m[k][1]))\n hmean = lambda a: len(a) / sum(1.0 / x for x in a)\n m = [(hmean(m[k]), k) for k in m]\n m = sorted(m, key=lambda x: x[1])\n m = sorted(m, key=lambda x: x[0], reverse=True)\n m = [k for score, k in m]\n return m[:top]\n",
"role": "user"
},
{
"content": "find_keywords",
"role": "assistant"
}
] |
def x(*args, **kwargs):
if len(args) == 1 and isinstance(args[0], (list, tuple)):
if args[0] not in ((PRESENT, PARTICIPLE), (PAST, PARTICIPLE)):
args = args[0]
if len(args) == 0 and len(kwargs) == 0:
t = INFINITIVE
else:
t = PRESENT
tense = kwargs.get('tense', args[0] if len(args) > 0 else t)
person = kwargs.get('person', args[1] if len(args) > 1 else 3) or None
number = kwargs.get('number', args[2] if len(args) > 2 else SINGULAR)
mood = kwargs.get('mood', args[3] if len(args) > 3 else INDICATIVE)
aspect = kwargs.get('aspect', args[4] if len(args) > 4 else IMPERFECTIVE)
negated = kwargs.get('negated', args[5] if len(args) > 5 else False)
if mood in (PERFECTIVE, IMPERFECTIVE):
mood, aspect = INDICATIVE, mood
if tense == INFINITIVE:
person = number = mood = aspect = None
negated = False
if tense in ((PRESENT, PARTICIPLE), PRESENT + PARTICIPLE, PARTICIPLE,
GERUND):
tense, aspect = PRESENT, PROGRESSIVE
if tense in ((PAST, PARTICIPLE), PAST + PARTICIPLE):
tense, aspect = PAST, PROGRESSIVE
if tense == IMPERFECT:
tense, aspect = PAST, IMPERFECTIVE
if tense == PRETERITE:
tense, aspect = PAST, PERFECTIVE
if aspect in (CONTINUOUS, PARTICIPLE, GERUND):
aspect = PROGRESSIVE
if aspect == PROGRESSIVE:
person = number = None
if tense == CONDITIONAL and mood == INDICATIVE:
tense, mood = PRESENT, CONDITIONAL
return TENSES_ID.get(tense.lower(), TENSES_ID.get((tense, person,
number, mood, aspect, negated)))
|
tense_id
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(*args, **kwargs):\n if len(args) == 1 and isinstance(args[0], (list, tuple)):\n if args[0] not in ((PRESENT, PARTICIPLE), (PAST, PARTICIPLE)):\n args = args[0]\n if len(args) == 0 and len(kwargs) == 0:\n t = INFINITIVE\n else:\n t = PRESENT\n tense = kwargs.get('tense', args[0] if len(args) > 0 else t)\n person = kwargs.get('person', args[1] if len(args) > 1 else 3) or None\n number = kwargs.get('number', args[2] if len(args) > 2 else SINGULAR)\n mood = kwargs.get('mood', args[3] if len(args) > 3 else INDICATIVE)\n aspect = kwargs.get('aspect', args[4] if len(args) > 4 else IMPERFECTIVE)\n negated = kwargs.get('negated', args[5] if len(args) > 5 else False)\n if mood in (PERFECTIVE, IMPERFECTIVE):\n mood, aspect = INDICATIVE, mood\n if tense == INFINITIVE:\n person = number = mood = aspect = None\n negated = False\n if tense in ((PRESENT, PARTICIPLE), PRESENT + PARTICIPLE, PARTICIPLE,\n GERUND):\n tense, aspect = PRESENT, PROGRESSIVE\n if tense in ((PAST, PARTICIPLE), PAST + PARTICIPLE):\n tense, aspect = PAST, PROGRESSIVE\n if tense == IMPERFECT:\n tense, aspect = PAST, IMPERFECTIVE\n if tense == PRETERITE:\n tense, aspect = PAST, PERFECTIVE\n if aspect in (CONTINUOUS, PARTICIPLE, GERUND):\n aspect = PROGRESSIVE\n if aspect == PROGRESSIVE:\n person = number = None\n if tense == CONDITIONAL and mood == INDICATIVE:\n tense, mood = PRESENT, CONDITIONAL\n return TENSES_ID.get(tense.lower(), TENSES_ID.get((tense, person,\n number, mood, aspect, negated)))\n",
"role": "user"
},
{
"content": "tense_id",
"role": "assistant"
}
] |
def x(function, *args, **kwargs):
return getattr(_module(kwargs.pop('language', 'en')), function)(*args,
**kwargs)
|
multilingual
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(function, *args, **kwargs):\n return getattr(_module(kwargs.pop('language', 'en')), function)(*args,\n **kwargs)\n",
"role": "user"
},
{
"content": "multilingual",
"role": "assistant"
}
] |
def x(s):
s = decode_utf8(s)
s = set(w.strip(PUNCTUATION) for w in s.replace("'", "' ").split())
n = float(len(s) or 1)
p = {}
for xx in LANGUAGES:
lexicon = _module(xx).__dict__['lexicon']
p[xx] = sum(1 for w in s if w in lexicon) / n
return max(p.items(), key=lambda kv: (kv[1], int(kv[0] == 'en')))
|
language
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(s):\n s = decode_utf8(s)\n s = set(w.strip(PUNCTUATION) for w in s.replace(\"'\", \"' \").split())\n n = float(len(s) or 1)\n p = {}\n for xx in LANGUAGES:\n lexicon = _module(xx).__dict__['lexicon']\n p[xx] = sum(1 for w in s if w in lexicon) / n\n return max(p.items(), key=lambda kv: (kv[1], int(kv[0] == 'en')))\n",
"role": "user"
},
{
"content": "language",
"role": "assistant"
}
] |
def x(self, method, *args):
if list.__len__(self) == 0:
self.load()
setattr(self, method, types.MethodType(getattr(list, method), self))
return getattr(list, method)(self, *args)
|
lazy
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, method, *args):\n if list.__len__(self) == 0:\n self.load()\n setattr(self, method, types.MethodType(getattr(list, method), self))\n return getattr(list, method)(self, *args)\n",
"role": "user"
},
{
"content": "lazy",
"role": "assistant"
}
] |
def x(self, token, tag, previous=None, next=None):
self._classifier.train(self._v(token, previous, next), type=tag)
|
train
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, token, tag, previous=None, next=None):\n self._classifier.train(self._v(token, previous, next), type=tag)\n",
"role": "user"
},
{
"content": "train",
"role": "assistant"
}
] |
def x(self, token, previous=None, next=None, **kwargs):
return self._classifier.classify(self._v(token, previous, next), **kwargs)
|
classify
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, token, previous=None, next=None, **kwargs):\n return self._classifier.classify(self._v(token, previous, next), **kwargs)\n",
"role": "user"
},
{
"content": "classify",
"role": "assistant"
}
] |
def x(self, token, previous=(None, None), next=(None, None)):
return [token[0], self._classifier.classify(self._v(token[0], previous,
next))]
|
apply
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, token, previous=(None, None), next=(None, None)):\n return [token[0], self._classifier.classify(self._v(token[0], previous,\n next))]\n",
"role": "user"
},
{
"content": "apply",
"role": "assistant"
}
] |
def x(self, token, previous=None, next=None):
def x(v, s1, s2):
if s2:
v[s1 + ' ' + s2] = 1
p, n = previous, next
p = ('', '') if not p else (p[0] or '', p[1] or '')
n = ('', '') if not n else (n[0] or '', n[1] or '')
v = {}
f(v, 'b', 'b')
f(v, 'h', token[0])
f(v, 'w', token[-6:] if token not in self.known or token in self.
unknown else '')
f(v, 'x', token[-3:])
f(v, '-x', p[0][-3:])
f(v, '+x', n[0][-3:])
f(v, '-t', p[1])
f(v, '-+', p[1] + n[1])
f(v, '+t', n[1])
return v
|
v
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, token, previous=None, next=None):\n\n def x(v, s1, s2):\n if s2:\n v[s1 + ' ' + s2] = 1\n p, n = previous, next\n p = ('', '') if not p else (p[0] or '', p[1] or '')\n n = ('', '') if not n else (n[0] or '', n[1] or '')\n v = {}\n f(v, 'b', 'b')\n f(v, 'h', token[0])\n f(v, 'w', token[-6:] if token not in self.known or token in self.\n unknown else '')\n f(v, 'x', token[-3:])\n f(v, '-x', p[0][-3:])\n f(v, '+x', n[0][-3:])\n f(v, '-t', p[1])\n f(v, '-+', p[1] + n[1])\n f(v, '+t', n[1])\n return v\n",
"role": "user"
},
{
"content": "v",
"role": "assistant"
}
] |
def x(self, token, previous=(None, None), next=(None, None)):
w = token[0]
for r in self:
if r[1] in self._cmd:
f, x, pos, cmd = bool(0), r[0], r[-2], r[1].lower()
if r[2] in self._cmd:
f, x, pos, cmd = bool(1), r[1], r[-2], r[2].lower().lstrip('f')
if f and token[1] != r[0]:
continue
if (cmd == 'word' and x == w or cmd == 'char' and x in w or cmd ==
'haspref' and w.startswith(x) or cmd == 'hassuf' and w.endswith
(x) or cmd == 'addpref' and x + w in self.known or cmd ==
'addsuf' and w + x in self.known or cmd == 'deletepref' and w.
startswith(x) and w[len(x):] in self.known or cmd ==
'deletesuf' and w.endswith(x) and w[:-len(x)] in self.known or
cmd == 'goodleft' and x == next[0] or cmd == 'goodright' and x ==
previous[0]):
token[1] = pos
return token
|
apply
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, token, previous=(None, None), next=(None, None)):\n w = token[0]\n for r in self:\n if r[1] in self._cmd:\n f, x, pos, cmd = bool(0), r[0], r[-2], r[1].lower()\n if r[2] in self._cmd:\n f, x, pos, cmd = bool(1), r[1], r[-2], r[2].lower().lstrip('f')\n if f and token[1] != r[0]:\n continue\n if (cmd == 'word' and x == w or cmd == 'char' and x in w or cmd ==\n 'haspref' and w.startswith(x) or cmd == 'hassuf' and w.endswith\n (x) or cmd == 'addpref' and x + w in self.known or cmd ==\n 'addsuf' and w + x in self.known or cmd == 'deletepref' and w.\n startswith(x) and w[len(x):] in self.known or cmd ==\n 'deletesuf' and w.endswith(x) and w[:-len(x)] in self.known or \n cmd == 'goodleft' and x == next[0] or cmd == 'goodright' and x ==\n previous[0]):\n token[1] = pos\n return token\n",
"role": "user"
},
{
"content": "apply",
"role": "assistant"
}
] |
def x(self, i, tag, affix, cmd='hassuf', tagged=None):
if affix.startswith('-') and affix.endswith('-'):
affix, cmd = affix[+1:-1], 'char'
if affix.startswith('-'):
affix, cmd = affix[+1:-0], 'hassuf'
if affix.endswith('-'):
affix, cmd = affix[+0:-1], 'haspref'
if tagged:
r = [tagged, affix, 'f' + cmd.lstrip('f'), tag, 'x']
else:
r = [affix, cmd.lstrip('f'), tag, 'x']
lazylist.insert(self, i, r)
|
insert
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, i, tag, affix, cmd='hassuf', tagged=None):\n if affix.startswith('-') and affix.endswith('-'):\n affix, cmd = affix[+1:-1], 'char'\n if affix.startswith('-'):\n affix, cmd = affix[+1:-0], 'hassuf'\n if affix.endswith('-'):\n affix, cmd = affix[+0:-1], 'haspref'\n if tagged:\n r = [tagged, affix, 'f' + cmd.lstrip('f'), tag, 'x']\n else:\n r = [affix, cmd.lstrip('f'), tag, 'x']\n lazylist.insert(self, i, r)\n",
"role": "user"
},
{
"content": "insert",
"role": "assistant"
}
] |
def x(self, tokens):
o = [('STAART', 'STAART')] * 3
t = o + tokens + o
for i, token in enumerate(t):
for r in self:
if token[1] == 'STAART':
continue
if token[1] != r[0] and r[0] != '*':
continue
cmd, x, y = r[2], r[3], r[4] if len(r) > 4 else ''
cmd = cmd.lower()
if cmd == 'prevtag' and x == t[i - 1][1
] or cmd == 'nexttag' and x == t[i + 1][1
] or cmd == 'prev2tag' and x == t[i - 2][1
] or cmd == 'next2tag' and x == t[i + 2][1
] or cmd == 'prev1or2tag' and x in (t[i - 1][1], t[i - 2][1]
) or cmd == 'next1or2tag' and x in (t[i + 1][1], t[i + 2][1]
) or cmd == 'prev1or2or3tag' and x in (t[i - 1][1], t[i - 2
][1], t[i - 3][1]) or cmd == 'next1or2or3tag' and x in (t[i +
1][1], t[i + 2][1], t[i + 3][1]
) or cmd == 'surroundtag' and x == t[i - 1][1] and y == t[i + 1
][1] or cmd == 'curwd' and x == t[i + 0][0
] or cmd == 'prevwd' and x == t[i - 1][0
] or cmd == 'nextwd' and x == t[i + 1][0
] or cmd == 'prev1or2wd' and x in (t[i - 1][0], t[i - 2][0]
) or cmd == 'next1or2wd' and x in (t[i + 1][0], t[i + 2][0]
) or cmd == 'prevwdtag' and x == t[i - 1][0] and y == t[i - 1][
1] or cmd == 'nextwdtag' and x == t[i + 1][0] and y == t[i + 1
][1] or cmd == 'wdprevtag' and x == t[i - 1][1] and y == t[
i + 0][0] or cmd == 'wdnexttag' and x == t[i + 0][0
] and y == t[i + 1][1] or cmd == 'wdand2aft' and x == t[i + 0][
0] and y == t[i + 2][0] or cmd == 'wdand2tagbfr' and x == t[
i - 2][1] and y == t[i + 0][0
] or cmd == 'wdand2tagaft' and x == t[i + 0][0] and y == t[
i + 2][1] or cmd == 'lbigram' and x == t[i - 1][0] and y == t[
i + 0][0] or cmd == 'rbigram' and x == t[i + 0][0] and y == t[
i + 1][0] or cmd == 'prevbigram' and x == t[i - 2][1
] and y == t[i - 1][1] or cmd == 'nextbigram' and x == t[i + 1
][1] and y == t[i + 2][1]:
t[i] = [t[i][0], r[1]]
return t[len(o):-len(o)]
|
apply
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, tokens):\n o = [('STAART', 'STAART')] * 3\n t = o + tokens + o\n for i, token in enumerate(t):\n for r in self:\n if token[1] == 'STAART':\n continue\n if token[1] != r[0] and r[0] != '*':\n continue\n cmd, x, y = r[2], r[3], r[4] if len(r) > 4 else ''\n cmd = cmd.lower()\n if cmd == 'prevtag' and x == t[i - 1][1\n ] or cmd == 'nexttag' and x == t[i + 1][1\n ] or cmd == 'prev2tag' and x == t[i - 2][1\n ] or cmd == 'next2tag' and x == t[i + 2][1\n ] or cmd == 'prev1or2tag' and x in (t[i - 1][1], t[i - 2][1]\n ) or cmd == 'next1or2tag' and x in (t[i + 1][1], t[i + 2][1]\n ) or cmd == 'prev1or2or3tag' and x in (t[i - 1][1], t[i - 2\n ][1], t[i - 3][1]) or cmd == 'next1or2or3tag' and x in (t[i +\n 1][1], t[i + 2][1], t[i + 3][1]\n ) or cmd == 'surroundtag' and x == t[i - 1][1] and y == t[i + 1\n ][1] or cmd == 'curwd' and x == t[i + 0][0\n ] or cmd == 'prevwd' and x == t[i - 1][0\n ] or cmd == 'nextwd' and x == t[i + 1][0\n ] or cmd == 'prev1or2wd' and x in (t[i - 1][0], t[i - 2][0]\n ) or cmd == 'next1or2wd' and x in (t[i + 1][0], t[i + 2][0]\n ) or cmd == 'prevwdtag' and x == t[i - 1][0] and y == t[i - 1][\n 1] or cmd == 'nextwdtag' and x == t[i + 1][0] and y == t[i + 1\n ][1] or cmd == 'wdprevtag' and x == t[i - 1][1] and y == t[\n i + 0][0] or cmd == 'wdnexttag' and x == t[i + 0][0\n ] and y == t[i + 1][1] or cmd == 'wdand2aft' and x == t[i + 0][\n 0] and y == t[i + 2][0] or cmd == 'wdand2tagbfr' and x == t[\n i - 2][1] and y == t[i + 0][0\n ] or cmd == 'wdand2tagaft' and x == t[i + 0][0] and y == t[\n i + 2][1] or cmd == 'lbigram' and x == t[i - 1][0] and y == t[\n i + 0][0] or cmd == 'rbigram' and x == t[i + 0][0] and y == t[\n i + 1][0] or cmd == 'prevbigram' and x == t[i - 2][1\n ] and y == t[i - 1][1] or cmd == 'nextbigram' and x == t[i + 1\n ][1] and y == t[i + 2][1]:\n t[i] = [t[i][0], r[1]]\n return t[len(o):-len(o)]\n",
"role": "user"
},
{
"content": "apply",
"role": "assistant"
}
] |
def x(self, i, tag1, tag2, cmd='prevtag', x=None, y=None):
if ' < ' in tag1 and not x and not y:
tag1, x = tag1.split(' < ')
cmd = 'prevtag'
if ' > ' in tag1 and not x and not y:
x, tag1 = tag1.split(' > ')
cmd = 'nexttag'
lazylist.insert(self, i, [tag1, tag2, cmd, x or '', y or ''])
|
insert
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, i, tag1, tag2, cmd='prevtag', x=None, y=None):\n if ' < ' in tag1 and not x and not y:\n tag1, x = tag1.split(' < ')\n cmd = 'prevtag'\n if ' > ' in tag1 and not x and not y:\n x, tag1 = tag1.split(' > ')\n cmd = 'nexttag'\n lazylist.insert(self, i, [tag1, tag2, cmd, x or '', y or ''])\n",
"role": "user"
},
{
"content": "insert",
"role": "assistant"
}
] |
def x(self, tokens):
i = 0
while i < len(tokens):
w = tokens[i][0].lower()
if RE_ENTITY1.match(w) or RE_ENTITY2.match(w) or RE_ENTITY3.match(w):
tokens[i][1] = self.tag
if w in self:
for e in self[w]:
e, tag = (e[:-1], '-' + e[-1].upper()) if e[-1
] in self._cmd else (e, '')
b = True
for j, e in enumerate(e):
if i + j >= len(tokens) or tokens[i + j][0].lower() != e:
b = False
break
if b:
for token in tokens[i:i + j + 1]:
token[1] = token[1] if token[1].startswith(self.tag
) else self.tag
token[1] += tag
i += j
break
i += 1
return tokens
|
apply
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, tokens):\n i = 0\n while i < len(tokens):\n w = tokens[i][0].lower()\n if RE_ENTITY1.match(w) or RE_ENTITY2.match(w) or RE_ENTITY3.match(w):\n tokens[i][1] = self.tag\n if w in self:\n for e in self[w]:\n e, tag = (e[:-1], '-' + e[-1].upper()) if e[-1\n ] in self._cmd else (e, '')\n b = True\n for j, e in enumerate(e):\n if i + j >= len(tokens) or tokens[i + j][0].lower() != e:\n b = False\n break\n if b:\n for token in tokens[i:i + j + 1]:\n token[1] = token[1] if token[1].startswith(self.tag\n ) else self.tag\n token[1] += tag\n i += j\n break\n i += 1\n return tokens\n",
"role": "user"
},
{
"content": "apply",
"role": "assistant"
}
] |
def x(self, entity, name='pers'):
e = map(lambda s: s.lower(), entity.split(' ') + [name])
self.setdefault(e[0], []).append(e)
|
append
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, entity, name='pers'):\n e = map(lambda s: s.lower(), entity.split(' ') + [name])\n self.setdefault(e[0], []).append(e)\n",
"role": "user"
},
{
"content": "append",
"role": "assistant"
}
] |
def x(self, string, **kwargs):
return find_keywords(string, parser=self, top=kwargs.pop('top', 10),
frequency=kwargs.pop('frequency', {}), **kwargs)
|
find_keywords
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, string, **kwargs):\n return find_keywords(string, parser=self, top=kwargs.pop('top', 10),\n frequency=kwargs.pop('frequency', {}), **kwargs)\n",
"role": "user"
},
{
"content": "find_keywords",
"role": "assistant"
}
] |
def x(self, string, **kwargs):
return find_tokens(string, punctuation=kwargs.get('punctuation',
PUNCTUATION), abbreviations=kwargs.get('abbreviations',
ABBREVIATIONS), replace=kwargs.get('replace', replacements),
linebreak='\\n{2,}')
|
find_tokens
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, string, **kwargs):\n return find_tokens(string, punctuation=kwargs.get('punctuation',\n PUNCTUATION), abbreviations=kwargs.get('abbreviations',\n ABBREVIATIONS), replace=kwargs.get('replace', replacements),\n linebreak='\\\\n{2,}')\n",
"role": "user"
},
{
"content": "find_tokens",
"role": "assistant"
}
] |
def x(self, tokens, **kwargs):
return find_tags(tokens, lexicon=kwargs.get('lexicon', self.lexicon or
{}), model=kwargs.get('model', self.model), morphology=kwargs.get(
'morphology', self.morphology), context=kwargs.get('context', self.
context), entities=kwargs.get('entities', self.entities), language=
kwargs.get('language', self.language), default=kwargs.get('default',
self.default), map=kwargs.get('map', None))
|
find_tags
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, tokens, **kwargs):\n return find_tags(tokens, lexicon=kwargs.get('lexicon', self.lexicon or\n {}), model=kwargs.get('model', self.model), morphology=kwargs.get(\n 'morphology', self.morphology), context=kwargs.get('context', self.\n context), entities=kwargs.get('entities', self.entities), language=\n kwargs.get('language', self.language), default=kwargs.get('default',\n self.default), map=kwargs.get('map', None))\n",
"role": "user"
},
{
"content": "find_tags",
"role": "assistant"
}
] |
def x(self, tokens, **kwargs):
return find_prepositions(find_chunks(tokens, language=kwargs.get(
'language', self.language)))
|
find_chunks
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, tokens, **kwargs):\n return find_prepositions(find_chunks(tokens, language=kwargs.get(\n 'language', self.language)))\n",
"role": "user"
},
{
"content": "find_chunks",
"role": "assistant"
}
] |
def x(self, s, tokenize=True, tags=True, chunks=True, relations=False,
lemmata=False, encoding='utf-8', **kwargs):
if tokenize is True:
s = self.find_tokens(s, **kwargs)
if isinstance(s, (list, tuple)):
s = [(isinstance(s, basestring) and s.split(' ') or s) for s in s]
if isinstance(s, basestring):
s = [s.split(' ') for s in s.split('\n')]
for i in range(len(s)):
for j in range(len(s[i])):
if isinstance(s[i][j], str):
s[i][j] = decode_string(s[i][j], encoding)
if tags or chunks or relations or lemmata:
s[i] = self.find_tags(s[i], **kwargs)
else:
s[i] = [[w] for w in s[i]]
if chunks or relations:
s[i] = self.find_chunks(s[i], **kwargs)
if relations:
s[i] = self.find_labels(s[i], **kwargs)
if lemmata:
s[i] = self.find_lemmata(s[i], **kwargs)
if not kwargs.get('collapse', True) or kwargs.get('split', False):
return s
format = ['word']
if tags:
format.append('part-of-speech')
if chunks:
format.extend(('chunk', 'preposition'))
if relations:
format.append('relation')
if lemmata:
format.append('lemma')
for i in range(len(s)):
for j in range(len(s[i])):
s[i][j][0] = s[i][j][0].replace('/', '&slash;')
s[i][j] = '/'.join(s[i][j])
s[i] = ' '.join(s[i])
s = '\n'.join(s)
s = TaggedString(s, format, language=kwargs.get('language', self.language))
return s
|
parse
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, s, tokenize=True, tags=True, chunks=True, relations=False,\n lemmata=False, encoding='utf-8', **kwargs):\n if tokenize is True:\n s = self.find_tokens(s, **kwargs)\n if isinstance(s, (list, tuple)):\n s = [(isinstance(s, basestring) and s.split(' ') or s) for s in s]\n if isinstance(s, basestring):\n s = [s.split(' ') for s in s.split('\\n')]\n for i in range(len(s)):\n for j in range(len(s[i])):\n if isinstance(s[i][j], str):\n s[i][j] = decode_string(s[i][j], encoding)\n if tags or chunks or relations or lemmata:\n s[i] = self.find_tags(s[i], **kwargs)\n else:\n s[i] = [[w] for w in s[i]]\n if chunks or relations:\n s[i] = self.find_chunks(s[i], **kwargs)\n if relations:\n s[i] = self.find_labels(s[i], **kwargs)\n if lemmata:\n s[i] = self.find_lemmata(s[i], **kwargs)\n if not kwargs.get('collapse', True) or kwargs.get('split', False):\n return s\n format = ['word']\n if tags:\n format.append('part-of-speech')\n if chunks:\n format.extend(('chunk', 'preposition'))\n if relations:\n format.append('relation')\n if lemmata:\n format.append('lemma')\n for i in range(len(s)):\n for j in range(len(s[i])):\n s[i][j][0] = s[i][j][0].replace('/', '&slash;')\n s[i][j] = '/'.join(s[i][j])\n s[i] = ' '.join(s[i])\n s = '\\n'.join(s)\n s = TaggedString(s, format, language=kwargs.get('language', self.language))\n return s\n",
"role": "user"
},
{
"content": "parse",
"role": "assistant"
}
] |
def x(self, sep=TOKENS):
if sep != TOKENS:
return unicode.split(self, sep)
if len(self) == 0:
return []
return [[[x.replace('&slash;', '/') for x in token.split('/')] for
token in sentence.split(' ')] for sentence in unicode.split(self, '\n')
]
|
split
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, sep=TOKENS):\n if sep != TOKENS:\n return unicode.split(self, sep)\n if len(self) == 0:\n return []\n return [[[x.replace('&slash;', '/') for x in token.split('/')] for\n token in sentence.split(' ')] for sentence in unicode.split(self, '\\n')\n ]\n",
"role": "user"
},
{
"content": "split",
"role": "assistant"
}
] |
def x(self):
a = set(TENSES[id] for id in self._format)
a = a.union(set(TENSES[id] for id in self._default.keys()))
a = a.union(set(TENSES[id] for id in self._default.values()))
a = sorted(x[:-2] for x in a if x[-2] is False)
return a
|
TENSES
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n a = set(TENSES[id] for id in self._format)\n a = a.union(set(TENSES[id] for id in self._default.keys()))\n a = a.union(set(TENSES[id] for id in self._default.values()))\n a = sorted(x[:-2] for x in a if x[-2] is False)\n return a\n",
"role": "user"
},
{
"content": "TENSES",
"role": "assistant"
}
] |
def x(self, verb, parse=True):
if dict.__len__(self) == 0:
self.load()
if verb.lower() in self._inverse:
return self._inverse[verb.lower()]
if verb in self._inverse:
return self._inverse[verb]
if parse is True:
return self.find_lemma(verb)
|
lemma
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, verb, parse=True):\n if dict.__len__(self) == 0:\n self.load()\n if verb.lower() in self._inverse:\n return self._inverse[verb.lower()]\n if verb in self._inverse:\n return self._inverse[verb]\n if parse is True:\n return self.find_lemma(verb)\n",
"role": "user"
},
{
"content": "lemma",
"role": "assistant"
}
] |
def x(self, verb, parse=True):
a = []
b = self.lemma(verb, parse=parse)
if b in self:
a = [x for x in self[b] if x != '']
elif parse is True:
a = self.find_lexeme(b)
u = []
[u.append(x) for x in a if x not in u]
return u
|
lexeme
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, verb, parse=True):\n a = []\n b = self.lemma(verb, parse=parse)\n if b in self:\n a = [x for x in self[b] if x != '']\n elif parse is True:\n a = self.find_lexeme(b)\n u = []\n [u.append(x) for x in a if x not in u]\n return u\n",
"role": "user"
},
{
"content": "lexeme",
"role": "assistant"
}
] |
def x(self, verb, *args, **kwargs):
id = tense_id(*args, **kwargs)
i1 = self._format.get(id)
i2 = self._format.get(self._default.get(id))
i3 = self._format.get(self._default.get(self._default.get(id)))
b = self.lemma(verb, parse=kwargs.get('parse', True))
v = []
if b in self:
v = self[b]
for i in (i1, i2, i3):
if i is not None and 0 <= i < len(v) and v[i]:
return v[i]
if kwargs.get('parse', True) is True:
v = self.find_lexeme(b)
for i in (i1, i2, i3):
if i is not None and 0 <= i < len(v) and v[i]:
return v[i]
|
conjugate
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, verb, *args, **kwargs):\n id = tense_id(*args, **kwargs)\n i1 = self._format.get(id)\n i2 = self._format.get(self._default.get(id))\n i3 = self._format.get(self._default.get(self._default.get(id)))\n b = self.lemma(verb, parse=kwargs.get('parse', True))\n v = []\n if b in self:\n v = self[b]\n for i in (i1, i2, i3):\n if i is not None and 0 <= i < len(v) and v[i]:\n return v[i]\n if kwargs.get('parse', True) is True:\n v = self.find_lexeme(b)\n for i in (i1, i2, i3):\n if i is not None and 0 <= i < len(v) and v[i]:\n return v[i]\n",
"role": "user"
},
{
"content": "conjugate",
"role": "assistant"
}
] |
def x(self, verb, parse=True):
verb = verb.lower()
a = set()
b = self.lemma(verb, parse=parse)
v = []
if b in self:
v = self[b]
elif parse is True:
v = self.find_lexeme(b)
for i, tense in enumerate(v):
if tense == verb:
for id, index in self._format.items():
if i == index:
a.add(id)
for id1, id2 in self._default.items():
if id2 in a:
a.add(id1)
for id1, id2 in self._default.items():
if id2 in a:
a.add(id1)
a = (TENSES[id][:-2] for id in a)
a = Tenses(sorted(a))
return a
|
tenses
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, verb, parse=True):\n verb = verb.lower()\n a = set()\n b = self.lemma(verb, parse=parse)\n v = []\n if b in self:\n v = self[b]\n elif parse is True:\n v = self.find_lexeme(b)\n for i, tense in enumerate(v):\n if tense == verb:\n for id, index in self._format.items():\n if i == index:\n a.add(id)\n for id1, id2 in self._default.items():\n if id2 in a:\n a.add(id1)\n for id1, id2 in self._default.items():\n if id2 in a:\n a.add(id1)\n a = (TENSES[id][:-2] for id in a)\n a = Tenses(sorted(a))\n return a\n",
"role": "user"
},
{
"content": "tenses",
"role": "assistant"
}
] |
def x(self, path=None):
if not path:
path = self._path
if not os.path.exists(path):
return
words, synsets, labels = {}, {}, {}
xml = cElementTree.parse(path)
xml = xml.getroot()
for w in xml.findall('word'):
if self._confidence is None or self._confidence <= float(w.attrib.
get('confidence', 0.0)):
w, pos, p, s, i, label, synset = w.attrib.get('form'
), w.attrib.get('pos'), w.attrib.get('polarity', 0.0
), w.attrib.get('subjectivity', 0.0), w.attrib.get('intensity',
1.0), w.attrib.get('label'), w.attrib.get(self._synset)
psi = float(p), float(s), float(i)
if w:
words.setdefault(w, {}).setdefault(pos, []).append(psi)
if w and label:
labels[w] = label
if synset:
synsets.setdefault(synset, []).append(psi)
self._language = xml.attrib.get('language', self._language)
for w in words:
words[w] = dict((pos, [avg(each) for each in zip(*psi)]) for pos,
psi in words[w].items())
for w, pos in list(words.items()):
words[w][None] = [avg(each) for each in zip(*pos.values())]
for id, psi in synsets.items():
synsets[id] = [avg(each) for each in zip(*psi)]
dict.update(self, words)
dict.update(self.labeler, labels)
dict.update(self._synsets, synsets)
|
load
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, path=None):\n if not path:\n path = self._path\n if not os.path.exists(path):\n return\n words, synsets, labels = {}, {}, {}\n xml = cElementTree.parse(path)\n xml = xml.getroot()\n for w in xml.findall('word'):\n if self._confidence is None or self._confidence <= float(w.attrib.\n get('confidence', 0.0)):\n w, pos, p, s, i, label, synset = w.attrib.get('form'\n ), w.attrib.get('pos'), w.attrib.get('polarity', 0.0\n ), w.attrib.get('subjectivity', 0.0), w.attrib.get('intensity',\n 1.0), w.attrib.get('label'), w.attrib.get(self._synset)\n psi = float(p), float(s), float(i)\n if w:\n words.setdefault(w, {}).setdefault(pos, []).append(psi)\n if w and label:\n labels[w] = label\n if synset:\n synsets.setdefault(synset, []).append(psi)\n self._language = xml.attrib.get('language', self._language)\n for w in words:\n words[w] = dict((pos, [avg(each) for each in zip(*psi)]) for pos,\n psi in words[w].items())\n for w, pos in list(words.items()):\n words[w][None] = [avg(each) for each in zip(*pos.values())]\n for id, psi in synsets.items():\n synsets[id] = [avg(each) for each in zip(*psi)]\n dict.update(self, words)\n dict.update(self.labeler, labels)\n dict.update(self._synsets, synsets)\n",
"role": "user"
},
{
"content": "load",
"role": "assistant"
}
] |
def x(self, id, pos=ADJECTIVE):
id = str(id).zfill(8)
if not id.startswith(('n-', 'v-', 'a-', 'r-')):
if pos == NOUN:
id = 'n-' + id
if pos == VERB:
id = 'v-' + id
if pos == ADJECTIVE:
id = 'a-' + id
if pos == ADVERB:
id = 'r-' + id
if dict.__len__(self) == 0:
self.load()
try:
return tuple(self._synsets[id])[:2]
except KeyError:
return tuple(self._synsets.get(re.sub('-0+', '-', id), (0.0, 0.0))[:2])
|
synset
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, id, pos=ADJECTIVE):\n id = str(id).zfill(8)\n if not id.startswith(('n-', 'v-', 'a-', 'r-')):\n if pos == NOUN:\n id = 'n-' + id\n if pos == VERB:\n id = 'v-' + id\n if pos == ADJECTIVE:\n id = 'a-' + id\n if pos == ADVERB:\n id = 'r-' + id\n if dict.__len__(self) == 0:\n self.load()\n try:\n return tuple(self._synsets[id])[:2]\n except KeyError:\n return tuple(self._synsets.get(re.sub('-0+', '-', id), (0.0, 0.0))[:2])\n",
"role": "user"
},
{
"content": "synset",
"role": "assistant"
}
] |
def x(self, words=[], negation=True):
a = []
m = None
n = None
for w, pos in words:
if w is None:
continue
if w in self and pos in self[w]:
p, s, i = self[w][pos]
if m is None:
a.append(dict(w=[w], p=p, s=s, i=i, n=1, x=self.labeler.get(w))
)
if m is not None:
a[-1]['w'].append(w)
a[-1]['p'] = max(-1.0, min(p * a[-1]['i'], +1.0))
a[-1]['s'] = max(-1.0, min(s * a[-1]['i'], +1.0))
a[-1]['i'] = i
a[-1]['x'] = self.labeler.get(w)
if n is not None:
a[-1]['w'].insert(0, n)
a[-1]['i'] = 1.0 / a[-1]['i']
a[-1]['n'] = -1
m = None
n = None
if pos and pos in self.modifiers or any(map(self[w].
__contains__, self.modifiers)):
m = w, pos
if negation and w in self.negations:
n = w
else:
if negation and w in self.negations:
n = w
elif n and len(w.strip("'")) > 1:
n = None
if n is not None and m is not None and (pos in self.modifiers or
self.modifier(m[0])):
a[-1]['w'].append(n)
a[-1]['n'] = -1
n = None
elif m and len(w) > 2:
m = None
if w == '!' and len(a) > 0:
a[-1]['w'].append('!')
a[-1]['p'] = max(-1.0, min(a[-1]['p'] * 1.25, +1.0))
if w == '(!)':
a.append(dict(w=[w], p=0.0, s=1.0, i=1.0, n=1, x=IRONY))
if w.isalpha() is False and len(w) <= 5 and w not in PUNCTUATION:
for (type, p), e in EMOTICONS.items():
if w in map(lambda e: e.lower(), e):
a.append(dict(w=[w], p=p, s=1.0, i=1.0, n=1, x=MOOD))
break
for i in range(len(a)):
w = a[i]['w']
p = a[i]['p']
s = a[i]['s']
n = a[i]['n']
x = a[i]['x']
a[i] = w, p * -0.5 if n < 0 else p, s, x
return a
|
assessments
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, words=[], negation=True):\n a = []\n m = None\n n = None\n for w, pos in words:\n if w is None:\n continue\n if w in self and pos in self[w]:\n p, s, i = self[w][pos]\n if m is None:\n a.append(dict(w=[w], p=p, s=s, i=i, n=1, x=self.labeler.get(w))\n )\n if m is not None:\n a[-1]['w'].append(w)\n a[-1]['p'] = max(-1.0, min(p * a[-1]['i'], +1.0))\n a[-1]['s'] = max(-1.0, min(s * a[-1]['i'], +1.0))\n a[-1]['i'] = i\n a[-1]['x'] = self.labeler.get(w)\n if n is not None:\n a[-1]['w'].insert(0, n)\n a[-1]['i'] = 1.0 / a[-1]['i']\n a[-1]['n'] = -1\n m = None\n n = None\n if pos and pos in self.modifiers or any(map(self[w].\n __contains__, self.modifiers)):\n m = w, pos\n if negation and w in self.negations:\n n = w\n else:\n if negation and w in self.negations:\n n = w\n elif n and len(w.strip(\"'\")) > 1:\n n = None\n if n is not None and m is not None and (pos in self.modifiers or\n self.modifier(m[0])):\n a[-1]['w'].append(n)\n a[-1]['n'] = -1\n n = None\n elif m and len(w) > 2:\n m = None\n if w == '!' and len(a) > 0:\n a[-1]['w'].append('!')\n a[-1]['p'] = max(-1.0, min(a[-1]['p'] * 1.25, +1.0))\n if w == '(!)':\n a.append(dict(w=[w], p=0.0, s=1.0, i=1.0, n=1, x=IRONY))\n if w.isalpha() is False and len(w) <= 5 and w not in PUNCTUATION:\n for (type, p), e in EMOTICONS.items():\n if w in map(lambda e: e.lower(), e):\n a.append(dict(w=[w], p=p, s=1.0, i=1.0, n=1, x=MOOD))\n break\n for i in range(len(a)):\n w = a[i]['w']\n p = a[i]['p']\n s = a[i]['s']\n n = a[i]['n']\n x = a[i]['x']\n a[i] = w, p * -0.5 if n < 0 else p, s, x\n return a\n",
"role": "user"
},
{
"content": "assessments",
"role": "assistant"
}
] |
def x(self, word, pos=None, polarity=0.0, subjectivity=0.0, intensity=1.0,
label=None):
w = self.setdefault(word, {})
w[pos] = w[None] = polarity, subjectivity, intensity
if label:
self.labeler[word] = label
|
annotate
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, word, pos=None, polarity=0.0, subjectivity=0.0, intensity=1.0,\n label=None):\n w = self.setdefault(word, {})\n w[pos] = w[None] = polarity, subjectivity, intensity\n if label:\n self.labeler[word] = label\n",
"role": "user"
},
{
"content": "annotate",
"role": "assistant"
}
] |
def x(self, s, path='spelling.txt'):
model = {}
for w in re.findall('[a-z]+', s.lower()):
model[w] = w in model and model[w] + 1 or 1
model = ('%s %s' % (k, v) for k, v in sorted(model.items()))
model = '\n'.join(model)
f = open(path, 'w')
f.write(model)
f.close()
|
train
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, s, path='spelling.txt'):\n model = {}\n for w in re.findall('[a-z]+', s.lower()):\n model[w] = w in model and model[w] + 1 or 1\n model = ('%s %s' % (k, v) for k, v in sorted(model.items()))\n model = '\\n'.join(model)\n f = open(path, 'w')\n f.write(model)\n f.close()\n",
"role": "user"
},
{
"content": "train",
"role": "assistant"
}
] |
def x(self, w):
split = [(w[:i], w[i:]) for i in range(len(w) + 1)]
delete, transpose, replace, insert = [(a + b[1:]) for a, b in split if b
], [(a + b[1] + b[0] + b[2:]) for a, b in split if len(b) > 1], [(a +
c + b[1:]) for a, b in split for c in Spelling.ALPHA if b], [(a + c +
b[0:]) for a, b in split for c in Spelling.ALPHA]
return set(delete + transpose + replace + insert)
|
edit1
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, w):\n split = [(w[:i], w[i:]) for i in range(len(w) + 1)]\n delete, transpose, replace, insert = [(a + b[1:]) for a, b in split if b\n ], [(a + b[1] + b[0] + b[2:]) for a, b in split if len(b) > 1], [(a +\n c + b[1:]) for a, b in split for c in Spelling.ALPHA if b], [(a + c +\n b[0:]) for a, b in split for c in Spelling.ALPHA]\n return set(delete + transpose + replace + insert)\n",
"role": "user"
},
{
"content": "edit1",
"role": "assistant"
}
] |
def x(self, w):
return set(e2 for e1 in self._edit1(w) for e2 in self._edit1(e1) if e2 in
self)
|
edit2
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, w):\n return set(e2 for e1 in self._edit1(w) for e2 in self._edit1(e1) if e2 in\n self)\n",
"role": "user"
},
{
"content": "edit2",
"role": "assistant"
}
] |
def x(self, w):
if len(self) == 0:
self.load()
if len(w) == 1:
return [(w, 1.0)]
if w in PUNCTUATION:
return [(w, 1.0)]
if w.replace('.', '').isdigit():
return [(w, 1.0)]
candidates = self._known([w]) or self._known(self._edit1(w)
) or self._known(self._edit2(w)) or [w]
candidates = [(self.get(c, 0.0), c) for c in candidates]
s = float(sum(p for p, w in candidates) or 1)
candidates = sorted(((p / s, w) for p, w in candidates), reverse=True)
candidates = [(w.istitle() and x.title() or x, p) for p, x in candidates]
return candidates
|
suggest
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, w):\n if len(self) == 0:\n self.load()\n if len(w) == 1:\n return [(w, 1.0)]\n if w in PUNCTUATION:\n return [(w, 1.0)]\n if w.replace('.', '').isdigit():\n return [(w, 1.0)]\n candidates = self._known([w]) or self._known(self._edit1(w)\n ) or self._known(self._edit2(w)) or [w]\n candidates = [(self.get(c, 0.0), c) for c in candidates]\n s = float(sum(p for p, w in candidates) or 1)\n candidates = sorted(((p / s, w) for p, w in candidates), reverse=True)\n candidates = [(w.istitle() and x.title() or x, p) for p, x in candidates]\n return candidates\n",
"role": "user"
},
{
"content": "suggest",
"role": "assistant"
}
] |
def x(*args, **kwargs):
args = [list(iterable) for iterable in args]
n = max(map(len, args))
v = kwargs.get('default', None)
return _zip(*[(i + [v] * (n - len(i))) for i in args])
|
zip
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(*args, **kwargs):\n args = [list(iterable) for iterable in args]\n n = max(map(len, args))\n v = kwargs.get('default', None)\n return _zip(*[(i + [v] * (n - len(i))) for i in args])\n",
"role": "user"
},
{
"content": "zip",
"role": "assistant"
}
] |
def x(sentence):
chunks = []
for word in sentence:
if word.chunk is not None:
if len(chunks) == 0 or chunks[-1] != word.chunk:
chunks.append(word.chunk)
else:
ch = Chink(sentence)
ch.append(word.copy(ch))
chunks.append(ch)
return chunks
|
chunked
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(sentence):\n chunks = []\n for word in sentence:\n if word.chunk is not None:\n if len(chunks) == 0 or chunks[-1] != word.chunk:\n chunks.append(word.chunk)\n else:\n ch = Chink(sentence)\n ch.append(word.copy(ch))\n chunks.append(ch)\n return chunks\n",
"role": "user"
},
{
"content": "chunked",
"role": "assistant"
}
] |
def x(string, token=[WORD, POS, CHUNK, PNP, REL, ANCHOR, LEMMA]):
return Text(string, token)
|
tree
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(string, token=[WORD, POS, CHUNK, PNP, REL, ANCHOR, LEMMA]):\n return Text(string, token)\n",
"role": "user"
},
{
"content": "tree",
"role": "assistant"
}
] |
def x(string):
string = string.replace('&', '&')
string = string.replace('<', '<')
string = string.replace('>', '>')
string = string.replace('"', '"')
string = string.replace(SLASH, '/')
return string
|
xml_encode
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(string):\n string = string.replace('&', '&')\n string = string.replace('<', '<')\n string = string.replace('>', '>')\n string = string.replace('\"', '"')\n string = string.replace(SLASH, '/')\n return string\n",
"role": "user"
},
{
"content": "xml_encode",
"role": "assistant"
}
] |
def x(string):
string = string.replace('&', '&')
string = string.replace('<', '<')
string = string.replace('>', '>')
string = string.replace('"', '"')
string = string.replace('/', SLASH)
return string
|
xml_decode
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(string):\n string = string.replace('&', '&')\n string = string.replace('<', '<')\n string = string.replace('>', '>')\n string = string.replace('"', '\"')\n string = string.replace('/', SLASH)\n return string\n",
"role": "user"
},
{
"content": "xml_decode",
"role": "assistant"
}
] |
def x(sentence, tab='\t', id=''):
uid = lambda *parts: ''.join([str(id), _UID_SEPARATOR] + [str(x) for x in
parts]).lstrip(_UID_SEPARATOR)
push = lambda indent: indent + tab
pop = lambda indent: indent[:-len(tab)]
indent = tab
xml = []
xml.append('<%s%s %s="%s" %s="%s">' % (XML_SENTENCE, XML_ID and
' %s="%s"' % (XML_ID, str(id)) or '', XML_TOKEN, ', '.join(sentence
.token), XML_LANGUAGE, sentence.language))
anchors = {}
for chunk in sentence.chunks:
if chunk.attachments:
anchors[chunk.start] = len(anchors) + 1
for word in sentence.words:
chunk = word.chunk
pnp = word.chunk and word.chunk.pnp or None
if pnp and pnp.start == chunk.start:
a = pnp.anchor and ' %s="%s"' % (XML_OF, uid('A', anchors.get(
pnp.anchor.start, ''))) or ''
xml.append(indent + '<%s %s="PNP"%s>' % (XML_CHUNK, XML_TYPE, a))
indent = push(indent)
if chunk and chunk.start == word.index:
if chunk.relations:
r1 = unzip(0, chunk.relations)
r2 = unzip(1, chunk.relations)
r1 = [(x is None and '-' or uid(x)) for x in r1]
r2 = [(x is None and '-' or x) for x in r2]
r1 = not len(unique(r1)) == 1 and '|'.join(r1) or (r1 + [None]
)[0]
r2 = not len(unique(r2)) == 1 and '|'.join(r2) or (r2 + [None]
)[0]
xml.append(indent + '<%s%s%s%s%s%s>' % (XML_CHUNK, chunk.type and
' %s="%s"' % (XML_TYPE, chunk.type) or '', chunk.relations and
chunk.role != None and ' %s="%s"' % (XML_RELATION, r2) or
'', chunk.relation and chunk.type == 'VP' and ' %s="%s"' %
(XML_ID, uid(chunk.relation)) or '', chunk.relation and
chunk.type != 'VP' and ' %s="%s"' % (XML_OF, r1) or '',
chunk.attachments and ' %s="%s"' % (XML_ANCHOR, uid('A',
anchors[chunk.start])) or ''))
indent = push(indent)
if not chunk:
xml.append(indent + '<%s>' % XML_CHINK)
indent = push(indent)
xml.append(indent + '<%s%s%s%s>%s</%s>' % (XML_WORD, word.type and
' %s="%s"' % (XML_TYPE, xml_encode(word.type)) or '', word.
lemma and ' %s="%s"' % (XML_LEMMA, xml_encode(word.lemma)) or
'', (' ' + ' '.join([('%s="%s"' % (k, v)) for k, v in word.
custom_tags.items() if v != None])).rstrip(), xml_encode(
unicode(word)), XML_WORD))
if not chunk:
indent = pop(indent)
xml.append(indent + '</%s>' % XML_CHINK)
if chunk and chunk.stop - 1 == word.index:
indent = pop(indent)
xml.append(indent + '</%s>' % XML_CHUNK)
if pnp and pnp.stop - 1 == word.index:
indent = pop(indent)
xml.append(indent + '</%s>' % XML_CHUNK)
xml.append('</%s>' % XML_SENTENCE)
return '\n'.join(xml).encode('utf-8')
|
parse_xml
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(sentence, tab='\\t', id=''):\n uid = lambda *parts: ''.join([str(id), _UID_SEPARATOR] + [str(x) for x in\n parts]).lstrip(_UID_SEPARATOR)\n push = lambda indent: indent + tab\n pop = lambda indent: indent[:-len(tab)]\n indent = tab\n xml = []\n xml.append('<%s%s %s=\"%s\" %s=\"%s\">' % (XML_SENTENCE, XML_ID and \n ' %s=\"%s\"' % (XML_ID, str(id)) or '', XML_TOKEN, ', '.join(sentence\n .token), XML_LANGUAGE, sentence.language))\n anchors = {}\n for chunk in sentence.chunks:\n if chunk.attachments:\n anchors[chunk.start] = len(anchors) + 1\n for word in sentence.words:\n chunk = word.chunk\n pnp = word.chunk and word.chunk.pnp or None\n if pnp and pnp.start == chunk.start:\n a = pnp.anchor and ' %s=\"%s\"' % (XML_OF, uid('A', anchors.get(\n pnp.anchor.start, ''))) or ''\n xml.append(indent + '<%s %s=\"PNP\"%s>' % (XML_CHUNK, XML_TYPE, a))\n indent = push(indent)\n if chunk and chunk.start == word.index:\n if chunk.relations:\n r1 = unzip(0, chunk.relations)\n r2 = unzip(1, chunk.relations)\n r1 = [(x is None and '-' or uid(x)) for x in r1]\n r2 = [(x is None and '-' or x) for x in r2]\n r1 = not len(unique(r1)) == 1 and '|'.join(r1) or (r1 + [None]\n )[0]\n r2 = not len(unique(r2)) == 1 and '|'.join(r2) or (r2 + [None]\n )[0]\n xml.append(indent + '<%s%s%s%s%s%s>' % (XML_CHUNK, chunk.type and\n ' %s=\"%s\"' % (XML_TYPE, chunk.type) or '', chunk.relations and\n chunk.role != None and ' %s=\"%s\"' % (XML_RELATION, r2) or\n '', chunk.relation and chunk.type == 'VP' and ' %s=\"%s\"' %\n (XML_ID, uid(chunk.relation)) or '', chunk.relation and \n chunk.type != 'VP' and ' %s=\"%s\"' % (XML_OF, r1) or '', \n chunk.attachments and ' %s=\"%s\"' % (XML_ANCHOR, uid('A',\n anchors[chunk.start])) or ''))\n indent = push(indent)\n if not chunk:\n xml.append(indent + '<%s>' % XML_CHINK)\n indent = push(indent)\n xml.append(indent + '<%s%s%s%s>%s</%s>' % (XML_WORD, word.type and \n ' %s=\"%s\"' % (XML_TYPE, xml_encode(word.type)) or '', word.\n lemma and ' %s=\"%s\"' % (XML_LEMMA, xml_encode(word.lemma)) or\n '', (' ' + ' '.join([('%s=\"%s\"' % (k, v)) for k, v in word.\n custom_tags.items() if v != None])).rstrip(), xml_encode(\n unicode(word)), XML_WORD))\n if not chunk:\n indent = pop(indent)\n xml.append(indent + '</%s>' % XML_CHINK)\n if chunk and chunk.stop - 1 == word.index:\n indent = pop(indent)\n xml.append(indent + '</%s>' % XML_CHUNK)\n if pnp and pnp.stop - 1 == word.index:\n indent = pop(indent)\n xml.append(indent + '</%s>' % XML_CHUNK)\n xml.append('</%s>' % XML_SENTENCE)\n return '\\n'.join(xml).encode('utf-8')\n",
"role": "user"
},
{
"content": "parse_xml",
"role": "assistant"
}
] |
def x(xml):
string = ''
dom = XML(xml)
for sentence in dom(XML_SENTENCE):
_anchors.clear()
_attachments.clear()
language = sentence.get(XML_LANGUAGE, 'en')
format = sentence.get(XML_TOKEN, [WORD, POS, CHUNK, PNP, REL,
ANCHOR, LEMMA])
format = not isinstance(format, basestring
) and format or format.replace(' ', '').split(',')
tokens = []
for chunk in sentence:
tokens.extend(_parse_tokens(chunk, format))
if ANCHOR in format:
A, P, a, i = _anchors, _attachments, 1, format.index(ANCHOR)
for id in sorted(A.keys()):
for token in A[id]:
token[i] += '-' + '-'.join([('A' + str(a + p)) for p in
range(len(P[id]))])
token[i] = token[i].strip('O-')
for p, pnp in enumerate(P[id]):
for token in pnp:
token[i] += '-' + 'P' + str(a + p)
token[i] = token[i].strip('O-')
a += len(P[id])
tokens = ['/'.join([tag for tag in token]) for token in tokens]
tokens = ' '.join(tokens)
string += tokens + '\n'
try:
if MBSP:
from mbsp import TokenString
return TokenString(string.strip(), tags=format, language=language)
except:
return TaggedString(string.strip(), tags=format, language=language)
|
parse_string
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(xml):\n string = ''\n dom = XML(xml)\n for sentence in dom(XML_SENTENCE):\n _anchors.clear()\n _attachments.clear()\n language = sentence.get(XML_LANGUAGE, 'en')\n format = sentence.get(XML_TOKEN, [WORD, POS, CHUNK, PNP, REL,\n ANCHOR, LEMMA])\n format = not isinstance(format, basestring\n ) and format or format.replace(' ', '').split(',')\n tokens = []\n for chunk in sentence:\n tokens.extend(_parse_tokens(chunk, format))\n if ANCHOR in format:\n A, P, a, i = _anchors, _attachments, 1, format.index(ANCHOR)\n for id in sorted(A.keys()):\n for token in A[id]:\n token[i] += '-' + '-'.join([('A' + str(a + p)) for p in\n range(len(P[id]))])\n token[i] = token[i].strip('O-')\n for p, pnp in enumerate(P[id]):\n for token in pnp:\n token[i] += '-' + 'P' + str(a + p)\n token[i] = token[i].strip('O-')\n a += len(P[id])\n tokens = ['/'.join([tag for tag in token]) for token in tokens]\n tokens = ' '.join(tokens)\n string += tokens + '\\n'\n try:\n if MBSP:\n from mbsp import TokenString\n return TokenString(string.strip(), tags=format, language=language)\n except:\n return TaggedString(string.strip(), tags=format, language=language)\n",
"role": "user"
},
{
"content": "parse_string",
"role": "assistant"
}
] |
def x(chunk, format=[WORD, POS, CHUNK, PNP, REL, ANCHOR, LEMMA]):
tokens = []
if not (chunk.tag == XML_CHUNK or chunk.tag == XML_CHINK):
return []
type = chunk.get(XML_TYPE, 'O')
if type == 'PNP':
for ch in chunk:
tokens.extend(_parse_tokens(ch, format))
if PNP in format:
i = format.index(PNP)
for j, token in enumerate(tokens):
token[i] = (j == 0 and 'B-' or 'I-') + 'PNP'
a = chunk.get(XML_OF).split(_UID_SEPARATOR)[-1]
if a:
_attachments.setdefault(a, [])
_attachments[a].append(tokens)
return tokens
relation = _parse_relation(chunk, type)
for word in filter(lambda n: n.tag == XML_WORD, chunk):
tokens.append(_parse_token(word, chunk=type, relation=relation,
format=format))
if CHUNK in format:
i = format.index(CHUNK)
for j, token in enumerate(tokens):
token[i] = token[i] != 'O' and (j == 0 and 'B-' or 'I-') + token[i
] or 'O'
a = chunk.get(XML_ANCHOR, '').split(_UID_SEPARATOR)[-1]
if a:
_anchors[a] = tokens
return tokens
|
parse_tokens
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(chunk, format=[WORD, POS, CHUNK, PNP, REL, ANCHOR, LEMMA]):\n tokens = []\n if not (chunk.tag == XML_CHUNK or chunk.tag == XML_CHINK):\n return []\n type = chunk.get(XML_TYPE, 'O')\n if type == 'PNP':\n for ch in chunk:\n tokens.extend(_parse_tokens(ch, format))\n if PNP in format:\n i = format.index(PNP)\n for j, token in enumerate(tokens):\n token[i] = (j == 0 and 'B-' or 'I-') + 'PNP'\n a = chunk.get(XML_OF).split(_UID_SEPARATOR)[-1]\n if a:\n _attachments.setdefault(a, [])\n _attachments[a].append(tokens)\n return tokens\n relation = _parse_relation(chunk, type)\n for word in filter(lambda n: n.tag == XML_WORD, chunk):\n tokens.append(_parse_token(word, chunk=type, relation=relation,\n format=format))\n if CHUNK in format:\n i = format.index(CHUNK)\n for j, token in enumerate(tokens):\n token[i] = token[i] != 'O' and (j == 0 and 'B-' or 'I-') + token[i\n ] or 'O'\n a = chunk.get(XML_ANCHOR, '').split(_UID_SEPARATOR)[-1]\n if a:\n _anchors[a] = tokens\n return tokens\n",
"role": "user"
},
{
"content": "parse_tokens",
"role": "assistant"
}
] |
def x(chunk, type='O'):
r1 = chunk.get(XML_RELATION)
r2 = chunk.get(XML_ID, chunk.get(XML_OF))
r1 = [(x != '-' and x or None) for x in r1.split('|')] or [None]
r2 = [(x != '-' and x or None) for x in r2.split('|')] or [None]
r2 = [(x is not None and x.split(_UID_SEPARATOR)[-1] or x) for x in r2]
if len(r1) < len(r2):
r1 = r1 + r1 * (len(r2) - len(r1))
if len(r2) < len(r1):
r2 = r2 + r2 * (len(r1) - len(r2))
return ';'.join(['-'.join([x for x in (type, r1, r2) if x]) for r1, r2 in
zip(r1, r2)])
|
parse_relation
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(chunk, type='O'):\n r1 = chunk.get(XML_RELATION)\n r2 = chunk.get(XML_ID, chunk.get(XML_OF))\n r1 = [(x != '-' and x or None) for x in r1.split('|')] or [None]\n r2 = [(x != '-' and x or None) for x in r2.split('|')] or [None]\n r2 = [(x is not None and x.split(_UID_SEPARATOR)[-1] or x) for x in r2]\n if len(r1) < len(r2):\n r1 = r1 + r1 * (len(r2) - len(r1))\n if len(r2) < len(r1):\n r2 = r2 + r2 * (len(r1) - len(r2))\n return ';'.join(['-'.join([x for x in (type, r1, r2) if x]) for r1, r2 in\n zip(r1, r2)])\n",
"role": "user"
},
{
"content": "parse_relation",
"role": "assistant"
}
] |
def x(word, chunk='O', pnp='O', relation='O', anchor='O', format=[WORD, POS,
CHUNK, PNP, REL, ANCHOR, LEMMA]):
tags = []
for tag in format:
if tag == WORD:
tags.append(xml_decode(word.value))
elif tag == POS:
tags.append(xml_decode(word.get(XML_TYPE, 'O')))
elif tag == CHUNK:
tags.append(chunk)
elif tag == PNP:
tags.append(pnp)
elif tag == REL:
tags.append(relation)
elif tag == ANCHOR:
tags.append(anchor)
elif tag == LEMMA:
tags.append(xml_decode(word.get(XML_LEMMA, '')))
else:
tags.append(xml_decode(word.get(tag, 'O')))
return tags
|
parse_token
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(word, chunk='O', pnp='O', relation='O', anchor='O', format=[WORD, POS,\n CHUNK, PNP, REL, ANCHOR, LEMMA]):\n tags = []\n for tag in format:\n if tag == WORD:\n tags.append(xml_decode(word.value))\n elif tag == POS:\n tags.append(xml_decode(word.get(XML_TYPE, 'O')))\n elif tag == CHUNK:\n tags.append(chunk)\n elif tag == PNP:\n tags.append(pnp)\n elif tag == REL:\n tags.append(relation)\n elif tag == ANCHOR:\n tags.append(anchor)\n elif tag == LEMMA:\n tags.append(xml_decode(word.get(XML_LEMMA, '')))\n else:\n tags.append(xml_decode(word.get(tag, 'O')))\n return tags\n",
"role": "user"
},
{
"content": "parse_token",
"role": "assistant"
}
] |
def x(sentence):
from nltk import tree
def x(pnp):
s = ' '.join([do_chunk(ch) for ch in pnp.chunks])
return '(PNP %s)' % s
def x(ch):
s = ' '.join([('(%s %s)' % (w.pos, w.string)) for w in ch.words])
s += ' '.join([do_pnp(pnp) for pnp in ch.attachments])
return '(%s %s)' % (ch.type, s)
T = ['(S']
v = []
for ch in sentence.chunked():
if not ch.pnp and isinstance(ch, Chink):
T.append('(%s %s)' % (ch.words[0].pos, ch.words[0].string))
elif not ch.pnp:
T.append(do_chunk(ch))
elif ch.pnp.anchor is None and ch.pnp not in v:
T.append(do_pnp(ch.pnp))
v.append(ch.pnp)
T.append(')')
return tree.bracket_parse(' '.join(T))
|
nltk_tree
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(sentence):\n from nltk import tree\n\n def x(pnp):\n s = ' '.join([do_chunk(ch) for ch in pnp.chunks])\n return '(PNP %s)' % s\n\n def x(ch):\n s = ' '.join([('(%s %s)' % (w.pos, w.string)) for w in ch.words])\n s += ' '.join([do_pnp(pnp) for pnp in ch.attachments])\n return '(%s %s)' % (ch.type, s)\n T = ['(S']\n v = []\n for ch in sentence.chunked():\n if not ch.pnp and isinstance(ch, Chink):\n T.append('(%s %s)' % (ch.words[0].pos, ch.words[0].string))\n elif not ch.pnp:\n T.append(do_chunk(ch))\n elif ch.pnp.anchor is None and ch.pnp not in v:\n T.append(do_pnp(ch.pnp))\n v.append(ch.pnp)\n T.append(')')\n return tree.bracket_parse(' '.join(T))\n",
"role": "user"
},
{
"content": "nltk_tree",
"role": "assistant"
}
] |
def x(sentence, font='Arial', colors=BLUE):
s = 'digraph sentence {\n'
s += '\tranksep=0.75;\n'
s += '\tnodesep=0.15;\n'
s += (
'\tnode [penwidth=1, fontname="%s", shape=record, margin=0.1, height=0.35];\n'
% font)
s += '\tedge [penwidth=1];\n'
s += '\t{ rank=same;\n'
for w in sentence.words:
s += '\t\tword%s [label="<f0>%s|<f1>%s"%s];\n' % (w.index, w.string,
w.type, _colorize(w, colors))
for w in sentence.words[:-1]:
s += '\t\tword%s -> word%s [color=none];\n' % (w.index, w.index + 1)
s += '\t}\n'
s += '\t{ rank=same;\n'
for i, ch in enumerate(sentence.chunks):
s += '\t\tchunk%s [label="<f0>%s"%s];\n' % (i + 1, '-'.join([x for
x in (ch.type, ch.role, str(ch.relation or '')) if x]) or '-',
_colorize(ch, colors))
for i, ch in enumerate(sentence.chunks[:-1]):
s += '\t\tchunk%s -> chunk%s [color=none];\n' % (i + 1, i + 2)
s += '}\n'
s += '\t{ rank=same;\n'
for i, ch in enumerate(sentence.pnp):
s += '\t\tpnp%s [label="<f0>PNP"%s];\n' % (i + 1, _colorize(ch, colors)
)
s += '\t}\n'
s += '\t{ rank=same;\n S [shape=circle, margin=0.25, penwidth=2]; }\n'
for i, ch in enumerate(sentence.chunks):
for w in ch:
s += '\tword%s -> chunk%s;\n' % (w.index, i + 1)
if ch.pnp:
s += '\tchunk%s -> pnp%s;\n' % (i + 1, sentence.pnp.index(ch.
pnp) + 1)
else:
s += '\tchunk%s -> S;\n' % (i + 1)
if ch.type == 'VP':
for r in ch.related:
s += (
'\tchunk%s -> chunk%s [style=dotted, arrowhead=none];\n' %
(i + 1, sentence.chunks.index(r) + 1))
for i, ch in enumerate(sentence.pnp):
if ch.anchor:
s += '\tpnp%s -> chunk%s;\n' % (i + 1, sentence.chunks.index(ch
.anchor) + 1)
s += '\tpnp%s -> S [color=none];\n' % (i + 1)
else:
s += '\tpnp%s -> S;\n' % (i + 1)
s += '}'
return s
|
graphviz_dot
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(sentence, font='Arial', colors=BLUE):\n s = 'digraph sentence {\\n'\n s += '\\tranksep=0.75;\\n'\n s += '\\tnodesep=0.15;\\n'\n s += (\n '\\tnode [penwidth=1, fontname=\"%s\", shape=record, margin=0.1, height=0.35];\\n'\n % font)\n s += '\\tedge [penwidth=1];\\n'\n s += '\\t{ rank=same;\\n'\n for w in sentence.words:\n s += '\\t\\tword%s [label=\"<f0>%s|<f1>%s\"%s];\\n' % (w.index, w.string,\n w.type, _colorize(w, colors))\n for w in sentence.words[:-1]:\n s += '\\t\\tword%s -> word%s [color=none];\\n' % (w.index, w.index + 1)\n s += '\\t}\\n'\n s += '\\t{ rank=same;\\n'\n for i, ch in enumerate(sentence.chunks):\n s += '\\t\\tchunk%s [label=\"<f0>%s\"%s];\\n' % (i + 1, '-'.join([x for\n x in (ch.type, ch.role, str(ch.relation or '')) if x]) or '-',\n _colorize(ch, colors))\n for i, ch in enumerate(sentence.chunks[:-1]):\n s += '\\t\\tchunk%s -> chunk%s [color=none];\\n' % (i + 1, i + 2)\n s += '}\\n'\n s += '\\t{ rank=same;\\n'\n for i, ch in enumerate(sentence.pnp):\n s += '\\t\\tpnp%s [label=\"<f0>PNP\"%s];\\n' % (i + 1, _colorize(ch, colors)\n )\n s += '\\t}\\n'\n s += '\\t{ rank=same;\\n S [shape=circle, margin=0.25, penwidth=2]; }\\n'\n for i, ch in enumerate(sentence.chunks):\n for w in ch:\n s += '\\tword%s -> chunk%s;\\n' % (w.index, i + 1)\n if ch.pnp:\n s += '\\tchunk%s -> pnp%s;\\n' % (i + 1, sentence.pnp.index(ch.\n pnp) + 1)\n else:\n s += '\\tchunk%s -> S;\\n' % (i + 1)\n if ch.type == 'VP':\n for r in ch.related:\n s += (\n '\\tchunk%s -> chunk%s [style=dotted, arrowhead=none];\\n' %\n (i + 1, sentence.chunks.index(r) + 1))\n for i, ch in enumerate(sentence.pnp):\n if ch.anchor:\n s += '\\tpnp%s -> chunk%s;\\n' % (i + 1, sentence.chunks.index(ch\n .anchor) + 1)\n s += '\\tpnp%s -> S [color=none];\\n' % (i + 1)\n else:\n s += '\\tpnp%s -> S;\\n' % (i + 1)\n s += '}'\n return s\n",
"role": "user"
},
{
"content": "graphviz_dot",
"role": "assistant"
}
] |
def x(sentence, fill=1, placeholder='-'):
tags = [WORD, POS, IOB, CHUNK, ROLE, REL, PNP, ANCHOR, LEMMA]
tags += [tag for tag in sentence.token if tag not in tags]
def x(token, tag):
if tag == WORD:
s = token.string
elif tag == POS:
s = token.type
elif tag == IOB:
s = token.chunk and (token.index == token.chunk.start and 'B' or
'I')
elif tag == CHUNK:
s = token.chunk and token.chunk.type
elif tag == ROLE:
s = token.chunk and token.chunk.role
elif tag == REL:
s = token.chunk and token.chunk.relation and str(token.chunk.
relation)
elif tag == PNP:
s = token.chunk and token.chunk.pnp and token.chunk.pnp.type
elif tag == ANCHOR:
s = token.chunk and token.chunk.anchor_id
elif tag == LEMMA:
s = token.lemma
else:
s = token.custom_tags.get(tag)
return s or placeholder
def x(column, fill=1, padding=3, align='left'):
n = max([len(x) for x in column] + [fill])
if align == 'left':
return [(x + ' ' * (n - len(x)) + ' ' * padding) for x in column]
if align == 'right':
return [(' ' * (n - len(x)) + x + ' ' * padding) for x in column]
columns = [[format(token, tag) for token in sentence] for tag in tags]
columns[3] = [(columns[3][i] + (iob == 'I' and ' ^' or '')) for i, iob in
enumerate(columns[2])]
del columns[2]
for i, header in enumerate(['word', 'tag', 'chunk', 'role', 'id', 'pnp',
'anchor', 'lemma'] + tags[9:]):
columns[i].insert(0, '')
columns[i].insert(0, header.upper())
for i, column in enumerate(columns):
columns[i] = outline(column, fill + 10 * (i == 0), align=('left',
'right')[i == 0])
if not MBSP:
del columns[6]
return '\n'.join([''.join([x[i] for x in columns]) for i in range(len(
columns[0]))])
|
table
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(sentence, fill=1, placeholder='-'):\n tags = [WORD, POS, IOB, CHUNK, ROLE, REL, PNP, ANCHOR, LEMMA]\n tags += [tag for tag in sentence.token if tag not in tags]\n\n def x(token, tag):\n if tag == WORD:\n s = token.string\n elif tag == POS:\n s = token.type\n elif tag == IOB:\n s = token.chunk and (token.index == token.chunk.start and 'B' or\n 'I')\n elif tag == CHUNK:\n s = token.chunk and token.chunk.type\n elif tag == ROLE:\n s = token.chunk and token.chunk.role\n elif tag == REL:\n s = token.chunk and token.chunk.relation and str(token.chunk.\n relation)\n elif tag == PNP:\n s = token.chunk and token.chunk.pnp and token.chunk.pnp.type\n elif tag == ANCHOR:\n s = token.chunk and token.chunk.anchor_id\n elif tag == LEMMA:\n s = token.lemma\n else:\n s = token.custom_tags.get(tag)\n return s or placeholder\n\n def x(column, fill=1, padding=3, align='left'):\n n = max([len(x) for x in column] + [fill])\n if align == 'left':\n return [(x + ' ' * (n - len(x)) + ' ' * padding) for x in column]\n if align == 'right':\n return [(' ' * (n - len(x)) + x + ' ' * padding) for x in column]\n columns = [[format(token, tag) for token in sentence] for tag in tags]\n columns[3] = [(columns[3][i] + (iob == 'I' and ' ^' or '')) for i, iob in\n enumerate(columns[2])]\n del columns[2]\n for i, header in enumerate(['word', 'tag', 'chunk', 'role', 'id', 'pnp',\n 'anchor', 'lemma'] + tags[9:]):\n columns[i].insert(0, '')\n columns[i].insert(0, header.upper())\n for i, column in enumerate(columns):\n columns[i] = outline(column, fill + 10 * (i == 0), align=('left',\n 'right')[i == 0])\n if not MBSP:\n del columns[6]\n return '\\n'.join([''.join([x[i] for x in columns]) for i in range(len(\n columns[0]))])\n",
"role": "user"
},
{
"content": "table",
"role": "assistant"
}
] |
def x(self):
ch, I, O, B = self.chunk, INSIDE + '-', OUTSIDE, BEGIN + '-'
tags = [OUTSIDE for i in range(len(self.sentence.token))]
for i, tag in enumerate(self.sentence.token):
if tag == WORD:
tags[i] = encode_entities(self.string)
elif tag == POS and self.type:
tags[i] = self.type
elif tag == CHUNK and ch and ch.type:
tags[i] = (self == ch[0] and B or I) + ch.type
elif tag == PNP and self.pnp:
tags[i] = (self == self.pnp[0] and B or I) + 'PNP'
elif tag == REL and ch and len(ch.relations) > 0:
tags[i] = ['-'.join([str(x) for x in [ch.type] + list(reversed(
r)) if x]) for r in ch.relations]
tags[i] = '*'.join(tags[i])
elif tag == ANCHOR and ch:
tags[i] = ch.anchor_id or OUTSIDE
elif tag == LEMMA:
tags[i] = encode_entities(self.lemma or '')
elif tag in self.custom_tags:
tags[i] = self.custom_tags.get(tag) or OUTSIDE
return tags
|
tags
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n ch, I, O, B = self.chunk, INSIDE + '-', OUTSIDE, BEGIN + '-'\n tags = [OUTSIDE for i in range(len(self.sentence.token))]\n for i, tag in enumerate(self.sentence.token):\n if tag == WORD:\n tags[i] = encode_entities(self.string)\n elif tag == POS and self.type:\n tags[i] = self.type\n elif tag == CHUNK and ch and ch.type:\n tags[i] = (self == ch[0] and B or I) + ch.type\n elif tag == PNP and self.pnp:\n tags[i] = (self == self.pnp[0] and B or I) + 'PNP'\n elif tag == REL and ch and len(ch.relations) > 0:\n tags[i] = ['-'.join([str(x) for x in [ch.type] + list(reversed(\n r)) if x]) for r in ch.relations]\n tags[i] = '*'.join(tags[i])\n elif tag == ANCHOR and ch:\n tags[i] = ch.anchor_id or OUTSIDE\n elif tag == LEMMA:\n tags[i] = encode_entities(self.lemma or '')\n elif tag in self.custom_tags:\n tags[i] = self.custom_tags.get(tag) or OUTSIDE\n return tags\n",
"role": "user"
},
{
"content": "tags",
"role": "assistant"
}
] |
def x(self, type=None):
i = self.index + 1
s = self.sentence
while i < len(s):
if type in (s[i].type, None):
return s[i]
i += 1
|
next
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, type=None):\n i = self.index + 1\n s = self.sentence\n while i < len(s):\n if type in (s[i].type, None):\n return s[i]\n i += 1\n",
"role": "user"
},
{
"content": "next",
"role": "assistant"
}
] |
def x(self, type=None):
i = self.index - 1
s = self.sentence
while i > 0:
if type in (s[i].type, None):
return s[i]
i -= 1
|
previous
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, type=None):\n i = self.index - 1\n s = self.sentence\n while i > 0:\n if type in (s[i].type, None):\n return s[i]\n i -= 1\n",
"role": "user"
},
{
"content": "previous",
"role": "assistant"
}
] |
def x(self):
if self.type == 'NP' and any(w.type.startswith('NNP') for w in self):
w = find(lambda w: w.type.startswith('NNP'), reversed(self))
elif self.type == 'NP':
w = find(lambda w: w.type.startswith('NN'), reversed(self))
elif self.type == 'VP':
w = find(lambda w: w.type.startswith('VB'), reversed(self))
elif self.type == 'PP':
w = find(lambda w: w.type.startswith(('IN', 'PP')), self)
elif self.type == 'PNP':
w = find(lambda w: w.type.startswith('NN'), reversed(self))
else:
w = None
if w is None:
w = self[-1]
return w
|
head
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self.type == 'NP' and any(w.type.startswith('NNP') for w in self):\n w = find(lambda w: w.type.startswith('NNP'), reversed(self))\n elif self.type == 'NP':\n w = find(lambda w: w.type.startswith('NN'), reversed(self))\n elif self.type == 'VP':\n w = find(lambda w: w.type.startswith('VB'), reversed(self))\n elif self.type == 'PP':\n w = find(lambda w: w.type.startswith(('IN', 'PP')), self)\n elif self.type == 'PNP':\n w = find(lambda w: w.type.startswith('NN'), reversed(self))\n else:\n w = None\n if w is None:\n w = self[-1]\n return w\n",
"role": "user"
},
{
"content": "head",
"role": "assistant"
}
] |
def x(self):
return [ch for ch in self.sentence.chunks if ch != self and intersects(
unzip(0, ch.relations), unzip(0, self.relations))]
|
related
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n return [ch for ch in self.sentence.chunks if ch != self and intersects(\n unzip(0, ch.relations), unzip(0, self.relations))]\n",
"role": "user"
},
{
"content": "related",
"role": "assistant"
}
] |
def x(self):
id = ''
f = lambda ch: filter(lambda k: self.sentence._anchors[k] == ch, self.
sentence._anchors)
if self.pnp and self.pnp.anchor:
id += '-' + '-'.join(f(self.pnp))
if self.anchor:
id += '-' + '-'.join(f(self))
if self.attachments:
id += '-' + '-'.join(f(self))
return id.strip('-') or None
|
anchor_id
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n id = ''\n f = lambda ch: filter(lambda k: self.sentence._anchors[k] == ch, self.\n sentence._anchors)\n if self.pnp and self.pnp.anchor:\n id += '-' + '-'.join(f(self.pnp))\n if self.anchor:\n id += '-' + '-'.join(f(self))\n if self.attachments:\n id += '-' + '-'.join(f(self))\n return id.strip('-') or None\n",
"role": "user"
},
{
"content": "anchor_id",
"role": "assistant"
}
] |
def x(self):
if self._modifiers is None:
is_modifier = lambda ch: ch.type in ('ADJP', 'ADVP'
) and ch.relation is None
for chunk in self.sentence.chunks:
chunk._modifiers = []
for chunk in filter(is_modifier, self.sentence.chunks):
anchor = chunk.nearest('VP')
if anchor:
anchor._modifiers.append(chunk)
return self._modifiers
|
modifiers
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self):\n if self._modifiers is None:\n is_modifier = lambda ch: ch.type in ('ADJP', 'ADVP'\n ) and ch.relation is None\n for chunk in self.sentence.chunks:\n chunk._modifiers = []\n for chunk in filter(is_modifier, self.sentence.chunks):\n anchor = chunk.nearest('VP')\n if anchor:\n anchor._modifiers.append(chunk)\n return self._modifiers\n",
"role": "user"
},
{
"content": "modifiers",
"role": "assistant"
}
] |
def x(self, type='VP'):
candidate, d = None, len(self.sentence.chunks)
if isinstance(self, PNPChunk):
i = self.sentence.chunks.index(self.chunks[0])
else:
i = self.sentence.chunks.index(self)
for j, chunk in enumerate(self.sentence.chunks):
if chunk.type.startswith(type) and abs(i - j) < d:
candidate, d = chunk, abs(i - j)
return candidate
|
nearest
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, type='VP'):\n candidate, d = None, len(self.sentence.chunks)\n if isinstance(self, PNPChunk):\n i = self.sentence.chunks.index(self.chunks[0])\n else:\n i = self.sentence.chunks.index(self)\n for j, chunk in enumerate(self.sentence.chunks):\n if chunk.type.startswith(type) and abs(i - j) < d:\n candidate, d = chunk, abs(i - j)\n return candidate\n",
"role": "user"
},
{
"content": "nearest",
"role": "assistant"
}
] |
def x(self, type=None):
i = self.stop
s = self.sentence
while i < len(s):
if s[i].chunk is not None and type in (s[i].chunk.type, None):
return s[i].chunk
i += 1
|
next
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, type=None):\n i = self.stop\n s = self.sentence\n while i < len(s):\n if s[i].chunk is not None and type in (s[i].chunk.type, None):\n return s[i].chunk\n i += 1\n",
"role": "user"
},
{
"content": "next",
"role": "assistant"
}
] |
def x(self, type=None):
i = self.start - 1
s = self.sentence
while i > 0:
if s[i].chunk is not None and type in (s[i].chunk.type, None):
return s[i].chunk
i -= 1
|
previous
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, type=None):\n i = self.start - 1\n s = self.sentence\n while i > 0:\n if s[i].chunk is not None and type in (s[i].chunk.type, None):\n return s[i].chunk\n i -= 1\n",
"role": "user"
},
{
"content": "previous",
"role": "assistant"
}
] |
def x(self, word, lemma=None, type=None, chunk=None, role=None, relation=
None, pnp=None, anchor=None, iob=None, custom={}):
self._do_word(word, lemma, type)
self._do_chunk(chunk, role, relation, iob)
self._do_conjunction()
self._do_relation()
self._do_pnp(pnp, anchor)
self._do_anchor(anchor)
self._do_custom(custom)
|
append
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, word, lemma=None, type=None, chunk=None, role=None, relation=\n None, pnp=None, anchor=None, iob=None, custom={}):\n self._do_word(word, lemma, type)\n self._do_chunk(chunk, role, relation, iob)\n self._do_conjunction()\n self._do_relation()\n self._do_pnp(pnp, anchor)\n self._do_anchor(anchor)\n self._do_custom(custom)\n",
"role": "user"
},
{
"content": "append",
"role": "assistant"
}
] |
def x(self, token, tags=[WORD, POS, CHUNK, PNP, REL, ANCHOR, LEMMA]):
p = {WORD: '', POS: None, IOB: None, CHUNK: None, PNP: None, REL: None,
ROLE: None, ANCHOR: None, LEMMA: None}
custom = {}
for k, v in izip(tags, token.split('/')):
if SLASH0 in v:
v = v.replace(SLASH, '/')
if k not in p:
custom[k] = None
if v != OUTSIDE or k == WORD or k == LEMMA:
(p if k not in custom else custom)[k] = v
ch = p[CHUNK]
if ch is not None and ch.startswith(('B-', 'I-')):
p[IOB], p[CHUNK] = ch[:1], ch[2:]
if p[REL] is not None:
ch, p[REL], p[ROLE] = self._parse_relation(p[REL])
if ch == 'PP' and self._previous and self._previous[REL] == p[REL
] and self._previous[ROLE] == p[ROLE]:
ch = 'NP'
if p[CHUNK] is None and ch != OUTSIDE:
p[CHUNK] = ch
self._previous = p
return p[WORD], p[LEMMA], p[POS], p[CHUNK], p[ROLE], p[REL], p[PNP], p[
ANCHOR], p[IOB], custom
|
parse_token
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, token, tags=[WORD, POS, CHUNK, PNP, REL, ANCHOR, LEMMA]):\n p = {WORD: '', POS: None, IOB: None, CHUNK: None, PNP: None, REL: None,\n ROLE: None, ANCHOR: None, LEMMA: None}\n custom = {}\n for k, v in izip(tags, token.split('/')):\n if SLASH0 in v:\n v = v.replace(SLASH, '/')\n if k not in p:\n custom[k] = None\n if v != OUTSIDE or k == WORD or k == LEMMA:\n (p if k not in custom else custom)[k] = v\n ch = p[CHUNK]\n if ch is not None and ch.startswith(('B-', 'I-')):\n p[IOB], p[CHUNK] = ch[:1], ch[2:]\n if p[REL] is not None:\n ch, p[REL], p[ROLE] = self._parse_relation(p[REL])\n if ch == 'PP' and self._previous and self._previous[REL] == p[REL\n ] and self._previous[ROLE] == p[ROLE]:\n ch = 'NP'\n if p[CHUNK] is None and ch != OUTSIDE:\n p[CHUNK] = ch\n self._previous = p\n return p[WORD], p[LEMMA], p[POS], p[CHUNK], p[ROLE], p[REL], p[PNP], p[\n ANCHOR], p[IOB], custom\n",
"role": "user"
},
{
"content": "parse_token",
"role": "assistant"
}
] |
def x(self, tag):
chunk, relation, role = None, [], []
if ';' in tag:
id = tag.split('*')[0][-2:]
id = id if id.startswith('-') else ''
tag = tag.replace(';', id + '*')
if '*' in tag:
tag = tag.split('*')
else:
tag = [tag]
for s in tag:
s = s.split('-')
n = len(s)
if n == 1:
chunk = s[0]
if n == 2:
chunk = s[0]
relation.append(s[1])
role.append(None)
if n >= 3:
chunk = s[0]
relation.append(s[2])
role.append(s[1])
if n > 1:
id = relation[-1]
if id.isdigit():
relation[-1] = int(id)
else:
relation[-1], role[-1] = None, id
return chunk, relation, role
|
parse_relation
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, tag):\n chunk, relation, role = None, [], []\n if ';' in tag:\n id = tag.split('*')[0][-2:]\n id = id if id.startswith('-') else ''\n tag = tag.replace(';', id + '*')\n if '*' in tag:\n tag = tag.split('*')\n else:\n tag = [tag]\n for s in tag:\n s = s.split('-')\n n = len(s)\n if n == 1:\n chunk = s[0]\n if n == 2:\n chunk = s[0]\n relation.append(s[1])\n role.append(None)\n if n >= 3:\n chunk = s[0]\n relation.append(s[2])\n role.append(s[1])\n if n > 1:\n id = relation[-1]\n if id.isdigit():\n relation[-1] = int(id)\n else:\n relation[-1], role[-1] = None, id\n return chunk, relation, role\n",
"role": "user"
},
{
"content": "parse_relation",
"role": "assistant"
}
] |
def x(self, word, lemma=None, type=None):
if lemma == "'s" and type in ('VB', 'VBZ'):
lemma = 'be'
self.words.append(Word(self, word, lemma, type, index=len(self.words)))
|
do_word
|
[
{
"content": "Given the source code of a python function, suggest a fitting name for the function.",
"role": "system"
},
{
"content": "def x(self, word, lemma=None, type=None):\n if lemma == \"'s\" and type in ('VB', 'VBZ'):\n lemma = 'be'\n self.words.append(Word(self, word, lemma, type, index=len(self.words)))\n",
"role": "user"
},
{
"content": "do_word",
"role": "assistant"
}
] |
Subsets and Splits
No community queries yet
The top public SQL queries from the community will appear here once available.