Search is not available for this dataset
repo stringlengths 2 152 ⌀ | file stringlengths 15 239 | code stringlengths 0 58.4M | file_length int64 0 58.4M | avg_line_length float64 0 1.81M | max_line_length int64 0 12.7M | extension_type stringclasses 364
values |
|---|---|---|---|---|---|---|
spaCy | spaCy-master/spacy/pipeline/spancat.py | from dataclasses import dataclass
from functools import partial
from typing import Any, Callable, Dict, Iterable, List, Optional, Tuple, Union, cast
import numpy
from thinc.api import Config, Model, Ops, Optimizer, get_current_ops, set_dropout_rate
from thinc.types import Floats2d, Ints1d, Ints2d, Ragged
from ..compa... | 30,032 | 37.112944 | 101 | py |
spaCy | spaCy-master/spacy/pipeline/textcat.py | from itertools import islice
from typing import Any, Callable, Dict, Iterable, List, Optional, Tuple
import numpy
from thinc.api import Config, Model, Optimizer, get_array_module, set_dropout_rate
from thinc.types import Floats2d
from ..errors import Errors
from ..language import Language
from ..scorer import Scorer
... | 14,943 | 35.537897 | 104 | py |
spaCy | spaCy-master/spacy/pipeline/textcat_multilabel.py | from itertools import islice
from typing import Any, Callable, Dict, Iterable, List, Optional
from thinc.api import Config, Model
from thinc.types import Floats2d
from ..errors import Errors
from ..language import Language
from ..scorer import Scorer
from ..tokens import Doc
from ..training import Example, validate_g... | 6,605 | 30.457143 | 94 | py |
spaCy | spaCy-master/spacy/pipeline/tok2vec.py | from itertools import islice
from typing import Any, Callable, Dict, Iterable, List, Optional, Sequence
from thinc.api import Config, Model, Optimizer, set_dropout_rate
from ..errors import Errors
from ..language import Language
from ..tokens import Doc
from ..training import Example, validate_examples, validate_get_... | 13,390 | 40.458204 | 104 | py |
spaCy | spaCy-master/spacy/pipeline/_edit_tree_internals/__init__.py | 0 | 0 | 0 | py | |
spaCy | spaCy-master/spacy/pipeline/_edit_tree_internals/schemas.py | from collections import defaultdict
from typing import Any, Dict, List, Union
from pydantic import BaseModel, Field, ValidationError
from pydantic.types import StrictBool, StrictInt, StrictStr
class MatchNodeSchema(BaseModel):
prefix_len: StrictInt = Field(..., title="Prefix length")
suffix_len: StrictInt = ... | 1,475 | 31.086957 | 99 | py |
spaCy | spaCy-master/spacy/pipeline/_parser_internals/__init__.py | 0 | 0 | 0 | py | |
spaCy | spaCy-master/spacy/pipeline/legacy/__init__.py | from .entity_linker import EntityLinker_v1
__all__ = ["EntityLinker_v1"]
| 74 | 17.75 | 42 | py |
spaCy | spaCy-master/spacy/pipeline/legacy/entity_linker.py | # This file is present to provide a prior version of the EntityLinker component
# for backwards compatability. For details see #9669.
import random
import warnings
from itertools import islice
from pathlib import Path
from typing import Any, Callable, Dict, Iterable, List, Optional, Union
import srsly
from thinc.api ... | 18,788 | 43.41844 | 120 | py |
spaCy | spaCy-master/spacy/tests/README.md | <a href="https://explosion.ai"><img src="https://explosion.ai/assets/img/logo.svg" width="125" height="125" align="right" /></a>
# spaCy tests
spaCy uses the [pytest](http://doc.pytest.org/) framework for testing. For more info on this, see the [pytest documentation](http://docs.pytest.org/en/latest/contents.html).
... | 8,728 | 62.253623 | 421 | md |
spaCy | spaCy-master/spacy/tests/__init__.py | 0 | 0 | 0 | py | |
spaCy | spaCy-master/spacy/tests/conftest.py | import pytest
from hypothesis import settings
from spacy.util import get_lang_class
# Functionally disable deadline settings for tests
# to prevent spurious test failures in CI builds.
settings.register_profile("no_deadlines", deadline=2 * 60 * 1000) # in ms
settings.load_profile("no_deadlines")
def pytest_addopti... | 11,609 | 21.764706 | 83 | py |
spaCy | spaCy-master/spacy/tests/enable_gpu.py | from spacy import require_gpu
require_gpu()
| 45 | 10.5 | 29 | py |
spaCy | spaCy-master/spacy/tests/test_architectures.py | import pytest
from catalogue import RegistryError
from thinc.api import Linear
from spacy import registry
def test_get_architecture():
@registry.architectures("my_test_function")
def create_model(nr_in, nr_out):
return Linear(nr_in, nr_out)
arch = registry.architectures.get("my_test_function")
... | 448 | 25.411765 | 57 | py |
spaCy | spaCy-master/spacy/tests/test_cli.py | import math
import os
import time
from collections import Counter
from pathlib import Path
from typing import Any, Dict, List, Tuple
import numpy
import pytest
import srsly
from click import NoSuchOption
from packaging.specifiers import SpecifierSet
from thinc.api import Config, ConfigValidationError
import spacy
fro... | 49,789 | 35.799704 | 229 | py |
spaCy | spaCy-master/spacy/tests/test_cli_app.py | import os
from pathlib import Path
import pytest
import srsly
from typer.testing import CliRunner
from spacy.cli._util import app, get_git_version
from spacy.tokens import Doc, DocBin
from .util import make_tempdir, normalize_whitespace
def has_git():
try:
get_git_version()
return True
exce... | 7,767 | 31.915254 | 111 | py |
spaCy | spaCy-master/spacy/tests/test_displacy.py | import numpy
import pytest
from spacy import displacy
from spacy.displacy.render import DependencyRenderer, EntityRenderer
from spacy.lang.en import English
from spacy.lang.fa import Persian
from spacy.tokens import Doc, Span
@pytest.mark.issue(2361)
def test_issue2361(de_vocab):
"""Test if < is escaped when ren... | 14,268 | 34.761905 | 92 | py |
spaCy | spaCy-master/spacy/tests/test_errors.py | from inspect import isclass
import pytest
from spacy.errors import ErrorsWithCodes
class Errors(metaclass=ErrorsWithCodes):
E001 = "error description"
def test_add_codes():
assert Errors.E001 == "[E001] error description"
with pytest.raises(AttributeError):
Errors.E002
assert isclass(Error... | 333 | 18.647059 | 52 | py |
spaCy | spaCy-master/spacy/tests/test_language.py | import itertools
import logging
from unittest import mock
import pytest
from thinc.api import CupyOps, NumpyOps, get_current_ops
import spacy
from spacy.lang.de import German
from spacy.lang.en import English
from spacy.language import Language
from spacy.scorer import Scorer
from spacy.tokens import Doc, Span
from s... | 26,993 | 32.616438 | 92 | py |
spaCy | spaCy-master/spacy/tests/test_misc.py | import ctypes
import os
from pathlib import Path
import pytest
from pydantic import ValidationError
from thinc.api import (
Config,
ConfigValidationError,
CupyOps,
MPSOps,
NumpyOps,
Optimizer,
get_current_ops,
set_current_ops,
)
from thinc.compat import has_cupy_gpu, has_torch_mps_gpu
... | 15,386 | 30.27439 | 88 | py |
spaCy | spaCy-master/spacy/tests/test_models.py | from typing import List
import numpy
import pytest
from numpy.testing import assert_array_almost_equal, assert_array_equal
from thinc.api import (
Adam,
Logistic,
Ragged,
Relu,
chain,
fix_random_seed,
reduce_mean,
set_dropout_rate,
)
from spacy.lang.en import English
from spacy.lang.en... | 9,082 | 30.648084 | 88 | py |
spaCy | spaCy-master/spacy/tests/test_pickles.py | import numpy
import pytest
import srsly
from spacy.attrs import NORM
from spacy.lang.en import English
from spacy.strings import StringStore
from spacy.tokens import Doc
from spacy.vocab import Vocab
@pytest.mark.parametrize("text1,text2", [("hello", "bye")])
def test_pickle_string_store(text1, text2):
stringsto... | 2,023 | 31.126984 | 77 | py |
spaCy | spaCy-master/spacy/tests/test_scorer.py | import pytest
from numpy.testing import assert_almost_equal, assert_array_almost_equal
from pytest import approx
from spacy.lang.en import English
from spacy.scorer import PRFScore, ROCAUCScore, Scorer, _roc_auc_score, _roc_curve
from spacy.tokens import Doc, Span
from spacy.training import Example
from spacy.training... | 17,478 | 31.488848 | 87 | py |
spaCy | spaCy-master/spacy/tests/test_ty.py | import spacy
from spacy import ty
def test_component_types():
nlp = spacy.blank("en")
tok2vec = nlp.create_pipe("tok2vec")
tagger = nlp.create_pipe("tagger")
entity_ruler = nlp.create_pipe("entity_ruler")
assert isinstance(tok2vec, ty.TrainableComponent)
assert isinstance(tagger, ty.TrainableC... | 748 | 38.421053 | 63 | py |
spaCy | spaCy-master/spacy/tests/util.py | import contextlib
import re
import tempfile
import numpy
import srsly
from thinc.api import get_current_ops
from spacy.tokens import Doc
from spacy.training import split_bilu_label
from spacy.util import make_tempdir # noqa: F401
from spacy.vocab import Vocab
@contextlib.contextmanager
def make_tempfile(mode="r"):... | 3,281 | 30.257143 | 79 | py |
spaCy | spaCy-master/spacy/tests/doc/__init__.py | 0 | 0 | 0 | py | |
spaCy | spaCy-master/spacy/tests/doc/test_add_entities.py | import pytest
from spacy import registry
from spacy.pipeline import EntityRecognizer
from spacy.pipeline.ner import DEFAULT_NER_MODEL
from spacy.tokens import Doc, Span
from spacy.training import Example
def _ner_example(ner):
doc = Doc(
ner.vocab,
words=["Joe", "loves", "visiting", "London", "du... | 1,804 | 30.12069 | 81 | py |
spaCy | spaCy-master/spacy/tests/doc/test_array.py | import numpy
import pytest
from spacy.attrs import DEP, MORPH, ORTH, POS, SHAPE
from spacy.tokens import Doc
@pytest.mark.issue(2203)
def test_issue2203(en_vocab):
"""Test that lemmas are set correctly in doc.from_array."""
words = ["I", "'ll", "survive"]
tags = ["PRP", "MD", "VB"]
lemmas = ["-PRON-"... | 4,988 | 35.416058 | 81 | py |
spaCy | spaCy-master/spacy/tests/doc/test_creation.py | import pytest
from spacy import util
from spacy.tokens import Doc
from spacy.vocab import Vocab
@pytest.fixture
def vocab():
return Vocab()
def test_empty_doc(vocab):
doc = Doc(vocab)
assert len(doc) == 0
def test_single_word(vocab):
doc = Doc(vocab, words=["a"])
assert doc.text == "a "
d... | 2,651 | 31.740741 | 80 | py |
spaCy | spaCy-master/spacy/tests/doc/test_doc_api.py | import warnings
import weakref
import numpy
import pytest
from numpy.testing import assert_array_equal
from thinc.api import NumpyOps, get_current_ops
from spacy.attrs import (
DEP,
ENT_IOB,
ENT_TYPE,
HEAD,
IS_ALPHA,
MORPH,
POS,
SENT_START,
TAG,
)
from spacy.lang.en import English
... | 35,906 | 34.835329 | 192 | py |
spaCy | spaCy-master/spacy/tests/doc/test_graph.py | from spacy.tokens.doc import Doc
from spacy.tokens.graph import Graph
from spacy.vocab import Vocab
def test_graph_init():
doc = Doc(Vocab(), words=["a", "b", "c", "d"])
graph = Graph(doc, name="hello")
assert graph.name == "hello"
assert graph.doc is doc
def test_graph_edges_and_nodes():
doc = ... | 1,819 | 36.142857 | 77 | py |
spaCy | spaCy-master/spacy/tests/doc/test_json_doc_conversion.py | import pytest
import srsly
import spacy
from spacy import schemas
from spacy.tokens import Doc, Span, Token
from .test_underscore import clean_underscore # noqa: F401
@pytest.fixture()
def doc(en_vocab):
words = ["c", "d", "e"]
spaces = [True, True, True]
pos = ["VERB", "NOUN", "NOUN"]
tags = ["VBP... | 13,660 | 34.575521 | 108 | py |
spaCy | spaCy-master/spacy/tests/doc/test_morphanalysis.py | import pytest
@pytest.fixture
def i_has(en_tokenizer):
doc = en_tokenizer("I has")
doc[0].set_morph({"PronType": "prs"})
doc[1].set_morph(
{
"VerbForm": "fin",
"Tense": "pres",
"Number": "sing",
"Person": "three",
}
)
return doc
de... | 3,145 | 30.148515 | 87 | py |
spaCy | spaCy-master/spacy/tests/doc/test_pickle_doc.py | from spacy.compat import pickle
from spacy.language import Language
def test_pickle_single_doc():
nlp = Language()
doc = nlp("pickle roundtrip")
data = pickle.dumps(doc, 1)
doc2 = pickle.loads(data)
assert doc2.text == "pickle roundtrip"
def test_list_of_docs_pickles_efficiently():
nlp = Lan... | 1,470 | 25.745455 | 53 | py |
spaCy | spaCy-master/spacy/tests/doc/test_retokenize_merge.py | import pytest
from spacy.attrs import LEMMA
from spacy.tokens import Doc, Token
from spacy.vocab import Vocab
def test_doc_retokenize_merge(en_tokenizer):
text = "WKRO played songs by the beach boys all night"
attrs = {
"tag": "NAMED",
"lemma": "LEMMA",
"ent_type": "TYPE",
"mo... | 18,956 | 37.219758 | 122 | py |
spaCy | spaCy-master/spacy/tests/doc/test_retokenize_split.py | import numpy
import pytest
from spacy.tokens import Doc, Token
from spacy.vocab import Vocab
@pytest.mark.issue(3540)
def test_issue3540(en_vocab):
words = ["I", "live", "in", "NewYork", "right", "now"]
tensor = numpy.asarray(
[[1.0, 1.1], [2.0, 2.1], [3.0, 3.1], [4.0, 4.1], [5.0, 5.1], [6.0, 6.1]],
... | 10,937 | 35.828283 | 92 | py |
spaCy | spaCy-master/spacy/tests/doc/test_span.py | import numpy
import pytest
from numpy.testing import assert_array_equal
from thinc.api import get_current_ops
from spacy.attrs import LENGTH, ORTH
from spacy.lang.en import English
from spacy.tokens import Doc, Span, Token
from spacy.util import filter_spans
from spacy.vocab import Vocab
from ..util import add_vecs_t... | 25,422 | 33.63624 | 137 | py |
spaCy | spaCy-master/spacy/tests/doc/test_span_group.py | from random import Random
from typing import List
import pytest
from spacy.matcher import Matcher
from spacy.tokens import Doc, Span, SpanGroup
from spacy.util import filter_spans
@pytest.fixture
def doc(en_tokenizer):
doc = en_tokenizer("0 1 2 3 4 5 6")
matcher = Matcher(en_tokenizer.vocab, validate=True)
... | 8,667 | 29.846975 | 98 | py |
spaCy | spaCy-master/spacy/tests/doc/test_token_api.py | import numpy
import pytest
from spacy.attrs import IS_ALPHA, IS_DIGIT, IS_LOWER, IS_PUNCT, IS_STOP, IS_TITLE
from spacy.symbols import VERB
from spacy.tokens import Doc
from spacy.training import Example
from spacy.vocab import Vocab
@pytest.fixture
def doc(en_vocab):
# fmt: off
words = ["This", "is", "a", "... | 11,165 | 36.59596 | 116 | py |
spaCy | spaCy-master/spacy/tests/doc/test_underscore.py | import pytest
from mock import Mock
from spacy.tokens import Doc, Span, Token
from spacy.tokens.underscore import Underscore
@pytest.fixture(scope="function", autouse=True)
def clean_underscore():
# reset the Underscore object after the test, to avoid having state copied across tests
yield
Underscore.doc... | 5,565 | 30.805714 | 91 | py |
spaCy | spaCy-master/spacy/tests/lang/__init__.py | 0 | 0 | 0 | py | |
spaCy | spaCy-master/spacy/tests/lang/test_attrs.py | import pytest
from spacy.attrs import ENT_IOB, IS_ALPHA, LEMMA, NORM, ORTH, intify_attrs
from spacy.lang.en.stop_words import STOP_WORDS
from spacy.lang.lex_attrs import (
is_ascii,
is_currency,
is_punct,
is_stop,
like_url,
word_shape,
)
@pytest.mark.parametrize("word", ["the"])
@pytest.mark.... | 3,522 | 24.904412 | 87 | py |
spaCy | spaCy-master/spacy/tests/lang/test_initialize.py | import pytest
from spacy.util import get_lang_class
# fmt: off
# Only include languages with no external dependencies
# excluded: ja, ko, th, vi, zh
LANGUAGES = ["af", "am", "ar", "az", "bg", "bn", "ca", "cs", "da", "de", "el",
"en", "es", "et", "eu", "fa", "fi", "fr", "ga", "gu", "he", "hi",
... | 922 | 35.92 | 78 | py |
spaCy | spaCy-master/spacy/tests/lang/test_lemmatizers.py | import pytest
from spacy import registry
from spacy.lookups import Lookups
from spacy.util import get_lang_class
# fmt: off
# Only include languages with no external dependencies
# excluded: ru, uk
# excluded for custom tables: es, pl
LANGUAGES = ["bn", "ca", "el", "en", "fa", "fr", "nb", "nl", "sv"]
# fmt: on
@pyt... | 1,910 | 33.745455 | 71 | py |
spaCy | spaCy-master/spacy/tests/lang/af/__init__.py | 0 | 0 | 0 | py | |
spaCy | spaCy-master/spacy/tests/lang/af/test_text.py | import pytest
def test_long_text(af_tokenizer):
# Excerpt: Universal Declaration of Human Rights; “'n” changed to “die” in first sentence
text = """
Hierdie Universele Verklaring van Menseregte as die algemene standaard vir die verwesenliking deur alle mense en nasies,
om te verseker dat elke individu en elk... | 923 | 39.173913 | 122 | py |
spaCy | spaCy-master/spacy/tests/lang/af/test_tokenizer.py | import pytest
AF_BASIC_TOKENIZATION_TESTS = [
(
"Elkeen het die reg tot lewe, vryheid en sekuriteit van persoon.",
[
"Elkeen",
"het",
"die",
"reg",
"tot",
"lewe",
",",
"vryheid",
"en",
... | 710 | 22.7 | 77 | py |
spaCy | spaCy-master/spacy/tests/lang/am/__init__.py | 0 | 0 | 0 | py | |
spaCy | spaCy-master/spacy/tests/lang/am/test_exception.py | 0 | 0 | 0 | py | |
spaCy | spaCy-master/spacy/tests/lang/am/test_text.py | import pytest
def test_am_tokenizer_handles_long_text(am_tokenizer):
text = """ሆሴ ሙጂካ በበጋ ወቅት በኦክስፎርድ ንግግር አንድያቀርቡ ሲጋበዙ ጭንቅላታቸው "ፈነዳ"።
“እጅግ ጥንታዊ” የእንግሊዝኛ ተናጋሪ ዩኒቨርስቲ፣ በአስር ሺዎች የሚቆጠሩ ዩሮዎችን ለተማሪዎች በማስተማር የሚያስከፍለው
እና ከማርጋሬት ታቸር እስከ ስቲቨን ሆኪንግ በአዳራሾቻቸው ውስጥ ንግግር ያደረጉበት የትምህርት ማዕከል፣ በሞንቴቪዴኦ
በሚገኘው የመንግስት ትምህርት ቤት የሰለጠ... | 1,259 | 23.230769 | 75 | py |
spaCy | spaCy-master/spacy/tests/lang/ar/__init__.py | 0 | 0 | 0 | py | |
spaCy | spaCy-master/spacy/tests/lang/ar/test_exceptions.py | import pytest
@pytest.mark.parametrize("text", ["ق.م", "إلخ", "ص.ب", "ت."])
def test_ar_tokenizer_handles_abbr(ar_tokenizer, text):
tokens = ar_tokenizer(text)
assert len(tokens) == 1
def test_ar_tokenizer_handles_exc_in_text(ar_tokenizer):
text = "تعود الكتابة الهيروغليفية إلى سنة 3200 ق.م"
tokens ... | 563 | 25.857143 | 61 | py |
spaCy | spaCy-master/spacy/tests/lang/ar/test_text.py | def test_ar_tokenizer_handles_long_text(ar_tokenizer):
text = """نجيب محفوظ مؤلف و كاتب روائي عربي، يعد من أهم الأدباء العرب خلال القرن العشرين.
ولد نجيب محفوظ في مدينة القاهرة، حيث ترعرع و تلقى تعليمه الجامعي في جامعتها،
فتمكن من نيل شهادة في الفلسفة. ألف محفوظ على مدار حياته الكثير من الأعمال الأدبية، ... | 539 | 53 | 118 | py |
spaCy | spaCy-master/spacy/tests/lang/bg/test_text.py | import pytest
@pytest.mark.parametrize(
"word,match",
[
("10", True),
("1", True),
("10000", True),
("1.000", True),
("бројка", False),
("999,23", True),
("едно", True),
("две", True),
("цифра", False),
("единайсет", True),
... | 685 | 21.129032 | 61 | py |
spaCy | spaCy-master/spacy/tests/lang/bg/test_tokenizer.py | import pytest
def test_bg_tokenizer_handles_final_diacritics(bg_tokenizer):
text = "Ня̀маше яйца̀. Ня̀маше яйца̀."
tokens = bg_tokenizer(text)
assert tokens[1].text == "яйца̀"
assert tokens[2].text == "."
| 223 | 23.888889 | 61 | py |
spaCy | spaCy-master/spacy/tests/lang/bn/__init__.py | 0 | 0 | 0 | py | |
spaCy | spaCy-master/spacy/tests/lang/bn/test_tokenizer.py | import pytest
# fmt: off
TESTCASES = [
# Punctuation tests
("আমি বাংলায় গান গাই!", ["আমি", "বাংলায়", "গান", "গাই", "!"]),
("আমি বাংলায় কথা কই।", ["আমি", "বাংলায়", "কথা", "কই", "।"]),
("বসুন্ধরা জনসম্মুখে দোষ স্বীকার করলো না?", ["বসুন্ধরা", "জনসম্মুখে", "দোষ", "স্বীকার", "করলো", "না", "?"]),
("ট... | 1,890 | 50.108108 | 148 | py |
spaCy | spaCy-master/spacy/tests/lang/ca/__init__.py | 0 | 0 | 0 | py | |
spaCy | spaCy-master/spacy/tests/lang/ca/test_exception.py | import pytest
@pytest.mark.parametrize(
"text,lemma",
[("aprox.", "aproximadament"), ("pàg.", "pàgina"), ("p.ex.", "per exemple")],
)
def test_ca_tokenizer_handles_abbr(ca_tokenizer, text, lemma):
tokens = ca_tokenizer(text)
assert len(tokens) == 1
def test_ca_tokenizer_handles_exc_in_text(ca_tokeni... | 617 | 20.310345 | 81 | py |
spaCy | spaCy-master/spacy/tests/lang/ca/test_prefix_suffix_infix.py | import pytest
@pytest.mark.parametrize(
"text,expected_tokens",
[
("d'un", ["d'", "un"]),
("s'ha", ["s'", "ha"]),
("del", ["d", "el"]),
("cantar-te", ["cantar", "-te"]),
("-hola", ["-", "hola"]),
],
)
def test_contractions(ca_tokenizer, text, expected_tokens):
"... | 493 | 25 | 62 | py |
spaCy | spaCy-master/spacy/tests/lang/ca/test_text.py | """Test that longer and mixed texts are tokenized correctly."""
import pytest
def test_ca_tokenizer_handles_long_text(ca_tokenizer):
text = """Una taula amb grans gerres de begudes i palles de coloraines com a reclam. Una carta
cridanera amb ofertes de tapes, paelles i sangria. Un cambrer amb un somriure que ... | 1,872 | 32.446429 | 98 | py |
spaCy | spaCy-master/spacy/tests/lang/cs/__init__.py | 0 | 0 | 0 | py | |
spaCy | spaCy-master/spacy/tests/lang/cs/test_text.py | import pytest
@pytest.mark.parametrize(
"text,match",
[
("10", True),
("1", True),
("10.000", True),
("1000", True),
("999,0", True),
("devatenáct", True),
("osmdesát", True),
("kvadrilion", True),
("Pes", False),
(",", False),
... | 508 | 20.208333 | 58 | py |
spaCy | spaCy-master/spacy/tests/lang/da/__init__.py | 0 | 0 | 0 | py | |
spaCy | spaCy-master/spacy/tests/lang/da/test_exceptions.py | import pytest
@pytest.mark.parametrize("text", ["ca.", "m.a.o.", "Jan.", "Dec.", "kr.", "jf."])
def test_da_tokenizer_handles_abbr(da_tokenizer, text):
tokens = da_tokenizer(text)
assert len(tokens) == 1
@pytest.mark.parametrize("text", ["Jul.", "jul.", "Tor.", "Tors."])
def test_da_tokenizer_handles_ambigu... | 1,815 | 29.266667 | 81 | py |
spaCy | spaCy-master/spacy/tests/lang/da/test_noun_chunks.py | import pytest
from spacy.tokens import Doc
def test_noun_chunks_is_parsed(da_tokenizer):
"""Test that noun_chunks raises Value Error for 'da' language if Doc is not parsed.
To check this test, we're constructing a Doc
with a new Vocab here and forcing is_parsed to 'False'
to make sure the noun chunks... | 2,060 | 27.625 | 87 | py |
spaCy | spaCy-master/spacy/tests/lang/da/test_prefix_suffix_infix.py | import pytest
@pytest.mark.parametrize("text", ["(under)"])
def test_da_tokenizer_splits_no_special(da_tokenizer, text):
tokens = da_tokenizer(text)
assert len(tokens) == 3
@pytest.mark.parametrize("text", ["ta'r", "Søren's", "Lars'"])
def test_da_tokenizer_handles_no_punct(da_tokenizer, text):
tokens =... | 5,417 | 31.059172 | 100 | py |
spaCy | spaCy-master/spacy/tests/lang/da/test_text.py | import pytest
from spacy.lang.da.lex_attrs import like_num
def test_da_tokenizer_handles_long_text(da_tokenizer):
text = """Der var så dejligt ude på landet. Det var sommer, kornet stod gult, havren grøn,
høet var rejst i stakke nede i de grønne enge, og der gik storken på sine lange,
røde ben og snakkede ægypti... | 1,208 | 27.785714 | 121 | py |
spaCy | spaCy-master/spacy/tests/lang/de/__init__.py | 0 | 0 | 0 | py | |
spaCy | spaCy-master/spacy/tests/lang/de/test_exceptions.py | import pytest
@pytest.mark.parametrize("text", ["auf'm", "du's", "über'm", "wir's"])
def test_de_tokenizer_splits_contractions(de_tokenizer, text):
tokens = de_tokenizer(text)
assert len(tokens) == 2
@pytest.mark.parametrize("text", ["z.B.", "d.h.", "Jan.", "Dez.", "Chr."])
def test_de_tokenizer_handles_abb... | 597 | 27.47619 | 74 | py |
spaCy | spaCy-master/spacy/tests/lang/de/test_noun_chunks.py | import pytest
def test_noun_chunks_is_parsed_de(de_tokenizer):
"""Test that noun_chunks raises Value Error for 'de' language if Doc is not parsed."""
doc = de_tokenizer("Er lag auf seinem")
with pytest.raises(ValueError):
list(doc.noun_chunks)
| 266 | 28.666667 | 90 | py |
spaCy | spaCy-master/spacy/tests/lang/de/test_parser.py | from spacy.tokens import Doc
def test_de_parser_noun_chunks_standard_de(de_vocab):
words = ["Eine", "Tasse", "steht", "auf", "dem", "Tisch", "."]
heads = [1, 2, 2, 2, 5, 3, 2]
pos = ["DET", "NOUN", "VERB", "ADP", "DET", "NOUN", "PUNCT"]
deps = ["nk", "sb", "ROOT", "mo", "nk", "nk", "punct"]
doc = ... | 1,186 | 39.931034 | 89 | py |
spaCy | spaCy-master/spacy/tests/lang/de/test_prefix_suffix_infix.py | import pytest
@pytest.mark.parametrize("text", ["(unter)"])
def test_de_tokenizer_splits_no_special(de_tokenizer, text):
tokens = de_tokenizer(text)
assert len(tokens) == 3
@pytest.mark.parametrize("text", ["unter'm"])
def test_de_tokenizer_splits_no_punct(de_tokenizer, text):
tokens = de_tokenizer(text... | 3,395 | 30.155963 | 88 | py |
spaCy | spaCy-master/spacy/tests/lang/de/test_text.py | import pytest
def test_de_tokenizer_handles_long_text(de_tokenizer):
text = """Die Verwandlung
Als Gregor Samsa eines Morgens aus unruhigen Träumen erwachte, fand er sich in
seinem Bett zu einem ungeheueren Ungeziefer verwandelt.
Er lag auf seinem panzerartig harten Rücken und sah, wenn er den Kopf ein wenig
ho... | 1,481 | 31.217391 | 88 | py |
spaCy | spaCy-master/spacy/tests/lang/dsb/__init__.py | 0 | 0 | 0 | py | |
spaCy | spaCy-master/spacy/tests/lang/dsb/test_text.py | import pytest
@pytest.mark.parametrize(
"text,match",
[
("10", True),
("1", True),
("10,000", True),
("10,00", True),
("jadno", True),
("dwanassćo", True),
("milion", True),
("sto", True),
("ceła", False),
("kopica", False),
... | 557 | 20.461538 | 59 | py |
spaCy | spaCy-master/spacy/tests/lang/dsb/test_tokenizer.py | import pytest
DSB_BASIC_TOKENIZATION_TESTS = [
(
"Ale eksistěrujo mimo togo ceła kopica narěcow, ako na pśikład slěpjańska.",
[
"Ale",
"eksistěrujo",
"mimo",
"togo",
"ceła",
"kopica",
"narěcow",
",",
... | 735 | 23.533333 | 84 | py |
spaCy | spaCy-master/spacy/tests/lang/el/__init__.py | 0 | 0 | 0 | py | |
spaCy | spaCy-master/spacy/tests/lang/el/test_exception.py | import pytest
@pytest.mark.parametrize("text", ["αριθ.", "τρισ.", "δισ.", "σελ."])
def test_el_tokenizer_handles_abbr(el_tokenizer, text):
tokens = el_tokenizer(text)
assert len(tokens) == 1
def test_el_tokenizer_handles_exc_in_text(el_tokenizer):
text = "Στα 14 τρισ. δολάρια το κόστος από την άνοδο της... | 442 | 28.533333 | 83 | py |
spaCy | spaCy-master/spacy/tests/lang/el/test_noun_chunks.py | import pytest
def test_noun_chunks_is_parsed_el(el_tokenizer):
"""Test that noun_chunks raises Value Error for 'el' language if Doc is not parsed."""
doc = el_tokenizer("είναι χώρα της νοτιοανατολικής")
with pytest.raises(ValueError):
list(doc.noun_chunks)
| 279 | 30.111111 | 90 | py |
spaCy | spaCy-master/spacy/tests/lang/el/test_text.py | import pytest
def test_el_tokenizer_handles_long_text(el_tokenizer):
text = """Η Ελλάδα (παλαιότερα Ελλάς), επίσημα γνωστή ως Ελληνική Δημοκρατία,\
είναι χώρα της νοτιοανατολικής Ευρώπης στο νοτιότερο άκρο της Βαλκανικής χερσονήσου.\
Συνορεύει στα βορειοδυτικά με την Αλβανία, στα βόρεια με την πρώην\
... | 1,195 | 36.375 | 121 | py |
spaCy | spaCy-master/spacy/tests/lang/en/__init__.py | 0 | 0 | 0 | py | |
spaCy | spaCy-master/spacy/tests/lang/en/test_customized_tokenizer.py | import re
import pytest
from spacy.lang.en import English
from spacy.tokenizer import Tokenizer
from spacy.util import compile_infix_regex, compile_prefix_regex, compile_suffix_regex
@pytest.fixture
def custom_en_tokenizer(en_vocab):
prefix_re = compile_prefix_regex(English.Defaults.prefixes)
suffix_re = co... | 3,774 | 23.354839 | 112 | py |
spaCy | spaCy-master/spacy/tests/lang/en/test_exceptions.py | import pytest
def test_en_tokenizer_handles_basic_contraction(en_tokenizer):
text = "don't giggle"
tokens = en_tokenizer(text)
assert len(tokens) == 3
assert tokens[1].text == "n't"
text = "i said don't!"
tokens = en_tokenizer(text)
assert len(tokens) == 5
assert tokens[4].text == "!"
... | 4,158 | 31.492188 | 88 | py |
spaCy | spaCy-master/spacy/tests/lang/en/test_indices.py | def test_en_simple_punct(en_tokenizer):
text = "to walk, do foo"
tokens = en_tokenizer(text)
assert tokens[0].idx == 0
assert tokens[1].idx == 3
assert tokens[2].idx == 7
assert tokens[3].idx == 9
assert tokens[4].idx == 12
def test_en_complex_punct(en_tokenizer):
text = "Tom (D., Ill.... | 723 | 26.846154 | 40 | py |
spaCy | spaCy-master/spacy/tests/lang/en/test_noun_chunks.py | import pytest
from spacy.tokens import Doc
@pytest.fixture
def doc(en_vocab):
words = ["Peter", "has", "chronic", "command", "and", "control", "issues"]
heads = [1, 1, 6, 6, 3, 3, 1]
deps = ["nsubj", "ROOT", "amod", "nmod", "cc", "conj", "dobj"]
pos = ["PROPN", "VERB", "ADJ", "NOUN", "CCONJ", "NOUN",... | 1,549 | 32.695652 | 90 | py |
spaCy | spaCy-master/spacy/tests/lang/en/test_parser.py | from spacy.tokens import Doc
def test_en_parser_noun_chunks_standard(en_vocab):
words = ["A", "base", "phrase", "should", "be", "recognized", "."]
heads = [2, 2, 5, 5, 5, 5, 5]
pos = ["DET", "ADJ", "NOUN", "AUX", "VERB", "VERB", "PUNCT"]
deps = ["det", "amod", "nsubjpass", "aux", "auxpass", "ROOT", "p... | 2,954 | 43.104478 | 110 | py |
spaCy | spaCy-master/spacy/tests/lang/en/test_prefix_suffix_infix.py | import pytest
@pytest.mark.parametrize("text", ["(can)"])
def test_en_tokenizer_splits_no_special(en_tokenizer, text):
tokens = en_tokenizer(text)
assert len(tokens) == 3
@pytest.mark.parametrize("text", ["can't"])
def test_en_tokenizer_splits_no_punct(en_tokenizer, text):
tokens = en_tokenizer(text)
... | 4,254 | 29.833333 | 80 | py |
spaCy | spaCy-master/spacy/tests/lang/en/test_punct.py | import pytest
from spacy.lang.punctuation import TOKENIZER_PREFIXES
from spacy.util import compile_prefix_regex
PUNCT_OPEN = ["(", "[", "{", "*"]
PUNCT_CLOSE = [")", "]", "}", "*"]
PUNCT_PAIRED = [("(", ")"), ("[", "]"), ("{", "}"), ("*", "*")]
@pytest.mark.parametrize("text", ["(", "((", "<"])
def test_en_tokenize... | 4,422 | 33.554688 | 88 | py |
spaCy | spaCy-master/spacy/tests/lang/en/test_sbd.py | import pytest
from spacy.tokens import Doc
from ...util import apply_transition_sequence
@pytest.mark.issue(309)
def test_issue309(en_vocab):
"""Test Issue #309: SBD fails on empty string"""
doc = Doc(en_vocab, words=[" "], heads=[0], deps=["ROOT"])
assert len(doc) == 1
sents = list(doc.sents)
a... | 1,708 | 35.361702 | 85 | py |
spaCy | spaCy-master/spacy/tests/lang/en/test_text.py | import pytest
from spacy.lang.en.lex_attrs import like_num
def test_en_tokenizer_handles_long_text(en_tokenizer):
text = """Tributes pour in for late British Labour Party leader
Tributes poured in from around the world Thursday
to the late Labour Party leader John Smith, who died earlier from a massive
heart at... | 1,959 | 26.605634 | 77 | py |
spaCy | spaCy-master/spacy/tests/lang/en/test_tokenizer.py | import pytest
@pytest.mark.issue(351)
def test_issue351(en_tokenizer):
doc = en_tokenizer(" This is a cat.")
assert doc[0].idx == 0
assert len(doc[0]) == 3
assert doc[1].idx == 3
@pytest.mark.issue(360)
def test_issue360(en_tokenizer):
"""Test tokenization of big ellipsis"""
tokens = en_to... | 5,924 | 32.100559 | 88 | py |
spaCy | spaCy-master/spacy/tests/lang/es/__init__.py | 0 | 0 | 0 | py | |
spaCy | spaCy-master/spacy/tests/lang/es/test_exception.py | import pytest
@pytest.mark.parametrize(
"text,lemma",
[
("aprox.", "aproximadamente"),
("esq.", "esquina"),
("pág.", "página"),
("p.ej.", "por ejemplo"),
],
)
def test_es_tokenizer_handles_abbr(es_tokenizer, text, lemma):
tokens = es_tokenizer(text)
assert len(token... | 546 | 22.782609 | 62 | py |
spaCy | spaCy-master/spacy/tests/lang/es/test_noun_chunks.py | import pytest
from spacy.tokens import Doc
# fmt: off
@pytest.mark.parametrize(
"words,heads,deps,pos,chunk_offsets",
[
# un gato -> "un gato"
(
["un", "gato"],
[1, 1],
["det", "ROOT"],
["DET", "NOUN"],
[(0, 2)],
),
#... | 9,901 | 60.8875 | 452 | py |
spaCy | spaCy-master/spacy/tests/lang/es/test_text.py | import pytest
from spacy.lang.es import Spanish
from spacy.lang.es.lex_attrs import like_num
@pytest.mark.issue(3803)
def test_issue3803():
"""Test that spanish num-like tokens have True for like_num attribute."""
nlp = Spanish()
text = "2 dos 1000 mil 12 doce"
doc = nlp(text)
assert [t.like_num... | 2,004 | 25.733333 | 83 | py |
spaCy | spaCy-master/spacy/tests/lang/et/__init__.py | 0 | 0 | 0 | py | |
spaCy | spaCy-master/spacy/tests/lang/et/test_text.py | import pytest
def test_long_text(et_tokenizer):
# Excerpt: European Convention on Human Rights
text = """
arvestades, et nimetatud deklaratsiooni eesmärk on tagada selles
kuulutatud õiguste üldine ja tõhus tunnustamine ning järgimine;
arvestades, et Euroopa Nõukogu eesmärk on saavutada tema
liikmete suurem üh... | 909 | 32.703704 | 66 | py |
spaCy | spaCy-master/spacy/tests/lang/et/test_tokenizer.py | import pytest
ET_BASIC_TOKENIZATION_TESTS = [
(
"Kedagi ei või piinata ega ebainimlikult või alandavalt kohelda "
"ega karistada.",
[
"Kedagi",
"ei",
"või",
"piinata",
"ega",
"ebainimlikult",
"või",
... | 733 | 23.466667 | 77 | py |
spaCy | spaCy-master/spacy/tests/lang/eu/__init__.py | 0 | 0 | 0 | py |