Search is not available for this dataset
repo
stringlengths
2
152
file
stringlengths
15
239
code
stringlengths
0
58.4M
file_length
int64
0
58.4M
avg_line_length
float64
0
1.81M
max_line_length
int64
0
12.7M
extension_type
stringclasses
364 values
spaCy
spaCy-master/spacy/pipeline/spancat.py
from dataclasses import dataclass from functools import partial from typing import Any, Callable, Dict, Iterable, List, Optional, Tuple, Union, cast import numpy from thinc.api import Config, Model, Ops, Optimizer, get_current_ops, set_dropout_rate from thinc.types import Floats2d, Ints1d, Ints2d, Ragged from ..compa...
30,032
37.112944
101
py
spaCy
spaCy-master/spacy/pipeline/textcat.py
from itertools import islice from typing import Any, Callable, Dict, Iterable, List, Optional, Tuple import numpy from thinc.api import Config, Model, Optimizer, get_array_module, set_dropout_rate from thinc.types import Floats2d from ..errors import Errors from ..language import Language from ..scorer import Scorer ...
14,943
35.537897
104
py
spaCy
spaCy-master/spacy/pipeline/textcat_multilabel.py
from itertools import islice from typing import Any, Callable, Dict, Iterable, List, Optional from thinc.api import Config, Model from thinc.types import Floats2d from ..errors import Errors from ..language import Language from ..scorer import Scorer from ..tokens import Doc from ..training import Example, validate_g...
6,605
30.457143
94
py
spaCy
spaCy-master/spacy/pipeline/tok2vec.py
from itertools import islice from typing import Any, Callable, Dict, Iterable, List, Optional, Sequence from thinc.api import Config, Model, Optimizer, set_dropout_rate from ..errors import Errors from ..language import Language from ..tokens import Doc from ..training import Example, validate_examples, validate_get_...
13,390
40.458204
104
py
spaCy
spaCy-master/spacy/pipeline/_edit_tree_internals/__init__.py
0
0
0
py
spaCy
spaCy-master/spacy/pipeline/_edit_tree_internals/schemas.py
from collections import defaultdict from typing import Any, Dict, List, Union from pydantic import BaseModel, Field, ValidationError from pydantic.types import StrictBool, StrictInt, StrictStr class MatchNodeSchema(BaseModel): prefix_len: StrictInt = Field(..., title="Prefix length") suffix_len: StrictInt = ...
1,475
31.086957
99
py
spaCy
spaCy-master/spacy/pipeline/_parser_internals/__init__.py
0
0
0
py
spaCy
spaCy-master/spacy/pipeline/legacy/__init__.py
from .entity_linker import EntityLinker_v1 __all__ = ["EntityLinker_v1"]
74
17.75
42
py
spaCy
spaCy-master/spacy/pipeline/legacy/entity_linker.py
# This file is present to provide a prior version of the EntityLinker component # for backwards compatability. For details see #9669. import random import warnings from itertools import islice from pathlib import Path from typing import Any, Callable, Dict, Iterable, List, Optional, Union import srsly from thinc.api ...
18,788
43.41844
120
py
spaCy
spaCy-master/spacy/tests/README.md
<a href="https://explosion.ai"><img src="https://explosion.ai/assets/img/logo.svg" width="125" height="125" align="right" /></a> # spaCy tests spaCy uses the [pytest](http://doc.pytest.org/) framework for testing. For more info on this, see the [pytest documentation](http://docs.pytest.org/en/latest/contents.html). ...
8,728
62.253623
421
md
spaCy
spaCy-master/spacy/tests/__init__.py
0
0
0
py
spaCy
spaCy-master/spacy/tests/conftest.py
import pytest from hypothesis import settings from spacy.util import get_lang_class # Functionally disable deadline settings for tests # to prevent spurious test failures in CI builds. settings.register_profile("no_deadlines", deadline=2 * 60 * 1000) # in ms settings.load_profile("no_deadlines") def pytest_addopti...
11,609
21.764706
83
py
spaCy
spaCy-master/spacy/tests/enable_gpu.py
from spacy import require_gpu require_gpu()
45
10.5
29
py
spaCy
spaCy-master/spacy/tests/test_architectures.py
import pytest from catalogue import RegistryError from thinc.api import Linear from spacy import registry def test_get_architecture(): @registry.architectures("my_test_function") def create_model(nr_in, nr_out): return Linear(nr_in, nr_out) arch = registry.architectures.get("my_test_function") ...
448
25.411765
57
py
spaCy
spaCy-master/spacy/tests/test_cli.py
import math import os import time from collections import Counter from pathlib import Path from typing import Any, Dict, List, Tuple import numpy import pytest import srsly from click import NoSuchOption from packaging.specifiers import SpecifierSet from thinc.api import Config, ConfigValidationError import spacy fro...
49,789
35.799704
229
py
spaCy
spaCy-master/spacy/tests/test_cli_app.py
import os from pathlib import Path import pytest import srsly from typer.testing import CliRunner from spacy.cli._util import app, get_git_version from spacy.tokens import Doc, DocBin from .util import make_tempdir, normalize_whitespace def has_git(): try: get_git_version() return True exce...
7,767
31.915254
111
py
spaCy
spaCy-master/spacy/tests/test_displacy.py
import numpy import pytest from spacy import displacy from spacy.displacy.render import DependencyRenderer, EntityRenderer from spacy.lang.en import English from spacy.lang.fa import Persian from spacy.tokens import Doc, Span @pytest.mark.issue(2361) def test_issue2361(de_vocab): """Test if < is escaped when ren...
14,268
34.761905
92
py
spaCy
spaCy-master/spacy/tests/test_errors.py
from inspect import isclass import pytest from spacy.errors import ErrorsWithCodes class Errors(metaclass=ErrorsWithCodes): E001 = "error description" def test_add_codes(): assert Errors.E001 == "[E001] error description" with pytest.raises(AttributeError): Errors.E002 assert isclass(Error...
333
18.647059
52
py
spaCy
spaCy-master/spacy/tests/test_language.py
import itertools import logging from unittest import mock import pytest from thinc.api import CupyOps, NumpyOps, get_current_ops import spacy from spacy.lang.de import German from spacy.lang.en import English from spacy.language import Language from spacy.scorer import Scorer from spacy.tokens import Doc, Span from s...
26,993
32.616438
92
py
spaCy
spaCy-master/spacy/tests/test_misc.py
import ctypes import os from pathlib import Path import pytest from pydantic import ValidationError from thinc.api import ( Config, ConfigValidationError, CupyOps, MPSOps, NumpyOps, Optimizer, get_current_ops, set_current_ops, ) from thinc.compat import has_cupy_gpu, has_torch_mps_gpu ...
15,386
30.27439
88
py
spaCy
spaCy-master/spacy/tests/test_models.py
from typing import List import numpy import pytest from numpy.testing import assert_array_almost_equal, assert_array_equal from thinc.api import ( Adam, Logistic, Ragged, Relu, chain, fix_random_seed, reduce_mean, set_dropout_rate, ) from spacy.lang.en import English from spacy.lang.en...
9,082
30.648084
88
py
spaCy
spaCy-master/spacy/tests/test_pickles.py
import numpy import pytest import srsly from spacy.attrs import NORM from spacy.lang.en import English from spacy.strings import StringStore from spacy.tokens import Doc from spacy.vocab import Vocab @pytest.mark.parametrize("text1,text2", [("hello", "bye")]) def test_pickle_string_store(text1, text2): stringsto...
2,023
31.126984
77
py
spaCy
spaCy-master/spacy/tests/test_scorer.py
import pytest from numpy.testing import assert_almost_equal, assert_array_almost_equal from pytest import approx from spacy.lang.en import English from spacy.scorer import PRFScore, ROCAUCScore, Scorer, _roc_auc_score, _roc_curve from spacy.tokens import Doc, Span from spacy.training import Example from spacy.training...
17,478
31.488848
87
py
spaCy
spaCy-master/spacy/tests/test_ty.py
import spacy from spacy import ty def test_component_types(): nlp = spacy.blank("en") tok2vec = nlp.create_pipe("tok2vec") tagger = nlp.create_pipe("tagger") entity_ruler = nlp.create_pipe("entity_ruler") assert isinstance(tok2vec, ty.TrainableComponent) assert isinstance(tagger, ty.TrainableC...
748
38.421053
63
py
spaCy
spaCy-master/spacy/tests/util.py
import contextlib import re import tempfile import numpy import srsly from thinc.api import get_current_ops from spacy.tokens import Doc from spacy.training import split_bilu_label from spacy.util import make_tempdir # noqa: F401 from spacy.vocab import Vocab @contextlib.contextmanager def make_tempfile(mode="r"):...
3,281
30.257143
79
py
spaCy
spaCy-master/spacy/tests/doc/__init__.py
0
0
0
py
spaCy
spaCy-master/spacy/tests/doc/test_add_entities.py
import pytest from spacy import registry from spacy.pipeline import EntityRecognizer from spacy.pipeline.ner import DEFAULT_NER_MODEL from spacy.tokens import Doc, Span from spacy.training import Example def _ner_example(ner): doc = Doc( ner.vocab, words=["Joe", "loves", "visiting", "London", "du...
1,804
30.12069
81
py
spaCy
spaCy-master/spacy/tests/doc/test_array.py
import numpy import pytest from spacy.attrs import DEP, MORPH, ORTH, POS, SHAPE from spacy.tokens import Doc @pytest.mark.issue(2203) def test_issue2203(en_vocab): """Test that lemmas are set correctly in doc.from_array.""" words = ["I", "'ll", "survive"] tags = ["PRP", "MD", "VB"] lemmas = ["-PRON-"...
4,988
35.416058
81
py
spaCy
spaCy-master/spacy/tests/doc/test_creation.py
import pytest from spacy import util from spacy.tokens import Doc from spacy.vocab import Vocab @pytest.fixture def vocab(): return Vocab() def test_empty_doc(vocab): doc = Doc(vocab) assert len(doc) == 0 def test_single_word(vocab): doc = Doc(vocab, words=["a"]) assert doc.text == "a " d...
2,651
31.740741
80
py
spaCy
spaCy-master/spacy/tests/doc/test_doc_api.py
import warnings import weakref import numpy import pytest from numpy.testing import assert_array_equal from thinc.api import NumpyOps, get_current_ops from spacy.attrs import ( DEP, ENT_IOB, ENT_TYPE, HEAD, IS_ALPHA, MORPH, POS, SENT_START, TAG, ) from spacy.lang.en import English ...
35,906
34.835329
192
py
spaCy
spaCy-master/spacy/tests/doc/test_graph.py
from spacy.tokens.doc import Doc from spacy.tokens.graph import Graph from spacy.vocab import Vocab def test_graph_init(): doc = Doc(Vocab(), words=["a", "b", "c", "d"]) graph = Graph(doc, name="hello") assert graph.name == "hello" assert graph.doc is doc def test_graph_edges_and_nodes(): doc = ...
1,819
36.142857
77
py
spaCy
spaCy-master/spacy/tests/doc/test_json_doc_conversion.py
import pytest import srsly import spacy from spacy import schemas from spacy.tokens import Doc, Span, Token from .test_underscore import clean_underscore # noqa: F401 @pytest.fixture() def doc(en_vocab): words = ["c", "d", "e"] spaces = [True, True, True] pos = ["VERB", "NOUN", "NOUN"] tags = ["VBP...
13,660
34.575521
108
py
spaCy
spaCy-master/spacy/tests/doc/test_morphanalysis.py
import pytest @pytest.fixture def i_has(en_tokenizer): doc = en_tokenizer("I has") doc[0].set_morph({"PronType": "prs"}) doc[1].set_morph( { "VerbForm": "fin", "Tense": "pres", "Number": "sing", "Person": "three", } ) return doc de...
3,145
30.148515
87
py
spaCy
spaCy-master/spacy/tests/doc/test_pickle_doc.py
from spacy.compat import pickle from spacy.language import Language def test_pickle_single_doc(): nlp = Language() doc = nlp("pickle roundtrip") data = pickle.dumps(doc, 1) doc2 = pickle.loads(data) assert doc2.text == "pickle roundtrip" def test_list_of_docs_pickles_efficiently(): nlp = Lan...
1,470
25.745455
53
py
spaCy
spaCy-master/spacy/tests/doc/test_retokenize_merge.py
import pytest from spacy.attrs import LEMMA from spacy.tokens import Doc, Token from spacy.vocab import Vocab def test_doc_retokenize_merge(en_tokenizer): text = "WKRO played songs by the beach boys all night" attrs = { "tag": "NAMED", "lemma": "LEMMA", "ent_type": "TYPE", "mo...
18,956
37.219758
122
py
spaCy
spaCy-master/spacy/tests/doc/test_retokenize_split.py
import numpy import pytest from spacy.tokens import Doc, Token from spacy.vocab import Vocab @pytest.mark.issue(3540) def test_issue3540(en_vocab): words = ["I", "live", "in", "NewYork", "right", "now"] tensor = numpy.asarray( [[1.0, 1.1], [2.0, 2.1], [3.0, 3.1], [4.0, 4.1], [5.0, 5.1], [6.0, 6.1]], ...
10,937
35.828283
92
py
spaCy
spaCy-master/spacy/tests/doc/test_span.py
import numpy import pytest from numpy.testing import assert_array_equal from thinc.api import get_current_ops from spacy.attrs import LENGTH, ORTH from spacy.lang.en import English from spacy.tokens import Doc, Span, Token from spacy.util import filter_spans from spacy.vocab import Vocab from ..util import add_vecs_t...
25,422
33.63624
137
py
spaCy
spaCy-master/spacy/tests/doc/test_span_group.py
from random import Random from typing import List import pytest from spacy.matcher import Matcher from spacy.tokens import Doc, Span, SpanGroup from spacy.util import filter_spans @pytest.fixture def doc(en_tokenizer): doc = en_tokenizer("0 1 2 3 4 5 6") matcher = Matcher(en_tokenizer.vocab, validate=True) ...
8,667
29.846975
98
py
spaCy
spaCy-master/spacy/tests/doc/test_token_api.py
import numpy import pytest from spacy.attrs import IS_ALPHA, IS_DIGIT, IS_LOWER, IS_PUNCT, IS_STOP, IS_TITLE from spacy.symbols import VERB from spacy.tokens import Doc from spacy.training import Example from spacy.vocab import Vocab @pytest.fixture def doc(en_vocab): # fmt: off words = ["This", "is", "a", "...
11,165
36.59596
116
py
spaCy
spaCy-master/spacy/tests/doc/test_underscore.py
import pytest from mock import Mock from spacy.tokens import Doc, Span, Token from spacy.tokens.underscore import Underscore @pytest.fixture(scope="function", autouse=True) def clean_underscore(): # reset the Underscore object after the test, to avoid having state copied across tests yield Underscore.doc...
5,565
30.805714
91
py
spaCy
spaCy-master/spacy/tests/lang/__init__.py
0
0
0
py
spaCy
spaCy-master/spacy/tests/lang/test_attrs.py
import pytest from spacy.attrs import ENT_IOB, IS_ALPHA, LEMMA, NORM, ORTH, intify_attrs from spacy.lang.en.stop_words import STOP_WORDS from spacy.lang.lex_attrs import ( is_ascii, is_currency, is_punct, is_stop, like_url, word_shape, ) @pytest.mark.parametrize("word", ["the"]) @pytest.mark....
3,522
24.904412
87
py
spaCy
spaCy-master/spacy/tests/lang/test_initialize.py
import pytest from spacy.util import get_lang_class # fmt: off # Only include languages with no external dependencies # excluded: ja, ko, th, vi, zh LANGUAGES = ["af", "am", "ar", "az", "bg", "bn", "ca", "cs", "da", "de", "el", "en", "es", "et", "eu", "fa", "fi", "fr", "ga", "gu", "he", "hi", ...
922
35.92
78
py
spaCy
spaCy-master/spacy/tests/lang/test_lemmatizers.py
import pytest from spacy import registry from spacy.lookups import Lookups from spacy.util import get_lang_class # fmt: off # Only include languages with no external dependencies # excluded: ru, uk # excluded for custom tables: es, pl LANGUAGES = ["bn", "ca", "el", "en", "fa", "fr", "nb", "nl", "sv"] # fmt: on @pyt...
1,910
33.745455
71
py
spaCy
spaCy-master/spacy/tests/lang/af/__init__.py
0
0
0
py
spaCy
spaCy-master/spacy/tests/lang/af/test_text.py
import pytest def test_long_text(af_tokenizer): # Excerpt: Universal Declaration of Human Rights; “'n” changed to “die” in first sentence text = """ Hierdie Universele Verklaring van Menseregte as die algemene standaard vir die verwesenliking deur alle mense en nasies, om te verseker dat elke individu en elk...
923
39.173913
122
py
spaCy
spaCy-master/spacy/tests/lang/af/test_tokenizer.py
import pytest AF_BASIC_TOKENIZATION_TESTS = [ ( "Elkeen het die reg tot lewe, vryheid en sekuriteit van persoon.", [ "Elkeen", "het", "die", "reg", "tot", "lewe", ",", "vryheid", "en", ...
710
22.7
77
py
spaCy
spaCy-master/spacy/tests/lang/am/__init__.py
0
0
0
py
spaCy
spaCy-master/spacy/tests/lang/am/test_exception.py
0
0
0
py
spaCy
spaCy-master/spacy/tests/lang/am/test_text.py
import pytest def test_am_tokenizer_handles_long_text(am_tokenizer): text = """ሆሴ ሙጂካ በበጋ ወቅት በኦክስፎርድ ንግግር አንድያቀርቡ ሲጋበዙ ጭንቅላታቸው "ፈነዳ"። “እጅግ ጥንታዊ” የእንግሊዝኛ ተናጋሪ ዩኒቨርስቲ፣ በአስር ሺዎች የሚቆጠሩ ዩሮዎችን ለተማሪዎች በማስተማር የሚያስከፍለው እና ከማርጋሬት ታቸር እስከ ስቲቨን ሆኪንግ በአዳራሾቻቸው ውስጥ ንግግር ያደረጉበት የትምህርት ማዕከል፣ በሞንቴቪዴኦ በሚገኘው የመንግስት ትምህርት ቤት የሰለጠ...
1,259
23.230769
75
py
spaCy
spaCy-master/spacy/tests/lang/ar/__init__.py
0
0
0
py
spaCy
spaCy-master/spacy/tests/lang/ar/test_exceptions.py
import pytest @pytest.mark.parametrize("text", ["ق.م", "إلخ", "ص.ب", "ت."]) def test_ar_tokenizer_handles_abbr(ar_tokenizer, text): tokens = ar_tokenizer(text) assert len(tokens) == 1 def test_ar_tokenizer_handles_exc_in_text(ar_tokenizer): text = "تعود الكتابة الهيروغليفية إلى سنة 3200 ق.م" tokens ...
563
25.857143
61
py
spaCy
spaCy-master/spacy/tests/lang/ar/test_text.py
def test_ar_tokenizer_handles_long_text(ar_tokenizer): text = """نجيب محفوظ مؤلف و كاتب روائي عربي، يعد من أهم الأدباء العرب خلال القرن العشرين. ولد نجيب محفوظ في مدينة القاهرة، حيث ترعرع و تلقى تعليمه الجامعي في جامعتها، فتمكن من نيل شهادة في الفلسفة. ألف محفوظ على مدار حياته الكثير من الأعمال الأدبية، ...
539
53
118
py
spaCy
spaCy-master/spacy/tests/lang/bg/test_text.py
import pytest @pytest.mark.parametrize( "word,match", [ ("10", True), ("1", True), ("10000", True), ("1.000", True), ("бројка", False), ("999,23", True), ("едно", True), ("две", True), ("цифра", False), ("единайсет", True), ...
685
21.129032
61
py
spaCy
spaCy-master/spacy/tests/lang/bg/test_tokenizer.py
import pytest def test_bg_tokenizer_handles_final_diacritics(bg_tokenizer): text = "Ня̀маше яйца̀. Ня̀маше яйца̀." tokens = bg_tokenizer(text) assert tokens[1].text == "яйца̀" assert tokens[2].text == "."
223
23.888889
61
py
spaCy
spaCy-master/spacy/tests/lang/bn/__init__.py
0
0
0
py
spaCy
spaCy-master/spacy/tests/lang/bn/test_tokenizer.py
import pytest # fmt: off TESTCASES = [ # Punctuation tests ("আমি বাংলায় গান গাই!", ["আমি", "বাংলায়", "গান", "গাই", "!"]), ("আমি বাংলায় কথা কই।", ["আমি", "বাংলায়", "কথা", "কই", "।"]), ("বসুন্ধরা জনসম্মুখে দোষ স্বীকার করলো না?", ["বসুন্ধরা", "জনসম্মুখে", "দোষ", "স্বীকার", "করলো", "না", "?"]), ("ট...
1,890
50.108108
148
py
spaCy
spaCy-master/spacy/tests/lang/ca/__init__.py
0
0
0
py
spaCy
spaCy-master/spacy/tests/lang/ca/test_exception.py
import pytest @pytest.mark.parametrize( "text,lemma", [("aprox.", "aproximadament"), ("pàg.", "pàgina"), ("p.ex.", "per exemple")], ) def test_ca_tokenizer_handles_abbr(ca_tokenizer, text, lemma): tokens = ca_tokenizer(text) assert len(tokens) == 1 def test_ca_tokenizer_handles_exc_in_text(ca_tokeni...
617
20.310345
81
py
spaCy
spaCy-master/spacy/tests/lang/ca/test_prefix_suffix_infix.py
import pytest @pytest.mark.parametrize( "text,expected_tokens", [ ("d'un", ["d'", "un"]), ("s'ha", ["s'", "ha"]), ("del", ["d", "el"]), ("cantar-te", ["cantar", "-te"]), ("-hola", ["-", "hola"]), ], ) def test_contractions(ca_tokenizer, text, expected_tokens): "...
493
25
62
py
spaCy
spaCy-master/spacy/tests/lang/ca/test_text.py
"""Test that longer and mixed texts are tokenized correctly.""" import pytest def test_ca_tokenizer_handles_long_text(ca_tokenizer): text = """Una taula amb grans gerres de begudes i palles de coloraines com a reclam. Una carta cridanera amb ofertes de tapes, paelles i sangria. Un cambrer amb un somriure que ...
1,872
32.446429
98
py
spaCy
spaCy-master/spacy/tests/lang/cs/__init__.py
0
0
0
py
spaCy
spaCy-master/spacy/tests/lang/cs/test_text.py
import pytest @pytest.mark.parametrize( "text,match", [ ("10", True), ("1", True), ("10.000", True), ("1000", True), ("999,0", True), ("devatenáct", True), ("osmdesát", True), ("kvadrilion", True), ("Pes", False), (",", False), ...
508
20.208333
58
py
spaCy
spaCy-master/spacy/tests/lang/da/__init__.py
0
0
0
py
spaCy
spaCy-master/spacy/tests/lang/da/test_exceptions.py
import pytest @pytest.mark.parametrize("text", ["ca.", "m.a.o.", "Jan.", "Dec.", "kr.", "jf."]) def test_da_tokenizer_handles_abbr(da_tokenizer, text): tokens = da_tokenizer(text) assert len(tokens) == 1 @pytest.mark.parametrize("text", ["Jul.", "jul.", "Tor.", "Tors."]) def test_da_tokenizer_handles_ambigu...
1,815
29.266667
81
py
spaCy
spaCy-master/spacy/tests/lang/da/test_noun_chunks.py
import pytest from spacy.tokens import Doc def test_noun_chunks_is_parsed(da_tokenizer): """Test that noun_chunks raises Value Error for 'da' language if Doc is not parsed. To check this test, we're constructing a Doc with a new Vocab here and forcing is_parsed to 'False' to make sure the noun chunks...
2,060
27.625
87
py
spaCy
spaCy-master/spacy/tests/lang/da/test_prefix_suffix_infix.py
import pytest @pytest.mark.parametrize("text", ["(under)"]) def test_da_tokenizer_splits_no_special(da_tokenizer, text): tokens = da_tokenizer(text) assert len(tokens) == 3 @pytest.mark.parametrize("text", ["ta'r", "Søren's", "Lars'"]) def test_da_tokenizer_handles_no_punct(da_tokenizer, text): tokens =...
5,417
31.059172
100
py
spaCy
spaCy-master/spacy/tests/lang/da/test_text.py
import pytest from spacy.lang.da.lex_attrs import like_num def test_da_tokenizer_handles_long_text(da_tokenizer): text = """Der var så dejligt ude på landet. Det var sommer, kornet stod gult, havren grøn, høet var rejst i stakke nede i de grønne enge, og der gik storken på sine lange, røde ben og snakkede ægypti...
1,208
27.785714
121
py
spaCy
spaCy-master/spacy/tests/lang/de/__init__.py
0
0
0
py
spaCy
spaCy-master/spacy/tests/lang/de/test_exceptions.py
import pytest @pytest.mark.parametrize("text", ["auf'm", "du's", "über'm", "wir's"]) def test_de_tokenizer_splits_contractions(de_tokenizer, text): tokens = de_tokenizer(text) assert len(tokens) == 2 @pytest.mark.parametrize("text", ["z.B.", "d.h.", "Jan.", "Dez.", "Chr."]) def test_de_tokenizer_handles_abb...
597
27.47619
74
py
spaCy
spaCy-master/spacy/tests/lang/de/test_noun_chunks.py
import pytest def test_noun_chunks_is_parsed_de(de_tokenizer): """Test that noun_chunks raises Value Error for 'de' language if Doc is not parsed.""" doc = de_tokenizer("Er lag auf seinem") with pytest.raises(ValueError): list(doc.noun_chunks)
266
28.666667
90
py
spaCy
spaCy-master/spacy/tests/lang/de/test_parser.py
from spacy.tokens import Doc def test_de_parser_noun_chunks_standard_de(de_vocab): words = ["Eine", "Tasse", "steht", "auf", "dem", "Tisch", "."] heads = [1, 2, 2, 2, 5, 3, 2] pos = ["DET", "NOUN", "VERB", "ADP", "DET", "NOUN", "PUNCT"] deps = ["nk", "sb", "ROOT", "mo", "nk", "nk", "punct"] doc = ...
1,186
39.931034
89
py
spaCy
spaCy-master/spacy/tests/lang/de/test_prefix_suffix_infix.py
import pytest @pytest.mark.parametrize("text", ["(unter)"]) def test_de_tokenizer_splits_no_special(de_tokenizer, text): tokens = de_tokenizer(text) assert len(tokens) == 3 @pytest.mark.parametrize("text", ["unter'm"]) def test_de_tokenizer_splits_no_punct(de_tokenizer, text): tokens = de_tokenizer(text...
3,395
30.155963
88
py
spaCy
spaCy-master/spacy/tests/lang/de/test_text.py
import pytest def test_de_tokenizer_handles_long_text(de_tokenizer): text = """Die Verwandlung Als Gregor Samsa eines Morgens aus unruhigen Träumen erwachte, fand er sich in seinem Bett zu einem ungeheueren Ungeziefer verwandelt. Er lag auf seinem panzerartig harten Rücken und sah, wenn er den Kopf ein wenig ho...
1,481
31.217391
88
py
spaCy
spaCy-master/spacy/tests/lang/dsb/__init__.py
0
0
0
py
spaCy
spaCy-master/spacy/tests/lang/dsb/test_text.py
import pytest @pytest.mark.parametrize( "text,match", [ ("10", True), ("1", True), ("10,000", True), ("10,00", True), ("jadno", True), ("dwanassćo", True), ("milion", True), ("sto", True), ("ceła", False), ("kopica", False), ...
557
20.461538
59
py
spaCy
spaCy-master/spacy/tests/lang/dsb/test_tokenizer.py
import pytest DSB_BASIC_TOKENIZATION_TESTS = [ ( "Ale eksistěrujo mimo togo ceła kopica narěcow, ako na pśikład slěpjańska.", [ "Ale", "eksistěrujo", "mimo", "togo", "ceła", "kopica", "narěcow", ",", ...
735
23.533333
84
py
spaCy
spaCy-master/spacy/tests/lang/el/__init__.py
0
0
0
py
spaCy
spaCy-master/spacy/tests/lang/el/test_exception.py
import pytest @pytest.mark.parametrize("text", ["αριθ.", "τρισ.", "δισ.", "σελ."]) def test_el_tokenizer_handles_abbr(el_tokenizer, text): tokens = el_tokenizer(text) assert len(tokens) == 1 def test_el_tokenizer_handles_exc_in_text(el_tokenizer): text = "Στα 14 τρισ. δολάρια το κόστος από την άνοδο της...
442
28.533333
83
py
spaCy
spaCy-master/spacy/tests/lang/el/test_noun_chunks.py
import pytest def test_noun_chunks_is_parsed_el(el_tokenizer): """Test that noun_chunks raises Value Error for 'el' language if Doc is not parsed.""" doc = el_tokenizer("είναι χώρα της νοτιοανατολικής") with pytest.raises(ValueError): list(doc.noun_chunks)
279
30.111111
90
py
spaCy
spaCy-master/spacy/tests/lang/el/test_text.py
import pytest def test_el_tokenizer_handles_long_text(el_tokenizer): text = """Η Ελλάδα (παλαιότερα Ελλάς), επίσημα γνωστή ως Ελληνική Δημοκρατία,\ είναι χώρα της νοτιοανατολικής Ευρώπης στο νοτιότερο άκρο της Βαλκανικής χερσονήσου.\ Συνορεύει στα βορειοδυτικά με την Αλβανία, στα βόρεια με την πρώην\ ...
1,195
36.375
121
py
spaCy
spaCy-master/spacy/tests/lang/en/__init__.py
0
0
0
py
spaCy
spaCy-master/spacy/tests/lang/en/test_customized_tokenizer.py
import re import pytest from spacy.lang.en import English from spacy.tokenizer import Tokenizer from spacy.util import compile_infix_regex, compile_prefix_regex, compile_suffix_regex @pytest.fixture def custom_en_tokenizer(en_vocab): prefix_re = compile_prefix_regex(English.Defaults.prefixes) suffix_re = co...
3,774
23.354839
112
py
spaCy
spaCy-master/spacy/tests/lang/en/test_exceptions.py
import pytest def test_en_tokenizer_handles_basic_contraction(en_tokenizer): text = "don't giggle" tokens = en_tokenizer(text) assert len(tokens) == 3 assert tokens[1].text == "n't" text = "i said don't!" tokens = en_tokenizer(text) assert len(tokens) == 5 assert tokens[4].text == "!" ...
4,158
31.492188
88
py
spaCy
spaCy-master/spacy/tests/lang/en/test_indices.py
def test_en_simple_punct(en_tokenizer): text = "to walk, do foo" tokens = en_tokenizer(text) assert tokens[0].idx == 0 assert tokens[1].idx == 3 assert tokens[2].idx == 7 assert tokens[3].idx == 9 assert tokens[4].idx == 12 def test_en_complex_punct(en_tokenizer): text = "Tom (D., Ill....
723
26.846154
40
py
spaCy
spaCy-master/spacy/tests/lang/en/test_noun_chunks.py
import pytest from spacy.tokens import Doc @pytest.fixture def doc(en_vocab): words = ["Peter", "has", "chronic", "command", "and", "control", "issues"] heads = [1, 1, 6, 6, 3, 3, 1] deps = ["nsubj", "ROOT", "amod", "nmod", "cc", "conj", "dobj"] pos = ["PROPN", "VERB", "ADJ", "NOUN", "CCONJ", "NOUN",...
1,549
32.695652
90
py
spaCy
spaCy-master/spacy/tests/lang/en/test_parser.py
from spacy.tokens import Doc def test_en_parser_noun_chunks_standard(en_vocab): words = ["A", "base", "phrase", "should", "be", "recognized", "."] heads = [2, 2, 5, 5, 5, 5, 5] pos = ["DET", "ADJ", "NOUN", "AUX", "VERB", "VERB", "PUNCT"] deps = ["det", "amod", "nsubjpass", "aux", "auxpass", "ROOT", "p...
2,954
43.104478
110
py
spaCy
spaCy-master/spacy/tests/lang/en/test_prefix_suffix_infix.py
import pytest @pytest.mark.parametrize("text", ["(can)"]) def test_en_tokenizer_splits_no_special(en_tokenizer, text): tokens = en_tokenizer(text) assert len(tokens) == 3 @pytest.mark.parametrize("text", ["can't"]) def test_en_tokenizer_splits_no_punct(en_tokenizer, text): tokens = en_tokenizer(text) ...
4,254
29.833333
80
py
spaCy
spaCy-master/spacy/tests/lang/en/test_punct.py
import pytest from spacy.lang.punctuation import TOKENIZER_PREFIXES from spacy.util import compile_prefix_regex PUNCT_OPEN = ["(", "[", "{", "*"] PUNCT_CLOSE = [")", "]", "}", "*"] PUNCT_PAIRED = [("(", ")"), ("[", "]"), ("{", "}"), ("*", "*")] @pytest.mark.parametrize("text", ["(", "((", "<"]) def test_en_tokenize...
4,422
33.554688
88
py
spaCy
spaCy-master/spacy/tests/lang/en/test_sbd.py
import pytest from spacy.tokens import Doc from ...util import apply_transition_sequence @pytest.mark.issue(309) def test_issue309(en_vocab): """Test Issue #309: SBD fails on empty string""" doc = Doc(en_vocab, words=[" "], heads=[0], deps=["ROOT"]) assert len(doc) == 1 sents = list(doc.sents) a...
1,708
35.361702
85
py
spaCy
spaCy-master/spacy/tests/lang/en/test_text.py
import pytest from spacy.lang.en.lex_attrs import like_num def test_en_tokenizer_handles_long_text(en_tokenizer): text = """Tributes pour in for late British Labour Party leader Tributes poured in from around the world Thursday to the late Labour Party leader John Smith, who died earlier from a massive heart at...
1,959
26.605634
77
py
spaCy
spaCy-master/spacy/tests/lang/en/test_tokenizer.py
import pytest @pytest.mark.issue(351) def test_issue351(en_tokenizer): doc = en_tokenizer(" This is a cat.") assert doc[0].idx == 0 assert len(doc[0]) == 3 assert doc[1].idx == 3 @pytest.mark.issue(360) def test_issue360(en_tokenizer): """Test tokenization of big ellipsis""" tokens = en_to...
5,924
32.100559
88
py
spaCy
spaCy-master/spacy/tests/lang/es/__init__.py
0
0
0
py
spaCy
spaCy-master/spacy/tests/lang/es/test_exception.py
import pytest @pytest.mark.parametrize( "text,lemma", [ ("aprox.", "aproximadamente"), ("esq.", "esquina"), ("pág.", "página"), ("p.ej.", "por ejemplo"), ], ) def test_es_tokenizer_handles_abbr(es_tokenizer, text, lemma): tokens = es_tokenizer(text) assert len(token...
546
22.782609
62
py
spaCy
spaCy-master/spacy/tests/lang/es/test_noun_chunks.py
import pytest from spacy.tokens import Doc # fmt: off @pytest.mark.parametrize( "words,heads,deps,pos,chunk_offsets", [ # un gato -> "un gato" ( ["un", "gato"], [1, 1], ["det", "ROOT"], ["DET", "NOUN"], [(0, 2)], ), #...
9,901
60.8875
452
py
spaCy
spaCy-master/spacy/tests/lang/es/test_text.py
import pytest from spacy.lang.es import Spanish from spacy.lang.es.lex_attrs import like_num @pytest.mark.issue(3803) def test_issue3803(): """Test that spanish num-like tokens have True for like_num attribute.""" nlp = Spanish() text = "2 dos 1000 mil 12 doce" doc = nlp(text) assert [t.like_num...
2,004
25.733333
83
py
spaCy
spaCy-master/spacy/tests/lang/et/__init__.py
0
0
0
py
spaCy
spaCy-master/spacy/tests/lang/et/test_text.py
import pytest def test_long_text(et_tokenizer): # Excerpt: European Convention on Human Rights text = """ arvestades, et nimetatud deklaratsiooni eesmärk on tagada selles kuulutatud õiguste üldine ja tõhus tunnustamine ning järgimine; arvestades, et Euroopa Nõukogu eesmärk on saavutada tema liikmete suurem üh...
909
32.703704
66
py
spaCy
spaCy-master/spacy/tests/lang/et/test_tokenizer.py
import pytest ET_BASIC_TOKENIZATION_TESTS = [ ( "Kedagi ei või piinata ega ebainimlikult või alandavalt kohelda " "ega karistada.", [ "Kedagi", "ei", "või", "piinata", "ega", "ebainimlikult", "või", ...
733
23.466667
77
py
spaCy
spaCy-master/spacy/tests/lang/eu/__init__.py
0
0
0
py