Update spaCy pipeline
Browse files
__pycache__/use_custom_tokenizer.cpython-312.pyc
CHANGED
Binary files a/__pycache__/use_custom_tokenizer.cpython-312.pyc and b/__pycache__/use_custom_tokenizer.cpython-312.pyc differ
|
|
de_trf_nrp-any-py3-none-any.whl
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:2eb7cfa5c314f82441bf8e915adb97edfe6e9d8238ea2a8a66527222e6983170
|
3 |
+
size 865702369
|
use_custom_tokenizer.py
CHANGED
@@ -1,12 +1,13 @@
|
|
1 |
from spacy.util import registry
|
2 |
-
|
3 |
-
import
|
|
|
|
|
|
|
4 |
|
5 |
@registry.tokenizers("customize_tokenizer")
|
6 |
def make_customize_tokenizer():
|
7 |
def customize_tokenizer(nlp):
|
8 |
-
|
9 |
-
script_dir = pathlib.Path(__file__).parent.resolve()
|
10 |
-
return tokenizer.from_disk(script_dir / "tokenizer")
|
11 |
|
12 |
return customize_tokenizer
|
|
|
1 |
from spacy.util import registry
|
2 |
+
|
3 |
+
from commercial_registry_ner.spacy.custom_tokenizer.custom_tokenizer import (
|
4 |
+
custom_tokenizer,
|
5 |
+
)
|
6 |
+
|
7 |
|
8 |
@registry.tokenizers("customize_tokenizer")
|
9 |
def make_customize_tokenizer():
|
10 |
def customize_tokenizer(nlp):
|
11 |
+
return custom_tokenizer(nlp)
|
|
|
|
|
12 |
|
13 |
return customize_tokenizer
|