Lazy-Val commited on
Commit
280e963
·
verified ·
1 Parent(s): 1501920

Update spaCy pipeline

Browse files
__pycache__/use_custom_tokenizer.cpython-312.pyc CHANGED
Binary files a/__pycache__/use_custom_tokenizer.cpython-312.pyc and b/__pycache__/use_custom_tokenizer.cpython-312.pyc differ
 
de_trf_nrp-any-py3-none-any.whl CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f3ceaf71530479286554f5f67226794741a6a306a84011daa39a2fa9c87709a2
3
- size 865702938
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:2eb7cfa5c314f82441bf8e915adb97edfe6e9d8238ea2a8a66527222e6983170
3
+ size 865702369
use_custom_tokenizer.py CHANGED
@@ -1,12 +1,13 @@
1
  from spacy.util import registry
2
- from spacy.tokenizer import Tokenizer
3
- import pathlib
 
 
 
4
 
5
  @registry.tokenizers("customize_tokenizer")
6
  def make_customize_tokenizer():
7
  def customize_tokenizer(nlp):
8
- tokenizer = Tokenizer(nlp.vocab)
9
- script_dir = pathlib.Path(__file__).parent.resolve()
10
- return tokenizer.from_disk(script_dir / "tokenizer")
11
 
12
  return customize_tokenizer
 
1
  from spacy.util import registry
2
+
3
+ from commercial_registry_ner.spacy.custom_tokenizer.custom_tokenizer import (
4
+ custom_tokenizer,
5
+ )
6
+
7
 
8
  @registry.tokenizers("customize_tokenizer")
9
  def make_customize_tokenizer():
10
  def customize_tokenizer(nlp):
11
+ return custom_tokenizer(nlp)
 
 
12
 
13
  return customize_tokenizer