Datasets:
Commit
·
5a54c9f
1
Parent(s):
9bbf8bc
Update parquet files
Browse files- .gitattributes +1 -0
- .github/workflows/ci.yaml +0 -47
- .github/workflows/push_to_hub.yaml +0 -26
- .gitignore +0 -171
- README.md +0 -33
- default/jsnli-train.parquet +3 -0
- default/jsnli-validation.parquet +0 -0
- jsnli.py +0 -111
- poetry.lock +0 -0
- pyproject.toml +0 -23
- tests/__init__.py +0 -0
- tests/jsnli_test.py +0 -35
.gitattributes
ADDED
@@ -0,0 +1 @@
|
|
|
|
|
1 |
+
default/jsnli-train.parquet filter=lfs diff=lfs merge=lfs -text
|
.github/workflows/ci.yaml
DELETED
@@ -1,47 +0,0 @@
|
|
1 |
-
name: CI
|
2 |
-
|
3 |
-
on:
|
4 |
-
push:
|
5 |
-
branches: [main]
|
6 |
-
paths-ignore:
|
7 |
-
- 'README.md'
|
8 |
-
pull_request:
|
9 |
-
branches: [main]
|
10 |
-
paths-ignore:
|
11 |
-
- 'README.md'
|
12 |
-
|
13 |
-
jobs:
|
14 |
-
test:
|
15 |
-
runs-on: ubuntu-latest
|
16 |
-
strategy:
|
17 |
-
matrix:
|
18 |
-
python-version: ['3.8', '3.9', '3.10']
|
19 |
-
|
20 |
-
steps:
|
21 |
-
- uses: actions/checkout@v2
|
22 |
-
- name: Set up Python ${{ matrix.python-version }}
|
23 |
-
uses: actions/setup-python@v2
|
24 |
-
with:
|
25 |
-
python-version: ${{ matrix.python-version }}
|
26 |
-
|
27 |
-
- name: Install dependencies
|
28 |
-
run: |
|
29 |
-
pip install -U pip setuptools wheel poetry
|
30 |
-
poetry install
|
31 |
-
- name: Format
|
32 |
-
run: |
|
33 |
-
poetry run black --check .
|
34 |
-
- name: Lint
|
35 |
-
run: |
|
36 |
-
poetry run flake8 . --ignore=E501,W503,E203
|
37 |
-
- name: Type check
|
38 |
-
run: |
|
39 |
-
poetry run mypy . \
|
40 |
-
--ignore-missing-imports \
|
41 |
-
--no-strict-optional \
|
42 |
-
--no-site-packages \
|
43 |
-
--cache-dir=/dev/null
|
44 |
-
|
45 |
-
- name: Run tests
|
46 |
-
run: |
|
47 |
-
poetry run pytest --color=yes -rf
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
.github/workflows/push_to_hub.yaml
DELETED
@@ -1,26 +0,0 @@
|
|
1 |
-
name: Sync to Hugging Face Hub
|
2 |
-
|
3 |
-
on:
|
4 |
-
workflow_run:
|
5 |
-
workflows:
|
6 |
-
- CI
|
7 |
-
branches:
|
8 |
-
- main
|
9 |
-
types:
|
10 |
-
- completed
|
11 |
-
|
12 |
-
jobs:
|
13 |
-
push_to_hub:
|
14 |
-
runs-on: ubuntu-latest
|
15 |
-
|
16 |
-
steps:
|
17 |
-
- name: Checkout repository
|
18 |
-
uses: actions/checkout@v2
|
19 |
-
|
20 |
-
- name: Push to Huggingface hub
|
21 |
-
env:
|
22 |
-
HF_TOKEN: ${{ secrets.HF_TOKEN }}
|
23 |
-
HF_USERNAME: ${{ secrets.HF_USERNAME }}
|
24 |
-
run: |
|
25 |
-
git fetch --unshallow
|
26 |
-
git push --force https://${HF_USERNAME}:${HF_TOKEN}@huggingface.co/datasets/${HF_USERNAME}/jsnli main
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
.gitignore
DELETED
@@ -1,171 +0,0 @@
|
|
1 |
-
# Created by https://www.toptal.com/developers/gitignore/api/python
|
2 |
-
# Edit at https://www.toptal.com/developers/gitignore?templates=python
|
3 |
-
|
4 |
-
### Python ###
|
5 |
-
# Byte-compiled / optimized / DLL files
|
6 |
-
__pycache__/
|
7 |
-
*.py[cod]
|
8 |
-
*$py.class
|
9 |
-
|
10 |
-
# C extensions
|
11 |
-
*.so
|
12 |
-
|
13 |
-
# Distribution / packaging
|
14 |
-
.Python
|
15 |
-
build/
|
16 |
-
develop-eggs/
|
17 |
-
dist/
|
18 |
-
downloads/
|
19 |
-
eggs/
|
20 |
-
.eggs/
|
21 |
-
lib/
|
22 |
-
lib64/
|
23 |
-
parts/
|
24 |
-
sdist/
|
25 |
-
var/
|
26 |
-
wheels/
|
27 |
-
share/python-wheels/
|
28 |
-
*.egg-info/
|
29 |
-
.installed.cfg
|
30 |
-
*.egg
|
31 |
-
MANIFEST
|
32 |
-
|
33 |
-
# PyInstaller
|
34 |
-
# Usually these files are written by a python script from a template
|
35 |
-
# before PyInstaller builds the exe, so as to inject date/other infos into it.
|
36 |
-
*.manifest
|
37 |
-
*.spec
|
38 |
-
|
39 |
-
# Installer logs
|
40 |
-
pip-log.txt
|
41 |
-
pip-delete-this-directory.txt
|
42 |
-
|
43 |
-
# Unit test / coverage reports
|
44 |
-
htmlcov/
|
45 |
-
.tox/
|
46 |
-
.nox/
|
47 |
-
.coverage
|
48 |
-
.coverage.*
|
49 |
-
.cache
|
50 |
-
nosetests.xml
|
51 |
-
coverage.xml
|
52 |
-
*.cover
|
53 |
-
*.py,cover
|
54 |
-
.hypothesis/
|
55 |
-
.pytest_cache/
|
56 |
-
cover/
|
57 |
-
|
58 |
-
# Translations
|
59 |
-
*.mo
|
60 |
-
*.pot
|
61 |
-
|
62 |
-
# Django stuff:
|
63 |
-
*.log
|
64 |
-
local_settings.py
|
65 |
-
db.sqlite3
|
66 |
-
db.sqlite3-journal
|
67 |
-
|
68 |
-
# Flask stuff:
|
69 |
-
instance/
|
70 |
-
.webassets-cache
|
71 |
-
|
72 |
-
# Scrapy stuff:
|
73 |
-
.scrapy
|
74 |
-
|
75 |
-
# Sphinx documentation
|
76 |
-
docs/_build/
|
77 |
-
|
78 |
-
# PyBuilder
|
79 |
-
.pybuilder/
|
80 |
-
target/
|
81 |
-
|
82 |
-
# Jupyter Notebook
|
83 |
-
.ipynb_checkpoints
|
84 |
-
|
85 |
-
# IPython
|
86 |
-
profile_default/
|
87 |
-
ipython_config.py
|
88 |
-
|
89 |
-
# pyenv
|
90 |
-
# For a library or package, you might want to ignore these files since the code is
|
91 |
-
# intended to run in multiple environments; otherwise, check them in:
|
92 |
-
.python-version
|
93 |
-
|
94 |
-
# pipenv
|
95 |
-
# According to pypa/pipenv#598, it is recommended to include Pipfile.lock in version control.
|
96 |
-
# However, in case of collaboration, if having platform-specific dependencies or dependencies
|
97 |
-
# having no cross-platform support, pipenv may install dependencies that don't work, or not
|
98 |
-
# install all needed dependencies.
|
99 |
-
#Pipfile.lock
|
100 |
-
|
101 |
-
# poetry
|
102 |
-
# Similar to Pipfile.lock, it is generally recommended to include poetry.lock in version control.
|
103 |
-
# This is especially recommended for binary packages to ensure reproducibility, and is more
|
104 |
-
# commonly ignored for libraries.
|
105 |
-
# https://python-poetry.org/docs/basic-usage/#commit-your-poetrylock-file-to-version-control
|
106 |
-
#poetry.lock
|
107 |
-
|
108 |
-
# pdm
|
109 |
-
# Similar to Pipfile.lock, it is generally recommended to include pdm.lock in version control.
|
110 |
-
#pdm.lock
|
111 |
-
# pdm stores project-wide configurations in .pdm.toml, but it is recommended to not include it
|
112 |
-
# in version control.
|
113 |
-
# https://pdm.fming.dev/#use-with-ide
|
114 |
-
.pdm.toml
|
115 |
-
|
116 |
-
# PEP 582; used by e.g. github.com/David-OConnor/pyflow and github.com/pdm-project/pdm
|
117 |
-
__pypackages__/
|
118 |
-
|
119 |
-
# Celery stuff
|
120 |
-
celerybeat-schedule
|
121 |
-
celerybeat.pid
|
122 |
-
|
123 |
-
# SageMath parsed files
|
124 |
-
*.sage.py
|
125 |
-
|
126 |
-
# Environments
|
127 |
-
.env
|
128 |
-
.venv
|
129 |
-
env/
|
130 |
-
venv/
|
131 |
-
ENV/
|
132 |
-
env.bak/
|
133 |
-
venv.bak/
|
134 |
-
|
135 |
-
# Spyder project settings
|
136 |
-
.spyderproject
|
137 |
-
.spyproject
|
138 |
-
|
139 |
-
# Rope project settings
|
140 |
-
.ropeproject
|
141 |
-
|
142 |
-
# mkdocs documentation
|
143 |
-
/site
|
144 |
-
|
145 |
-
# mypy
|
146 |
-
.mypy_cache/
|
147 |
-
.dmypy.json
|
148 |
-
dmypy.json
|
149 |
-
|
150 |
-
# Pyre type checker
|
151 |
-
.pyre/
|
152 |
-
|
153 |
-
# pytype static type analyzer
|
154 |
-
.pytype/
|
155 |
-
|
156 |
-
# Cython debug symbols
|
157 |
-
cython_debug/
|
158 |
-
|
159 |
-
# PyCharm
|
160 |
-
# JetBrains specific template is maintained in a separate JetBrains.gitignore that can
|
161 |
-
# be found at https://github.com/github/gitignore/blob/main/Global/JetBrains.gitignore
|
162 |
-
# and can be added to the global gitignore or merged into this file. For a more nuclear
|
163 |
-
# option (not recommended) you can uncomment the following to ignore the entire idea folder.
|
164 |
-
#.idea/
|
165 |
-
|
166 |
-
### Python Patch ###
|
167 |
-
# Poetry local configuration file - https://python-poetry.org/docs/configuration/#local-configuration
|
168 |
-
poetry.toml
|
169 |
-
|
170 |
-
|
171 |
-
# End of https://www.toptal.com/developers/gitignore/api/python
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
README.md
DELETED
@@ -1,33 +0,0 @@
|
|
1 |
-
---
|
2 |
-
language:
|
3 |
-
- ja
|
4 |
-
|
5 |
-
license:
|
6 |
-
- cc-by-sa-4.0
|
7 |
-
|
8 |
-
multilinguality:
|
9 |
-
- monolingual
|
10 |
-
|
11 |
-
task_categories:
|
12 |
-
- text-classification
|
13 |
-
|
14 |
-
task_ids:
|
15 |
-
- natural-language-inference
|
16 |
-
- multi-input-text-classification
|
17 |
-
|
18 |
-
tags:
|
19 |
-
- natural-language-inference
|
20 |
-
- nli
|
21 |
-
- jsnli
|
22 |
-
|
23 |
-
datasets:
|
24 |
-
- without-filtering
|
25 |
-
- with-filtering
|
26 |
-
|
27 |
-
metrics:
|
28 |
-
- accuracy
|
29 |
-
---
|
30 |
-
|
31 |
-
# Dataset Card for JSNLI
|
32 |
-
|
33 |
-
[](https://github.com/shunk031/huggingface-datasets_jsnli/actions/workflows/ci.yaml)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
default/jsnli-train.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:090e9a46e99de1f43a48e0a139b1dba6be1f74a401ada985744d3980ed74ec73
|
3 |
+
size 44382398
|
default/jsnli-validation.parquet
ADDED
Binary file (300 kB). View file
|
|
jsnli.py
DELETED
@@ -1,111 +0,0 @@
|
|
1 |
-
import logging
|
2 |
-
import os
|
3 |
-
|
4 |
-
import datasets as ds
|
5 |
-
|
6 |
-
logger = logging.getLogger(__name__)
|
7 |
-
|
8 |
-
_CITATION = """\
|
9 |
-
- 吉越 卓見, 河原 大輔, 黒橋 禎夫: 機械翻訳を用いた自然言語推論データセットの多言語化, 第244回自然言語処理研究会, (2020.7.3).
|
10 |
-
- Samuel R. Bowman, Gabor Angeli, Christopher Potts, and Christopher D. Manning. 2015. A large annotated corpus for learning natural language inference. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing (EMNLP).
|
11 |
-
- Peter Young, Alice Lai, Micah Hodosh, and Julia Hockenmaier. "From image descriptions to visual denotations: New similarity metrics for semantic inference over event descriptions." Transactions of the Association for Computational Linguistics 2 (2014): 67-78.
|
12 |
-
"""
|
13 |
-
|
14 |
-
_DESCRIPTION = """\
|
15 |
-
== 日本語SNLI(JSNLI)データセット ==
|
16 |
-
|
17 |
-
SNLI コーパスを日本語に翻訳した自然言語推論データセット
|
18 |
-
学習データは元データを翻訳し、計算機によるフィルタリングによって作成
|
19 |
-
評価データは日本語として意味が通るか、翻訳後のラベルが元のラベルと一致しているかどうかの2段階のクラウドソーシングによりデータをフィルタリング
|
20 |
-
"""
|
21 |
-
|
22 |
-
_HOMEPAGE = "https://nlp.ist.i.kyoto-u.ac.jp/?%E6%97%A5%E6%9C%AC%E8%AA%9ESNLI%28JSNLI%29%E3%83%87%E3%83%BC%E3%82%BF%E3%82%BB%E3%83%83%E3%83%88"
|
23 |
-
|
24 |
-
_LICENSE = """\
|
25 |
-
CC BY-SA 4.0
|
26 |
-
"""
|
27 |
-
|
28 |
-
_URL = "https://nlp.ist.i.kyoto-u.ac.jp/DLcounter/lime.cgi?down=https://nlp.ist.i.kyoto-u.ac.jp/nl-resource/JSNLI/jsnli_1.1.zip&name=JSNLI.zip"
|
29 |
-
|
30 |
-
|
31 |
-
class JSNLIDataset(ds.GeneratorBasedBuilder):
|
32 |
-
VERSION = ds.Version("1.1.0") # type: ignore
|
33 |
-
DEFAULT_CONFIG_NAME: str = "with-filtering" # type: ignore
|
34 |
-
BUILDER_CONFIG = [
|
35 |
-
ds.BuilderConfig(
|
36 |
-
name="with-filtering",
|
37 |
-
version=VERSION, # type: ignore
|
38 |
-
description="SNLIの学習データに機械翻訳を適用した後、BLEUスコアの閾値0.1でフィルタリングを施したもの。BERTにこの学習データを学習させることにより、93.0%の精度を記録した。(533,005ペア)",
|
39 |
-
),
|
40 |
-
ds.BuilderConfig(
|
41 |
-
name="without-filtering",
|
42 |
-
version=VERSION, # type: ignore
|
43 |
-
description="SNLIの学習データに機械翻訳を適用したもの。フィルタリングは行っていない。(548,014ペア)",
|
44 |
-
),
|
45 |
-
]
|
46 |
-
|
47 |
-
def _info(self) -> ds.DatasetInfo:
|
48 |
-
features = ds.Features(
|
49 |
-
{
|
50 |
-
"label": ds.Value("string"),
|
51 |
-
"premise": ds.Value("string"),
|
52 |
-
"hypothesis": ds.Value("string"),
|
53 |
-
}
|
54 |
-
)
|
55 |
-
return ds.DatasetInfo(
|
56 |
-
description=_DESCRIPTION,
|
57 |
-
features=features,
|
58 |
-
homepage=_HOMEPAGE,
|
59 |
-
license=_LICENSE,
|
60 |
-
citation=_CITATION,
|
61 |
-
)
|
62 |
-
|
63 |
-
def _split_generators(self, dl_manager: ds.DownloadManager):
|
64 |
-
jsnli_base_dir = dl_manager.download_and_extract(_URL)
|
65 |
-
jsnli_dir = os.path.join(
|
66 |
-
jsnli_base_dir, f"jsnli_{self.VERSION.major}.{self.VERSION.minor}" # type: ignore
|
67 |
-
)
|
68 |
-
|
69 |
-
train_w_filtering_path = os.path.join(jsnli_dir, "train_w_filtering.tsv")
|
70 |
-
train_wo_filtering_path = os.path.join(jsnli_dir, "train_wo_filtering.tsv")
|
71 |
-
|
72 |
-
dev_path = os.path.join(jsnli_dir, "dev.tsv")
|
73 |
-
|
74 |
-
if "without-filtering" in self.config.name:
|
75 |
-
tng_path = train_wo_filtering_path
|
76 |
-
else:
|
77 |
-
# use `with-filtering` dataset as default setting
|
78 |
-
tng_path = train_w_filtering_path
|
79 |
-
|
80 |
-
tng_gen_kwargs = {
|
81 |
-
"tsv_path": tng_path,
|
82 |
-
}
|
83 |
-
val_gen_kwargs = {
|
84 |
-
"tsv_path": dev_path,
|
85 |
-
}
|
86 |
-
|
87 |
-
return [
|
88 |
-
ds.SplitGenerator(
|
89 |
-
name=ds.Split.TRAIN, # type: ignore
|
90 |
-
gen_kwargs=tng_gen_kwargs, # type: ignore
|
91 |
-
),
|
92 |
-
ds.SplitGenerator(
|
93 |
-
name=ds.Split.VALIDATION, # type: ignore
|
94 |
-
gen_kwargs=val_gen_kwargs, # type: ignore
|
95 |
-
),
|
96 |
-
]
|
97 |
-
|
98 |
-
def _generate_examples( # type: ignore
|
99 |
-
self,
|
100 |
-
tsv_path: str,
|
101 |
-
):
|
102 |
-
with open(tsv_path, "r") as rf:
|
103 |
-
for sentence_id, line in enumerate(rf):
|
104 |
-
label, premise, hypothesis = line.replace("\n", "").split("\t")
|
105 |
-
|
106 |
-
example_dict = {
|
107 |
-
"label": label,
|
108 |
-
"premise": premise,
|
109 |
-
"hypothesis": hypothesis,
|
110 |
-
}
|
111 |
-
yield sentence_id, example_dict
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
poetry.lock
DELETED
The diff for this file is too large to render.
See raw diff
|
|
pyproject.toml
DELETED
@@ -1,23 +0,0 @@
|
|
1 |
-
[tool.poetry]
|
2 |
-
name = "huggingface-datasets-jsnli"
|
3 |
-
version = "0.1.0"
|
4 |
-
description = ""
|
5 |
-
authors = ["Shunsuke KITADA <[email protected]>"]
|
6 |
-
readme = "README.md"
|
7 |
-
packages = []
|
8 |
-
|
9 |
-
[tool.poetry.dependencies]
|
10 |
-
python = ">=3.8.1,<4.0"
|
11 |
-
datasets = "^2.7.1"
|
12 |
-
|
13 |
-
|
14 |
-
[tool.poetry.group.dev.dependencies]
|
15 |
-
black = "^22.10.0"
|
16 |
-
isort = "^5.10.1"
|
17 |
-
flake8 = "^6.0.0"
|
18 |
-
mypy = "^0.991"
|
19 |
-
pytest = "^7.2.0"
|
20 |
-
|
21 |
-
[build-system]
|
22 |
-
requires = ["poetry-core"]
|
23 |
-
build-backend = "poetry.core.masonry.api"
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
tests/__init__.py
DELETED
File without changes
|
tests/jsnli_test.py
DELETED
@@ -1,35 +0,0 @@
|
|
1 |
-
import datasets as ds
|
2 |
-
import pytest
|
3 |
-
|
4 |
-
|
5 |
-
@pytest.fixture
|
6 |
-
def dataset_path() -> str:
|
7 |
-
return "jsnli.py"
|
8 |
-
|
9 |
-
|
10 |
-
@pytest.mark.parametrize(
|
11 |
-
"dataset_name, expected_num_train,",
|
12 |
-
(
|
13 |
-
("without-filtering", 548014),
|
14 |
-
("with-filtering", 533005),
|
15 |
-
),
|
16 |
-
)
|
17 |
-
def test_load_dataset(
|
18 |
-
dataset_path: str,
|
19 |
-
dataset_name: str,
|
20 |
-
expected_num_train: int,
|
21 |
-
expected_num_valid: int = 3916,
|
22 |
-
):
|
23 |
-
dataset = ds.load_dataset(path=dataset_path, name=dataset_name)
|
24 |
-
|
25 |
-
assert dataset["train"].num_rows == expected_num_train # type: ignore
|
26 |
-
assert dataset["validation"].num_rows == expected_num_valid # type: ignore
|
27 |
-
|
28 |
-
|
29 |
-
def test_load_dataset_default(
|
30 |
-
dataset_path: str, expected_num_train: int = 533005, expected_num_valid: int = 3916
|
31 |
-
):
|
32 |
-
|
33 |
-
dataset = ds.load_dataset(path=dataset_path)
|
34 |
-
assert dataset["train"].num_rows == expected_num_train # type: ignore
|
35 |
-
assert dataset["validation"].num_rows == expected_num_valid # type: ignore
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|