Datasets:
Add 'wnli' config data files
Browse files- README.md +34 -0
- dataset_infos.json +16 -30
- wnli/test-00000-of-00001.parquet +3 -0
- wnli/train-00000-of-00001.parquet +3 -0
- wnli/validation-00000-of-00001.parquet +3 -0
README.md
CHANGED
@@ -279,6 +279,32 @@ dataset_info:
|
|
279 |
num_examples: 1379
|
280 |
download_size: 761235
|
281 |
dataset_size: 1140829
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
282 |
configs:
|
283 |
- config_name: cola
|
284 |
data_files:
|
@@ -360,6 +386,14 @@ configs:
|
|
360 |
path: stsb/validation-*
|
361 |
- split: test
|
362 |
path: stsb/test-*
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
363 |
train-eval-index:
|
364 |
- config: cola
|
365 |
task: text-classification
|
|
|
279 |
num_examples: 1379
|
280 |
download_size: 761235
|
281 |
dataset_size: 1140829
|
282 |
+
- config_name: wnli
|
283 |
+
features:
|
284 |
+
- name: sentence1
|
285 |
+
dtype: string
|
286 |
+
- name: sentence2
|
287 |
+
dtype: string
|
288 |
+
- name: label
|
289 |
+
dtype:
|
290 |
+
class_label:
|
291 |
+
names:
|
292 |
+
'0': not_entailment
|
293 |
+
'1': entailment
|
294 |
+
- name: idx
|
295 |
+
dtype: int32
|
296 |
+
splits:
|
297 |
+
- name: train
|
298 |
+
num_bytes: 107109
|
299 |
+
num_examples: 635
|
300 |
+
- name: validation
|
301 |
+
num_bytes: 12162
|
302 |
+
num_examples: 71
|
303 |
+
- name: test
|
304 |
+
num_bytes: 37889
|
305 |
+
num_examples: 146
|
306 |
+
download_size: 61250
|
307 |
+
dataset_size: 157160
|
308 |
configs:
|
309 |
- config_name: cola
|
310 |
data_files:
|
|
|
386 |
path: stsb/validation-*
|
387 |
- split: test
|
388 |
path: stsb/test-*
|
389 |
+
- config_name: wnli
|
390 |
+
data_files:
|
391 |
+
- split: train
|
392 |
+
path: wnli/train-*
|
393 |
+
- split: validation
|
394 |
+
path: wnli/validation-*
|
395 |
+
- split: test
|
396 |
+
path: wnli/test-*
|
397 |
train-eval-index:
|
398 |
- config: cola
|
399 |
task: text-classification
|
dataset_infos.json
CHANGED
@@ -593,39 +593,32 @@
|
|
593 |
},
|
594 |
"wnli": {
|
595 |
"description": "GLUE, the General Language Understanding Evaluation benchmark\n(https://gluebenchmark.com/) is a collection of resources for training,\nevaluating, and analyzing natural language understanding systems.\n\n",
|
596 |
-
"citation": "@inproceedings{levesque2012winograd,\n title={The winograd schema challenge},\n author={Levesque, Hector and Davis, Ernest and Morgenstern, Leora},\n booktitle={Thirteenth International Conference on the Principles of Knowledge Representation and Reasoning},\n year={2012}\n}\n@inproceedings{wang2019glue,\n title={{GLUE}: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding},\n author={Wang, Alex and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R.},\n note={In the Proceedings of ICLR.},\n year={2019}\n}\n
|
597 |
"homepage": "https://cs.nyu.edu/faculty/davise/papers/WinogradSchemas/WS.html",
|
598 |
"license": "",
|
599 |
"features": {
|
600 |
"sentence1": {
|
601 |
"dtype": "string",
|
602 |
-
"id": null,
|
603 |
"_type": "Value"
|
604 |
},
|
605 |
"sentence2": {
|
606 |
"dtype": "string",
|
607 |
-
"id": null,
|
608 |
"_type": "Value"
|
609 |
},
|
610 |
"label": {
|
611 |
-
"num_classes": 2,
|
612 |
"names": [
|
613 |
"not_entailment",
|
614 |
"entailment"
|
615 |
],
|
616 |
-
"names_file": null,
|
617 |
-
"id": null,
|
618 |
"_type": "ClassLabel"
|
619 |
},
|
620 |
"idx": {
|
621 |
"dtype": "int32",
|
622 |
-
"id": null,
|
623 |
"_type": "Value"
|
624 |
}
|
625 |
},
|
626 |
-
"
|
627 |
-
"
|
628 |
-
"builder_name": "glue",
|
629 |
"config_name": "wnli",
|
630 |
"version": {
|
631 |
"version_str": "1.0.0",
|
@@ -635,35 +628,28 @@
|
|
635 |
"patch": 0
|
636 |
},
|
637 |
"splits": {
|
638 |
-
"test": {
|
639 |
-
"name": "test",
|
640 |
-
"num_bytes": 37992,
|
641 |
-
"num_examples": 146,
|
642 |
-
"dataset_name": "glue"
|
643 |
-
},
|
644 |
"train": {
|
645 |
"name": "train",
|
646 |
-
"num_bytes":
|
647 |
"num_examples": 635,
|
648 |
-
"dataset_name":
|
649 |
},
|
650 |
"validation": {
|
651 |
"name": "validation",
|
652 |
-
"num_bytes":
|
653 |
"num_examples": 71,
|
654 |
-
"dataset_name":
|
655 |
-
}
|
656 |
-
|
657 |
-
|
658 |
-
|
659 |
-
"
|
660 |
-
"
|
661 |
}
|
662 |
},
|
663 |
-
"download_size":
|
664 |
-
"
|
665 |
-
"
|
666 |
-
"size_in_bytes": 186723
|
667 |
},
|
668 |
"ax": {
|
669 |
"description": "GLUE, the General Language Understanding Evaluation benchmark\n(https://gluebenchmark.com/) is a collection of resources for training,\nevaluating, and analyzing natural language understanding systems.\n\n",
|
|
|
593 |
},
|
594 |
"wnli": {
|
595 |
"description": "GLUE, the General Language Understanding Evaluation benchmark\n(https://gluebenchmark.com/) is a collection of resources for training,\nevaluating, and analyzing natural language understanding systems.\n\n",
|
596 |
+
"citation": "@inproceedings{levesque2012winograd,\n title={The winograd schema challenge},\n author={Levesque, Hector and Davis, Ernest and Morgenstern, Leora},\n booktitle={Thirteenth International Conference on the Principles of Knowledge Representation and Reasoning},\n year={2012}\n}\n@inproceedings{wang2019glue,\n title={{GLUE}: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding},\n author={Wang, Alex and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R.},\n note={In the Proceedings of ICLR.},\n year={2019}\n}\n",
|
597 |
"homepage": "https://cs.nyu.edu/faculty/davise/papers/WinogradSchemas/WS.html",
|
598 |
"license": "",
|
599 |
"features": {
|
600 |
"sentence1": {
|
601 |
"dtype": "string",
|
|
|
602 |
"_type": "Value"
|
603 |
},
|
604 |
"sentence2": {
|
605 |
"dtype": "string",
|
|
|
606 |
"_type": "Value"
|
607 |
},
|
608 |
"label": {
|
|
|
609 |
"names": [
|
610 |
"not_entailment",
|
611 |
"entailment"
|
612 |
],
|
|
|
|
|
613 |
"_type": "ClassLabel"
|
614 |
},
|
615 |
"idx": {
|
616 |
"dtype": "int32",
|
|
|
617 |
"_type": "Value"
|
618 |
}
|
619 |
},
|
620 |
+
"builder_name": "glue-ci",
|
621 |
+
"dataset_name": "glue-ci",
|
|
|
622 |
"config_name": "wnli",
|
623 |
"version": {
|
624 |
"version_str": "1.0.0",
|
|
|
628 |
"patch": 0
|
629 |
},
|
630 |
"splits": {
|
|
|
|
|
|
|
|
|
|
|
|
|
631 |
"train": {
|
632 |
"name": "train",
|
633 |
+
"num_bytes": 107109,
|
634 |
"num_examples": 635,
|
635 |
+
"dataset_name": null
|
636 |
},
|
637 |
"validation": {
|
638 |
"name": "validation",
|
639 |
+
"num_bytes": 12162,
|
640 |
"num_examples": 71,
|
641 |
+
"dataset_name": null
|
642 |
+
},
|
643 |
+
"test": {
|
644 |
+
"name": "test",
|
645 |
+
"num_bytes": 37889,
|
646 |
+
"num_examples": 146,
|
647 |
+
"dataset_name": null
|
648 |
}
|
649 |
},
|
650 |
+
"download_size": 61250,
|
651 |
+
"dataset_size": 157160,
|
652 |
+
"size_in_bytes": 218410
|
|
|
653 |
},
|
654 |
"ax": {
|
655 |
"description": "GLUE, the General Language Understanding Evaluation benchmark\n(https://gluebenchmark.com/) is a collection of resources for training,\nevaluating, and analyzing natural language understanding systems.\n\n",
|
wnli/test-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:39f6607b5400cfc068ab613dd2ceadbdddf2c85aa56b8694d602e4612a170cab
|
3 |
+
size 12837
|
wnli/train-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:69539c43069eb3f052f8b86abe7ca2cd4f23645e152f044e7d19d1e107458588
|
3 |
+
size 38011
|
wnli/validation-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3945356789e3b1b488f5469beb230abd2c5d4b79ec3319048ce63ab405eb14eb
|
3 |
+
size 10402
|