Datasets:
Add 'ax' config data files
Browse files- README.md +25 -0
- ax/test-00000-of-00001.parquet +3 -0
- dataset_infos.json +8 -22
README.md
CHANGED
@@ -24,6 +24,27 @@ task_ids:
|
|
24 |
paperswithcode_id: glue
|
25 |
pretty_name: GLUE (General Language Understanding Evaluation benchmark)
|
26 |
dataset_info:
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
27 |
- config_name: cola
|
28 |
features:
|
29 |
- name: sentence
|
@@ -306,6 +327,10 @@ dataset_info:
|
|
306 |
download_size: 61250
|
307 |
dataset_size: 157160
|
308 |
configs:
|
|
|
|
|
|
|
|
|
309 |
- config_name: cola
|
310 |
data_files:
|
311 |
- split: train
|
|
|
24 |
paperswithcode_id: glue
|
25 |
pretty_name: GLUE (General Language Understanding Evaluation benchmark)
|
26 |
dataset_info:
|
27 |
+
- config_name: ax
|
28 |
+
features:
|
29 |
+
- name: premise
|
30 |
+
dtype: string
|
31 |
+
- name: hypothesis
|
32 |
+
dtype: string
|
33 |
+
- name: label
|
34 |
+
dtype:
|
35 |
+
class_label:
|
36 |
+
names:
|
37 |
+
'0': entailment
|
38 |
+
'1': neutral
|
39 |
+
'2': contradiction
|
40 |
+
- name: idx
|
41 |
+
dtype: int32
|
42 |
+
splits:
|
43 |
+
- name: test
|
44 |
+
num_bytes: 237694
|
45 |
+
num_examples: 1104
|
46 |
+
download_size: 79191
|
47 |
+
dataset_size: 237694
|
48 |
- config_name: cola
|
49 |
features:
|
50 |
- name: sentence
|
|
|
327 |
download_size: 61250
|
328 |
dataset_size: 157160
|
329 |
configs:
|
330 |
+
- config_name: ax
|
331 |
+
data_files:
|
332 |
+
- split: test
|
333 |
+
path: ax/test-*
|
334 |
- config_name: cola
|
335 |
data_files:
|
336 |
- split: train
|
ax/test-00000-of-00001.parquet
ADDED
@@ -0,0 +1,3 @@
|
|
|
|
|
|
|
|
|
1 |
+
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:3a47cbc39a3617a4121f9c30d5f0c77a0f37e96427dbaef01647cb0c4ff92638
|
3 |
+
size 79191
|
dataset_infos.json
CHANGED
@@ -653,40 +653,33 @@
|
|
653 |
},
|
654 |
"ax": {
|
655 |
"description": "GLUE, the General Language Understanding Evaluation benchmark\n(https://gluebenchmark.com/) is a collection of resources for training,\nevaluating, and analyzing natural language understanding systems.\n\n",
|
656 |
-
"citation": "\n@inproceedings{wang2019glue,\n title={{GLUE}: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding},\n author={Wang, Alex and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R.},\n note={In the Proceedings of ICLR.},\n year={2019}\n}\n
|
657 |
"homepage": "https://gluebenchmark.com/diagnostics",
|
658 |
"license": "",
|
659 |
"features": {
|
660 |
"premise": {
|
661 |
"dtype": "string",
|
662 |
-
"id": null,
|
663 |
"_type": "Value"
|
664 |
},
|
665 |
"hypothesis": {
|
666 |
"dtype": "string",
|
667 |
-
"id": null,
|
668 |
"_type": "Value"
|
669 |
},
|
670 |
"label": {
|
671 |
-
"num_classes": 3,
|
672 |
"names": [
|
673 |
"entailment",
|
674 |
"neutral",
|
675 |
"contradiction"
|
676 |
],
|
677 |
-
"names_file": null,
|
678 |
-
"id": null,
|
679 |
"_type": "ClassLabel"
|
680 |
},
|
681 |
"idx": {
|
682 |
"dtype": "int32",
|
683 |
-
"id": null,
|
684 |
"_type": "Value"
|
685 |
}
|
686 |
},
|
687 |
-
"
|
688 |
-
"
|
689 |
-
"builder_name": "glue",
|
690 |
"config_name": "ax",
|
691 |
"version": {
|
692 |
"version_str": "1.0.0",
|
@@ -698,20 +691,13 @@
|
|
698 |
"splits": {
|
699 |
"test": {
|
700 |
"name": "test",
|
701 |
-
"num_bytes":
|
702 |
"num_examples": 1104,
|
703 |
-
"dataset_name":
|
704 |
-
}
|
705 |
-
},
|
706 |
-
"download_checksums": {
|
707 |
-
"https://dl.fbaipublicfiles.com/glue/data/AX.tsv": {
|
708 |
-
"num_bytes": 222257,
|
709 |
-
"checksum": "0e13510b1bb14436ff7e2ee82338f0efb0133ecf2e73507a697dc210db3f05fd"
|
710 |
}
|
711 |
},
|
712 |
-
"download_size":
|
713 |
-
"
|
714 |
-
"
|
715 |
-
"size_in_bytes": 460649
|
716 |
}
|
717 |
}
|
|
|
653 |
},
|
654 |
"ax": {
|
655 |
"description": "GLUE, the General Language Understanding Evaluation benchmark\n(https://gluebenchmark.com/) is a collection of resources for training,\nevaluating, and analyzing natural language understanding systems.\n\n",
|
656 |
+
"citation": "\n@inproceedings{wang2019glue,\n title={{GLUE}: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding},\n author={Wang, Alex and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R.},\n note={In the Proceedings of ICLR.},\n year={2019}\n}\n",
|
657 |
"homepage": "https://gluebenchmark.com/diagnostics",
|
658 |
"license": "",
|
659 |
"features": {
|
660 |
"premise": {
|
661 |
"dtype": "string",
|
|
|
662 |
"_type": "Value"
|
663 |
},
|
664 |
"hypothesis": {
|
665 |
"dtype": "string",
|
|
|
666 |
"_type": "Value"
|
667 |
},
|
668 |
"label": {
|
|
|
669 |
"names": [
|
670 |
"entailment",
|
671 |
"neutral",
|
672 |
"contradiction"
|
673 |
],
|
|
|
|
|
674 |
"_type": "ClassLabel"
|
675 |
},
|
676 |
"idx": {
|
677 |
"dtype": "int32",
|
|
|
678 |
"_type": "Value"
|
679 |
}
|
680 |
},
|
681 |
+
"builder_name": "glue-ci",
|
682 |
+
"dataset_name": "glue-ci",
|
|
|
683 |
"config_name": "ax",
|
684 |
"version": {
|
685 |
"version_str": "1.0.0",
|
|
|
691 |
"splits": {
|
692 |
"test": {
|
693 |
"name": "test",
|
694 |
+
"num_bytes": 237694,
|
695 |
"num_examples": 1104,
|
696 |
+
"dataset_name": null
|
|
|
|
|
|
|
|
|
|
|
|
|
697 |
}
|
698 |
},
|
699 |
+
"download_size": 79191,
|
700 |
+
"dataset_size": 237694,
|
701 |
+
"size_in_bytes": 316885
|
|
|
702 |
}
|
703 |
}
|