lhoestq HF Staff commited on
Commit
545ce75
·
verified ·
1 Parent(s): c613e45

Add 'wnli' config data files

Browse files
README.md CHANGED
@@ -279,6 +279,32 @@ dataset_info:
279
  num_examples: 1379
280
  download_size: 761235
281
  dataset_size: 1140829
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
282
  configs:
283
  - config_name: cola
284
  data_files:
@@ -360,6 +386,14 @@ configs:
360
  path: stsb/validation-*
361
  - split: test
362
  path: stsb/test-*
 
 
 
 
 
 
 
 
363
  train-eval-index:
364
  - config: cola
365
  task: text-classification
 
279
  num_examples: 1379
280
  download_size: 761235
281
  dataset_size: 1140829
282
+ - config_name: wnli
283
+ features:
284
+ - name: sentence1
285
+ dtype: string
286
+ - name: sentence2
287
+ dtype: string
288
+ - name: label
289
+ dtype:
290
+ class_label:
291
+ names:
292
+ '0': not_entailment
293
+ '1': entailment
294
+ - name: idx
295
+ dtype: int32
296
+ splits:
297
+ - name: train
298
+ num_bytes: 107109
299
+ num_examples: 635
300
+ - name: validation
301
+ num_bytes: 12162
302
+ num_examples: 71
303
+ - name: test
304
+ num_bytes: 37889
305
+ num_examples: 146
306
+ download_size: 61250
307
+ dataset_size: 157160
308
  configs:
309
  - config_name: cola
310
  data_files:
 
386
  path: stsb/validation-*
387
  - split: test
388
  path: stsb/test-*
389
+ - config_name: wnli
390
+ data_files:
391
+ - split: train
392
+ path: wnli/train-*
393
+ - split: validation
394
+ path: wnli/validation-*
395
+ - split: test
396
+ path: wnli/test-*
397
  train-eval-index:
398
  - config: cola
399
  task: text-classification
dataset_infos.json CHANGED
@@ -593,39 +593,32 @@
593
  },
594
  "wnli": {
595
  "description": "GLUE, the General Language Understanding Evaluation benchmark\n(https://gluebenchmark.com/) is a collection of resources for training,\nevaluating, and analyzing natural language understanding systems.\n\n",
596
- "citation": "@inproceedings{levesque2012winograd,\n title={The winograd schema challenge},\n author={Levesque, Hector and Davis, Ernest and Morgenstern, Leora},\n booktitle={Thirteenth International Conference on the Principles of Knowledge Representation and Reasoning},\n year={2012}\n}\n@inproceedings{wang2019glue,\n title={{GLUE}: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding},\n author={Wang, Alex and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R.},\n note={In the Proceedings of ICLR.},\n year={2019}\n}\n\nNote that each GLUE dataset has its own citation. Please see the source to see\nthe correct citation for each contained dataset.",
597
  "homepage": "https://cs.nyu.edu/faculty/davise/papers/WinogradSchemas/WS.html",
598
  "license": "",
599
  "features": {
600
  "sentence1": {
601
  "dtype": "string",
602
- "id": null,
603
  "_type": "Value"
604
  },
605
  "sentence2": {
606
  "dtype": "string",
607
- "id": null,
608
  "_type": "Value"
609
  },
610
  "label": {
611
- "num_classes": 2,
612
  "names": [
613
  "not_entailment",
614
  "entailment"
615
  ],
616
- "names_file": null,
617
- "id": null,
618
  "_type": "ClassLabel"
619
  },
620
  "idx": {
621
  "dtype": "int32",
622
- "id": null,
623
  "_type": "Value"
624
  }
625
  },
626
- "post_processed": null,
627
- "supervised_keys": null,
628
- "builder_name": "glue",
629
  "config_name": "wnli",
630
  "version": {
631
  "version_str": "1.0.0",
@@ -635,35 +628,28 @@
635
  "patch": 0
636
  },
637
  "splits": {
638
- "test": {
639
- "name": "test",
640
- "num_bytes": 37992,
641
- "num_examples": 146,
642
- "dataset_name": "glue"
643
- },
644
  "train": {
645
  "name": "train",
646
- "num_bytes": 107517,
647
  "num_examples": 635,
648
- "dataset_name": "glue"
649
  },
650
  "validation": {
651
  "name": "validation",
652
- "num_bytes": 12215,
653
  "num_examples": 71,
654
- "dataset_name": "glue"
655
- }
656
- },
657
- "download_checksums": {
658
- "https://dl.fbaipublicfiles.com/glue/data/WNLI.zip": {
659
- "num_bytes": 28999,
660
- "checksum": "ae0e8e4d16f4d46d4a0a566ec7ecceccfd3fbfaa4a7a4b4e02848c0f2561ac46"
661
  }
662
  },
663
- "download_size": 28999,
664
- "post_processing_size": null,
665
- "dataset_size": 157724,
666
- "size_in_bytes": 186723
667
  },
668
  "ax": {
669
  "description": "GLUE, the General Language Understanding Evaluation benchmark\n(https://gluebenchmark.com/) is a collection of resources for training,\nevaluating, and analyzing natural language understanding systems.\n\n",
 
593
  },
594
  "wnli": {
595
  "description": "GLUE, the General Language Understanding Evaluation benchmark\n(https://gluebenchmark.com/) is a collection of resources for training,\nevaluating, and analyzing natural language understanding systems.\n\n",
596
+ "citation": "@inproceedings{levesque2012winograd,\n title={The winograd schema challenge},\n author={Levesque, Hector and Davis, Ernest and Morgenstern, Leora},\n booktitle={Thirteenth International Conference on the Principles of Knowledge Representation and Reasoning},\n year={2012}\n}\n@inproceedings{wang2019glue,\n title={{GLUE}: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding},\n author={Wang, Alex and Singh, Amanpreet and Michael, Julian and Hill, Felix and Levy, Omer and Bowman, Samuel R.},\n note={In the Proceedings of ICLR.},\n year={2019}\n}\n",
597
  "homepage": "https://cs.nyu.edu/faculty/davise/papers/WinogradSchemas/WS.html",
598
  "license": "",
599
  "features": {
600
  "sentence1": {
601
  "dtype": "string",
 
602
  "_type": "Value"
603
  },
604
  "sentence2": {
605
  "dtype": "string",
 
606
  "_type": "Value"
607
  },
608
  "label": {
 
609
  "names": [
610
  "not_entailment",
611
  "entailment"
612
  ],
 
 
613
  "_type": "ClassLabel"
614
  },
615
  "idx": {
616
  "dtype": "int32",
 
617
  "_type": "Value"
618
  }
619
  },
620
+ "builder_name": "glue-ci",
621
+ "dataset_name": "glue-ci",
 
622
  "config_name": "wnli",
623
  "version": {
624
  "version_str": "1.0.0",
 
628
  "patch": 0
629
  },
630
  "splits": {
 
 
 
 
 
 
631
  "train": {
632
  "name": "train",
633
+ "num_bytes": 107109,
634
  "num_examples": 635,
635
+ "dataset_name": null
636
  },
637
  "validation": {
638
  "name": "validation",
639
+ "num_bytes": 12162,
640
  "num_examples": 71,
641
+ "dataset_name": null
642
+ },
643
+ "test": {
644
+ "name": "test",
645
+ "num_bytes": 37889,
646
+ "num_examples": 146,
647
+ "dataset_name": null
648
  }
649
  },
650
+ "download_size": 61250,
651
+ "dataset_size": 157160,
652
+ "size_in_bytes": 218410
 
653
  },
654
  "ax": {
655
  "description": "GLUE, the General Language Understanding Evaluation benchmark\n(https://gluebenchmark.com/) is a collection of resources for training,\nevaluating, and analyzing natural language understanding systems.\n\n",
wnli/test-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:39f6607b5400cfc068ab613dd2ceadbdddf2c85aa56b8694d602e4612a170cab
3
+ size 12837
wnli/train-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:69539c43069eb3f052f8b86abe7ca2cd4f23645e152f044e7d19d1e107458588
3
+ size 38011
wnli/validation-00000-of-00001.parquet ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:3945356789e3b1b488f5469beb230abd2c5d4b79ec3319048ce63ab405eb14eb
3
+ size 10402