Datasets:
Update README.md
Browse files
README.md
CHANGED
@@ -13,7 +13,7 @@ dataset_info:
|
|
13 |
num_bytes: 484098.71867881546
|
14 |
num_examples: 2108
|
15 |
download_size: 1259998
|
16 |
-
dataset_size: 2419575
|
17 |
configs:
|
18 |
- config_name: default
|
19 |
data_files:
|
@@ -21,4 +21,17 @@ configs:
|
|
21 |
path: data/train-*
|
22 |
- split: test
|
23 |
path: data/test-*
|
|
|
|
|
|
|
|
|
|
|
|
|
24 |
---
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
13 |
num_bytes: 484098.71867881546
|
14 |
num_examples: 2108
|
15 |
download_size: 1259998
|
16 |
+
dataset_size: 2419575
|
17 |
configs:
|
18 |
- config_name: default
|
19 |
data_files:
|
|
|
21 |
path: data/train-*
|
22 |
- split: test
|
23 |
path: data/test-*
|
24 |
+
license: apache-2.0
|
25 |
+
task_categories:
|
26 |
+
- text-generation
|
27 |
+
language:
|
28 |
+
- ha
|
29 |
+
pretty_name: Hausa DataMix
|
30 |
---
|
31 |
+
|
32 |
+
|
33 |
+
This dataset is created using the composition of diverse Hausa datasets available on Hugging Face.
|
34 |
+
Each entry in a given split is composed of a `text` and `source` column. The `text` contains the textual data, meanwhile the `source`
|
35 |
+
contains the Hub ID from which this data sample has been retrieved from.
|
36 |
+
|
37 |
+
This dataset was created in the spirit of the [HausaLM](https://github.com/thiomajid/hausa_lm) project.
|