akhooli commited on
Commit
06397c9
·
verified ·
1 Parent(s): 039810b

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +8 -10
README.md CHANGED
@@ -10,23 +10,24 @@ model-index:
10
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
11
  should probably proofread and complete it, then remove this comment. -->
12
 
13
- # ModernBERT-ar-base-small4
14
 
15
- This model is a fine-tuned version of [](https://huggingface.co/) on an unknown dataset.
 
 
 
16
 
17
  ## Model description
18
 
19
- More information needed
20
 
21
  ## Intended uses & limitations
22
 
23
- More information needed
24
 
25
  ## Training and evaluation data
26
 
27
- More information needed
28
-
29
- ## Training procedure
30
 
31
  ### Training hyperparameters
32
 
@@ -45,9 +46,6 @@ The following hyperparameters were used during training:
45
  - training_steps: 50000
46
  - mixed_precision_training: Native AMP
47
 
48
- ### Training results
49
-
50
-
51
 
52
  ### Framework versions
53
 
 
10
  <!-- This model card has been generated automatically according to the information the Trainer had access to. You
11
  should probably proofread and complete it, then remove this comment. -->
12
 
13
+ # ModernBERT-ar-base-tiny
14
 
15
+ This model was trained on [Fineweb2 Ar sample](https://huggingface.co/datasets/akhooli/fineweb2_ar_24_sample) dataset.
16
+ The tokenizer was also trained using the same dataset.
17
+ See [sample code](https://colab.research.google.com/drive/1CUsUsJQV4ZzJar2987yAzCTn8ve4hR5b?usp=sharing)
18
+ (usage and training) and [initial post](https://www.linkedin.com/posts/akhooli_a-micro-arabic-modern-bert-a-couple-weeks-activity-7282005813357875202-SAGk)
19
 
20
  ## Model description
21
 
22
+ ModernBERT Arabic (MLM) experiment.
23
 
24
  ## Intended uses & limitations
25
 
26
+ Educational and explorational uses only. Limited data, not fully trained.
27
 
28
  ## Training and evaluation data
29
 
30
+ Evaluation on 5% of the data, uses 2 GPUs.
 
 
31
 
32
  ### Training hyperparameters
33
 
 
46
  - training_steps: 50000
47
  - mixed_precision_training: Native AMP
48
 
 
 
 
49
 
50
  ### Framework versions
51