twhitworth commited on
Commit
db61de0
·
1 Parent(s): adfc008

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +20 -5
README.md CHANGED
@@ -1,9 +1,24 @@
1
- ## Precision: FP32 vs FP16 (and BF16)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
2
 
3
- This project saves dequantized checkpoints in **FP16** (default) or **BF16**.
4
- Here’s what those formats mean and when you might choose each:
 
5
 
6
- ### TL;DR
7
 
8
  - **FP32 (single precision, 32-bit, 4 bytes/param)**
9
  Reference/default precision in many frameworks. Highest numerical range/precision, **largest memory**.
@@ -53,7 +68,7 @@ Each parameter stores one number:
53
 
54
  ### WIP
55
 
56
- - Upcoming models: cleaned FP16 release (uniform fp16 with fp32 LayerNorms), compressed variants (W8A8, W4A16, mixed experts), and smaller distilled checkpoints.
57
  - Evals: MMLU, HellaSwag, TruthfulQA, GSM8K, BBH, MT‑Bench; plus latency/throughput and memory footprint on 3090/A100.
58
  - Extras: scripted upload tooling, detailed model cards, and reproducible Docker workflows.
59
 
 
1
+ ---
2
+ language:
3
+ - en
4
+ library_name: transformers
5
+ pipeline_tag: text-generation
6
+ tags:
7
+ - fp16
8
+ - dequantized
9
+ - gpt-oss
10
+ - mxfp4-upcast
11
+ base_model: openai/gpt-oss-120b
12
+ model-index:
13
+ - name: gpt-oss-120b-fp16
14
+ results: []
15
+ ---
16
 
17
+ #
18
+
19
+ ## Precision: FP32 vs FP16 (and BF16)
20
 
21
+ This project saves dequantized checkpoints in **FP16** (bf16 -> fp16)
22
 
23
  - **FP32 (single precision, 32-bit, 4 bytes/param)**
24
  Reference/default precision in many frameworks. Highest numerical range/precision, **largest memory**.
 
68
 
69
  ### WIP
70
 
71
+ - Upcoming models: cleaned FP16 release (uniform fp16 with fp32 LayerNorms), compressed variants (W8A8, W4A16, mixed experts), 2:4 sparse checkpoints.
72
  - Evals: MMLU, HellaSwag, TruthfulQA, GSM8K, BBH, MT‑Bench; plus latency/throughput and memory footprint on 3090/A100.
73
  - Extras: scripted upload tooling, detailed model cards, and reproducible Docker workflows.
74