cihangxie nielsr HF Staff commited on
Commit
e3668e6
·
verified ·
1 Parent(s): 446b6db

Add model card metadata and link to code (#1)

Browse files

- Add model card metadata and link to code (5acf4fea4bf58c84cc44c7390251823af88bd7ef)


Co-authored-by: Niels Rogge <[email protected]>

Files changed (1) hide show
  1. README.md +13 -2
README.md CHANGED
@@ -1,3 +1,10 @@
 
 
 
 
 
 
 
1
  <div align="center">
2
  <h1>
3
  <b>m1</b>: Unleash the Potential of Test-Time Scaling for Medical Reasoning in Large Language Models
@@ -7,10 +14,13 @@ A simple test-time scaling strategy, with minimal fine-tuning, can unlock strong
7
  </p>
8
  </div>
9
 
10
- ## Introduction
11
 
 
 
 
12
 
13
- Hi! Welcome to the huggingface repository for m1 (https://github.com/UCSC-VLAA/m1)!
14
 
15
  **m1** is a medical LLM designed to enhance reasoning through efficient test-time scaling. It enables lightweight models to match or exceed the performance of much larger counterparts by extending inference-time “thinking.” Unlike methods that rely on complex RL or expert supervision, m1 achieves strong results through:
16
 
@@ -19,3 +29,4 @@ Hi! Welcome to the huggingface repository for m1 (https://github.com/UCSC-VLAA/m
19
  - **Scaling reasoning at inference using token budgets**, which consistently improves performance across medical QA tasks—up to an optimal ~4K token budget, beyond which performance may degrade due to overthinking.
20
 
21
  - **Identifying medical knowledge as the key bottleneck**, revealing that additional reasoning alone cannot overcome knowledge gaps; instead, improvements require better data quality and increased model capacity.
 
 
1
+ ---
2
+ license: mit
3
+ library_name: transformers
4
+ pipeline_tag: question-answering
5
+ ---
6
+
7
+ ```markdown
8
  <div align="center">
9
  <h1>
10
  <b>m1</b>: Unleash the Potential of Test-Time Scaling for Medical Reasoning in Large Language Models
 
14
  </p>
15
  </div>
16
 
17
+ This repository contains the model presented in the paper [m1: Unleash the Potential of Test-Time Scaling for Medical Reasoning in Large Language Models](https://huggingface.co/papers/2504.00869).
18
 
19
+ Code: https://github.com/UCSC-VLAA/m1
20
+
21
+ ## ⚡ Introduction
22
 
23
+ Hi! Welcome to the huggingface repository for m1!
24
 
25
  **m1** is a medical LLM designed to enhance reasoning through efficient test-time scaling. It enables lightweight models to match or exceed the performance of much larger counterparts by extending inference-time “thinking.” Unlike methods that rely on complex RL or expert supervision, m1 achieves strong results through:
26
 
 
29
  - **Scaling reasoning at inference using token budgets**, which consistently improves performance across medical QA tasks—up to an optimal ~4K token budget, beyond which performance may degrade due to overthinking.
30
 
31
  - **Identifying medical knowledge as the key bottleneck**, revealing that additional reasoning alone cannot overcome knowledge gaps; instead, improvements require better data quality and increased model capacity.
32
+ ```