anujga
anujga
AI & ML interests
None yet
Recent Activity
updated
a collection
24 days ago
O1
updated
a collection
24 days ago
O1
updated
a collection
about 1 month ago
O1
Organizations
None yet
Retrieval
Aggregates
Persona
Sft
Rl
-
RLCD: Reinforcement Learning from Contrast Distillation for Language Model Alignment
Paper • 2307.12950 • Published • 10 -
HumanLLMs/Human-Like-DPO-Dataset
Viewer • Updated • 10.9k • 949 • 229 -
sam-paech/gutenberg3-generalfiction-scifi-fantasy-romance-adventure-dpo
Viewer • Updated • 5.65k • 73 • 28 -
RLHFlow/Deepseek-PRM-Data
Viewer • Updated • 253k • 394 • 17
Benchmark
Datasets
agent
data/vision
Multi-lingual
Special
PT
Pt-classify
O1
Programming
Architecture
-
UT5: Pretraining Non autoregressive T5 with unrolled denoising
Paper • 2311.08552 • Published • 8 -
Rethinking Attention: Exploring Shallow Feed-Forward Neural Networks as an Alternative to Attention Layers in Transformers
Paper • 2311.10642 • Published • 26 -
Densing Law of LLMs
Paper • 2412.04315 • Published • 19
Theory
data/tool
chemistry
rl-papers
Multi-lingual
Retrieval
Special
Aggregates
PT
Persona
Pt-classify
Sft
O1
Rl
-
RLCD: Reinforcement Learning from Contrast Distillation for Language Model Alignment
Paper • 2307.12950 • Published • 10 -
HumanLLMs/Human-Like-DPO-Dataset
Viewer • Updated • 10.9k • 949 • 229 -
sam-paech/gutenberg3-generalfiction-scifi-fantasy-romance-adventure-dpo
Viewer • Updated • 5.65k • 73 • 28 -
RLHFlow/Deepseek-PRM-Data
Viewer • Updated • 253k • 394 • 17
Programming
Benchmark
Architecture
-
UT5: Pretraining Non autoregressive T5 with unrolled denoising
Paper • 2311.08552 • Published • 8 -
Rethinking Attention: Exploring Shallow Feed-Forward Neural Networks as an Alternative to Attention Layers in Transformers
Paper • 2311.10642 • Published • 26 -
Densing Law of LLMs
Paper • 2412.04315 • Published • 19
Datasets
Theory
agent
data/tool
data/vision
chemistry