johnr0
's Collections
Chain-of-Verification Reduces Hallucination in Large Language Models
Paper
•
2309.11495
•
Published
•
38
EIPE-text: Evaluation-Guided Iterative Plan Extraction for Long-Form
Narrative Text Generation
Paper
•
2310.08185
•
Published
•
8
The Consensus Game: Language Model Generation via Equilibrium Search
Paper
•
2310.09139
•
Published
•
14
In-Context Pretraining: Language Modeling Beyond Document Boundaries
Paper
•
2310.10638
•
Published
•
30
Reward-Augmented Decoding: Efficient Controlled Text Generation With a
Unidirectional Reward Model
Paper
•
2310.09520
•
Published
•
12
Self-RAG: Learning to Retrieve, Generate, and Critique through
Self-Reflection
Paper
•
2310.11511
•
Published
•
76
VeRA: Vector-based Random Matrix Adaptation
Paper
•
2310.11454
•
Published
•
30
Safe RLHF: Safe Reinforcement Learning from Human Feedback
Paper
•
2310.12773
•
Published
•
28
In-Context Learning Creates Task Vectors
Paper
•
2310.15916
•
Published
•
43
Deja Vu: Contextual Sparsity for Efficient LLMs at Inference Time
Paper
•
2310.17157
•
Published
•
14
Controlled Decoding from Language Models
Paper
•
2310.17022
•
Published
•
15
Tell Your Model Where to Attend: Post-hoc Attention Steering for LLMs
Paper
•
2311.02262
•
Published
•
14
S-LoRA: Serving Thousands of Concurrent LoRA Adapters
Paper
•
2311.03285
•
Published
•
31
Prompt Cache: Modular Attention Reuse for Low-Latency Inference
Paper
•
2311.04934
•
Published
•
32
System 2 Attention (is something you might need too)
Paper
•
2311.11829
•
Published
•
42
Adapters: A Unified Library for Parameter-Efficient and Modular Transfer
Learning
Paper
•
2311.11077
•
Published
•
28
Tuning Language Models by Proxy
Paper
•
2401.08565
•
Published
•
23
Self-Rewarding Language Models
Paper
•
2401.10020
•
Published
•
146
Collaborative Development of NLP models
Paper
•
2305.12219
•
Published
Suppressing Pink Elephants with Direct Principle Feedback
Paper
•
2402.07896
•
Published
•
11
A Tale of Tails: Model Collapse as a Change of Scaling Laws
Paper
•
2402.07043
•
Published
•
15
Direct Language Model Alignment from Online AI Feedback
Paper
•
2402.04792
•
Published
•
31