-
Mistral 7B
Paper • 2310.06825 • Published • 46 -
Llama 2: Open Foundation and Fine-Tuned Chat Models
Paper • 2307.09288 • Published • 244 -
OpenChat: Advancing Open-source Language Models with Mixed-Quality Data
Paper • 2309.11235 • Published • 15 -
DeepSeek-R1: Incentivizing Reasoning Capability in LLMs via Reinforcement Learning
Paper • 2501.12948 • Published • 330
Collections
Discover the best community collections!
Collections including paper arxiv:2310.06825
-
Qwen2.5 Technical Report
Paper • 2412.15115 • Published • 346 -
Qwen2.5-Coder Technical Report
Paper • 2409.12186 • Published • 141 -
Qwen2.5-Math Technical Report: Toward Mathematical Expert Model via Self-Improvement
Paper • 2409.12122 • Published • 3 -
Qwen2.5-VL Technical Report
Paper • 2502.13923 • Published • 136
-
ReAct: Synergizing Reasoning and Acting in Language Models
Paper • 2210.03629 • Published • 18 -
Attention Is All You Need
Paper • 1706.03762 • Published • 52 -
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Paper • 1810.04805 • Published • 16 -
Jamba: A Hybrid Transformer-Mamba Language Model
Paper • 2403.19887 • Published • 107
-
Attention Is All You Need
Paper • 1706.03762 • Published • 52 -
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
Paper • 1810.04805 • Published • 16 -
DistilBERT, a distilled version of BERT: smaller, faster, cheaper and lighter
Paper • 1910.01108 • Published • 14 -
Language Models are Few-Shot Learners
Paper • 2005.14165 • Published • 13
-
Mistral 7B
Paper • 2310.06825 • Published • 46 -
Instruction Tuning with Human Curriculum
Paper • 2310.09518 • Published • 3 -
RAFT: Adapting Language Model to Domain Specific RAG
Paper • 2403.10131 • Published • 69 -
Instruction-tuned Language Models are Better Knowledge Learners
Paper • 2402.12847 • Published • 26