It's Not Easy Being Wrong: Large Language Models Struggle with Process of Elimination Reasoning Paper • 2311.07532 • Published Nov 13, 2023
KARL: Knowledge-Aware Retrieval and Representations aid Retention and Learning in Students Paper • 2402.12291 • Published Feb 19, 2024
Artifacts or Abduction: How Do LLMs Answer Multiple-Choice Questions Without the Question? Paper • 2402.12483 • Published Feb 19, 2024
Which of These Best Describes Multiple Choice Evaluation with LLMs? A) Forced B) Flawed C) Fixable D) All of the Above Paper • 2502.14127 • Published 4 days ago • 2
Which of These Best Describes Multiple Choice Evaluation with LLMs? A) Forced B) Flawed C) Fixable D) All of the Above Paper • 2502.14127 • Published 4 days ago • 2
Which of These Best Describes Multiple Choice Evaluation with LLMs? A) Forced B) Flawed C) Fixable D) All of the Above Paper • 2502.14127 • Published 4 days ago • 2 • 2
Alignment Personalization Collection Collection of datasets and models for our paper "Whose Boat Does it Float? Improving Personalization in Preference Tuning via Inferred User Personas" • 5 items • Updated 12 days ago