PaperPulse logo
FeedTopicsAI Researcher FeedBlogPodcastAccount

Stay Updated

Get the latest research delivered to your inbox

Platform

  • Home
  • About Us
  • Search Papers
  • Research Topics
  • Researcher Feed

Resources

  • Newsletter
  • Blog
  • Podcast
PaperPulse•

AI-powered research discovery platform

© 2024 PaperPulse. All rights reserved.

ABCD: All Biases Come Disguised

ArXivSource

Mateusz Nowak, Xavier Cadet, Peter Chin

cs.CL
cs.LG
|
Feb 19, 2026
3 views

One-line Summary

The paper introduces a bias-reduced evaluation protocol for LLMs in multiple-choice questions that improves robustness to answer permutations with minimal performance loss.

Plain-language Overview

When evaluating large language models (LLMs) using multiple-choice questions, biases can arise from the position of answers, labels, and examples used in the prompts. This study identifies these biases and proposes a new method to reduce them by using uniform, unordered labels and requiring the model to consider the entire answer. This approach enhances the consistency of LLMs' performance across different arrangements of answers, showing that the models' true capabilities can be assessed more accurately. The method maintains high performance while reducing variability in results, making it a more reliable evaluation tool.

Technical Details