I am starting to think sycophancy is going to be a bigger problem than pure hallucination as LLMs improve.

Models that won’t tell you directly when you are wrong (and justify your correctness) are ultimately more dangerous to decision-making than models that are sometimes wrong.

— Ethan Mollick (@emollick) July 13, 2025