LLM Error Stubbornness

Cluster focuses on large language models like ChatGPT stubbornly insisting on incorrect answers, refusing to admit mistakes when corrected, and related issues like hallucinations, prompt failures, and workarounds such as regenerating responses.

➡️ Stable 1.4x AI & Machine Learning
5,047
Comments
16
Years Active
5
Top Authors
#8466
Topic ID

Activity Over Time

2010
1
2011
1
2012
3
2013
1
2015
1
2016
5
2017
5
2018
8
2019
16
2020
51
2021
46
2022
224
2023
1,401
2024
1,060
2025
2,047
2026
181

Keywords

AI OP LLM chatgpt.com CNN GPT OpenAi openai.com API CC llm prompt gpt chat context chatgpt temperature model prompts answer

Sample Comments

olabyne Aug 11, 2023 View on HN

That's a OpenAi problem, not a LLM problem

RecycledEle Apr 21, 2023 View on HN

Tip: Don't point out contradictions to a LLM AI. Just tell it to redo the wrong part. They can not figure out that they are wrong (yet.) They only spit out best effort attempts.

saimiam Dec 4, 2022 View on HN

This is super interesting. Can you share some prompts that you recall had believable but wrong responses?

kovezd Sep 6, 2024 View on HN

I find this behavior when using ChatGPT.If I ask for a task, and the output is not the one expected: I ask for the motivation that lead to the bad decisions. Then, ChatGPT proceeds to retry the task "incorporating" my feedback, not answering my question!!

SamFold Nov 7, 2023 View on HN

I tried this and it seemed to break ChatGPT, it blurted out something which made no sense and then offered to regenerate it. How is it supposed to work?

vidarh Aug 10, 2025 View on HN

It doesn't work this time because there are plenty of models, including GPT5 Thinking that can handle this correctly, and so it is clear this isn't a systemic issue that can't be trained out of them.

kamalkishor1991 Mar 28, 2024 View on HN

Thanks. I have also noticed this in some cases. Sometime GPT-3 does not follow the prompt. Please retry with a new query. It works in most cases. I will improve the prompt or to make sure this does not happen.

msp26 Sep 12, 2024 View on HN

That's unfortunate. When an LLM makes a mistake it's very helpful to read the CoT and see what went wrong (input error/instruction error/random shit)

brandall10 Jul 25, 2024 View on HN

I haven't played with this model, but rarely do I find working w/ Claude or GPT-4 for that to be the case. If you say it's incorrect, it will give you another answer instead of insisting on correctness.

npew Mar 17, 2022 View on HN

I work at OpenAI. This was unfortunately a bug, and we've fixed it since.