با برنامه Player FM !
LLMs Cannot Find Reasoning Errors, but They Can Correct Them!
Manage episode 421464394 series 3474148
This story was originally published on HackerNoon at: https://hackernoon.com/llms-cannot-find-reasoning-errors-but-they-can-correct-them.
In this paper, we break down the self-correction process into two core components: mistake finding and output correction.
Check more stories related to machine-learning at: https://hackernoon.com/c/machine-learning. You can also check exclusive content about #llms, #llm-mistake-finding, #llm-output-correction, #big-bench-mistake, #chain-of-thought, #nlp, #self-consistency, #zero-shot-prompting, and more.
This story was written by: @textmodels. Learn more about this writer by checking @textmodels's about page, and for more stories, please visit hackernoon.com.
Large Language Models (LLMs) have dominated the field of NLP in recent years. LLMs have demonstrated the ability to solve tasks with zero- or few-shot prompting. Recent literature has focused on the concept of self-correction, i.e. having an LLM correct its own outputs. Attempts to self-correct logical or reasoning errors often cause correct answers to become incorrect, resulting in worse performances overall. In this paper, we break down the self-Correction process into two core components: mistake finding and output correction. For mistake finding, we release BIG-Bench Mistake, a dataset of logical mistakes in Chain-of-Thought reasoning traces. For output
316 قسمت
Manage episode 421464394 series 3474148
This story was originally published on HackerNoon at: https://hackernoon.com/llms-cannot-find-reasoning-errors-but-they-can-correct-them.
In this paper, we break down the self-correction process into two core components: mistake finding and output correction.
Check more stories related to machine-learning at: https://hackernoon.com/c/machine-learning. You can also check exclusive content about #llms, #llm-mistake-finding, #llm-output-correction, #big-bench-mistake, #chain-of-thought, #nlp, #self-consistency, #zero-shot-prompting, and more.
This story was written by: @textmodels. Learn more about this writer by checking @textmodels's about page, and for more stories, please visit hackernoon.com.
Large Language Models (LLMs) have dominated the field of NLP in recent years. LLMs have demonstrated the ability to solve tasks with zero- or few-shot prompting. Recent literature has focused on the concept of self-correction, i.e. having an LLM correct its own outputs. Attempts to self-correct logical or reasoning errors often cause correct answers to become incorrect, resulting in worse performances overall. In this paper, we break down the self-Correction process into two core components: mistake finding and output correction. For mistake finding, we release BIG-Bench Mistake, a dataset of logical mistakes in Chain-of-Thought reasoning traces. For output
316 قسمت
כל הפרקים
×به Player FM خوش آمدید!
Player FM در سراسر وب را برای یافتن پادکست های با کیفیت اسکن می کند تا همین الان لذت ببرید. این بهترین برنامه ی پادکست است که در اندروید، آیفون و وب کار می کند. ثبت نام کنید تا اشتراک های شما در بین دستگاه های مختلف همگام سازی شود.