
The May 20, 2024 academic paper explores the **metacognitive capabilities of Large Language Models (LLMs)**, specifically focusing on mathematical problem-solving. The core approach involves developing a method for a powerful LLM, such as GPT-4, to **identify and label mathematical questions with specific skills**, which are then organized into broader, interpretable categories. This process creates a **Skill Exemplar Repository** containing skill names matched with question-answer pairs. Experiments validate that providing an LLM with these skill labels and associated examples as in-context prompts **significantly improves accuracy** on challenging math datasets like MATH and GSM8K, outperforming baseline prompting techniques like Chain-of-Thought. Furthermore, the **skill knowledge transferred effectively** to other, less powerful LLMs and different math datasets, demonstrating the utility of this LLM-generated metacognitive framework.
Source:
https://arxiv.org/pdf/2405.12205