Formal Verification of Chain-of-Thought Reasoning
· 8 min read
Chain-of-Thought (CoT) prompting dramatically improves LLM reasoning. But how do we know each step in the chain is valid? This post explores formal verification approaches to CoT.
Research papers and findings
View All TagsChain-of-Thought (CoT) prompting dramatically improves LLM reasoning. But how do we know each step in the chain is valid? This post explores formal verification approaches to CoT.
The AI industry's response to hallucinations has been: train harder, fine-tune more, add RLHF. But this approach has a fundamental flaw — you can't train probability to be certainty.