Despite their impressive capabilities, large pre-trained language models (LMs) struggle with consistent reasoning; recently, prompting LMs to generate explanations that self-guide the inference has emerged as a promising direction to amend this. However, these approaches are fundamentally bounded by the correctness of explanations, which themselves are often noisy and inconsistent.
In this work, we develop Maieutic Prompting, which infers a correct answer to a question even from the noisy and inconsistent generations of LM. Maieutic Prompting induces a tree of explanations abductively (eg. ‘X is true, because’ …) and recursively, then frames the inference as a Max-SAT [using Morgadoet al2014] satisfiability problem over these explanations and their logical relations.
We test Maieutic Prompting for true/false QA on 3 challenging benchmarks that require complex commonsense reasoning. Maieutic Prompting achieves up to 20% better accuracy than state-of-the-art prompting methods, and as a fully unsupervised approach, performs competitively with supervised models.
We also show that Maieutic Prompting improves robustness in inference while providing interpretable rationales.
Figure 6: Examples of Maieutic Prompting. We present a case where Maieutic Prompting correctly infers the ground-truth answer (above), and a case where the inferred answer is different from the ground-truth. Even in the latter case, the generated explanations make sense and logically lead to the inferred answer. We provide more examples in Appendix B.