“Language Models That Seek for Knowledge: Modular Search & Generation for Dialogue and Prompt Completion”, 2022-03-24 ():
Language models (LMs) have recently been shown to generate more factual responses by employing modularity ( et al 2021) in combination with retrieval ( et al 2021).
We extend the recent approach of et al 2021 to include internet search as a module. Our SeeKeR (Search engine → Knowledge → Response) method thus applies a single LM to 3 modular tasks in succession: search, generating knowledge, and generating a final response. We show that, when using SeeKeR as a dialogue model, it outperforms the state-of-the-art model BlenderBot 2 ( et al 2021) on open-domain knowledge-grounded conversations for the same number of parameters, in terms of consistency, knowledge and per-turn engagingness. SeeKeR applied to topical prompt completions as a standard language model outperforms GPT-2 ( et al 2019) and GPT-3 ( et al 2020) in terms of factuality and topicality, despite GPT-3 being a vastly larger model.
Our code and models are made publicly available.