Simply adding "Repeat the question before answering it." somehow make the models answer the trick question correctly. 🤔 Probable explanations:✨ 📌 Repeating the question in the model's context, significantly increasing the likelihood of the model detecting any potential "gotchas." 📌 One hypothesis is that maybe it puts the model into more of a completion mode vs answering from a chat instruct mode. 📌 Another, albeit less likely, reason could be that the model might assume the user’s question contains mistakes (e.g., the user intended to ask about a Schrödinger cat instead of a dead cat). However, if the question is in the assistant’s part of the context, the model trusts it to be accurate. 📚 The related Paper is EchoPrompt which proposes this techniques to rephrase original Prompt/queries before answering them. improves the Zero-shot-CoT performance of code-davinci-002 by 5% in numerical tasks and 13% in reading comprehension tasks.
Anthropic \ Tracing Model Outputs to the Training Data
Nicolay Gerold added
Nathan Storey added
ChatGPT as muse, not oracle
sari and added
Innovations like prompting, chain-of-thought reasoning, retrieval grounding, and others are needed to educate models.
Ben Auffarth • Generative AI with LangChain: Build large language model (LLM) apps with Python, ChatGPT, and other LLMs
consistent
answers, lighting the same bulbs each time you perform the task. Otherwise, you may run afoul of an interference phenomenon called “retrieval-induced forgetting”
This effect has been produced in many experiments but is not yet well understood. For an overview, see Murayama et al, Forgetting as a c... See more
Andy Matuschak • How to Write Good Prompts
We find that models learn just as fast with many prompts that are intentionally irrelevant or even pathologically misleading as they do with instructively “good” prompts. Further, such patterns hold even for models as large as 175 billion parameters (Brown et al., 2020) as well as the recently proposed instruction-tuned models which are trained on
... See moreAlbert Webson • Do Prompt-Based Models Really Understand the Meaning of their Prompts?
Isabelle Levent added
"Violation of Expectation via Metacognitive Prompting Reduces Theory of Mind Prediction Error in Large Language Models"
This paper investigates how Violation of Expectation (VoE) and metacognitive prompting can be used to reduce Theory of Mind prediction errors in Large Language Models (LLMs) in the context of human-AI interaction.
arxiv.orgBhaumik Patel added
Shortwave — rajhesh.panchanadhan@gmail.com [Gmail alternative]
Nicolay Gerold added