Simply adding "Repeat the question before answering it." somehow make the models answer the trick question correctly. 🤔 Probable explanations:✨ 📌 Repeating the question in the model's context, significantly increasing the likelihood of the model detecting any potential "gotchas." 📌 One hypothesis is that maybe it puts the model into more of a completion mode vs answering from a chat instruct mode. 📌 Another, albeit less likely, reason could be that the model might assume the user’s question contains mistakes (e.g., the user intended to ask about a Schrödinger cat instead of a dead cat). However, if the question is in the assistant’s part of the context, the model trusts it to be accurate. 📚 The related Paper is EchoPrompt which proposes this techniques to rephrase original Prompt/queries before answering them. improves the Zero-shot-CoT performance of code-davinci-002 by 5% in numerical tasks and 13% in reading comprehension tasks.

Thumbnail of www-x-com-rohanpaul-ai-status-1830230678673223737
Simply adding "Repeat the question before answering it." somehow make the models answer the trick question correctly. 🤔 Probable explanations:✨ 📌 Repeating the question in the model's context, significantly increasing the likelihood of the model detecting any potential "gotchas." 📌 One hypothesis is that maybe it puts the model into more of a completion mode vs answering from a chat instruct mode. 📌 Another, albeit less likely, reason could be that the model might assume the user’s question contains mistakes (e.g., the user intended to ask about a Schrödinger cat instead of a dead cat). However, if the question is in the assistant’s part of the context, the model trusts it to be accurate. 📚 The related Paper is EchoPrompt which proposes this techniques to rephrase original Prompt/queries before answering them. improves the Zero-shot-CoT performance of code-davinci-002 by 5% in numerical tasks and 13% in reading comprehension tasks.

Anthropic \ Tracing Model Outputs to the Training Data

Nicolay Gerold added

ChatGPT as muse, not oracle

sari and added

Ben Auffarth • Generative AI with LangChain: Build large language model (LLM) apps with Python, ChatGPT, and other LLMs

Andy Matuschak • How to Write Good Prompts

added

Albert Webson • Do Prompt-Based Models Really Understand the Meaning of their Prompts?

"Violation of Expectation via Metacognitive Prompting Reduces Theory of Mind Prediction Error in Large Language Models"

This paper investigates how Violation of Expectation (VoE) and metacognitive prompting can be used to reduce Theory of Mind prediction errors in Large Language Models (LLMs) in the context of human-AI interaction.

arxiv.org

Shortwave — rajhesh.panchanadhan@gmail.com [Gmail alternative]