šØBREAKING: US @NIST publishes 1st draft of its "AI Risk Management Framework: Generative AI Profile." Important information & quotes:
ā”ļøThis is a comprehensive document that contains an overview of risks unique to or exacerbated by generative AI (GAI) and an extensive list of actions to manage GAI's... See more
"Role prompting"... telling the model to assume a role has never been a good way to elicit capabilities/style/etc.
For instance, if you ask one of the Claude models to simulate Bing Sydney, assuming you can get it to consent, the simulation will probably be very inaccurate. But if you use a prompt that tricks them into... See more
New Paper: Stanford researcher (@joon_s_pk) discovers how to clone human personalities and inject them into AI Agents š§
This builds on last year's paper which put 1000's of fully automated agents in a simulated town.
The results are wild. š https://t.co/IpG4R9sfcQ
IMO @AnthropicAI is very close to making a breakthrough in productizable interpretability.
For ~4 years all we've had to really control LLMs is temperature/top_p and logit bias. We recently got `seed` and constrained structured output, with `interactive=false` on the way.
But now Claude... See more