The fact that most individual neurons are uninterpretable presents a serious roadblock to a mechanistic understanding of language models. We demonstrate a method for decomposing groups of neurons into interpretable features with the potential to move past that roadblock.
Sarah Wang • What Builders Talk About When They Talk About AI | Andreessen Horowitz
Nicolay Gerold added
Hagen Peters added
“testing with concept activation vectors,” or TCAV, which offers a way to use such human concepts to understand the internal workings of the network.
Brian Christian • The Alignment Problem
Sofia Quaglia • How the brains of social animals synchronise and expand one another
Keely Adler added
Nathan Storey added
Generative AI is quite good at certain parts of the value chain of knowledge work, but thinks quite differently from humans. Anthropic, a company that focuses on understanding how AI works, has found that humans working side-by-side with expert AI assistants to perform various tasks produce superior performance compared to either the AI or a human
... See moreNoah Smith • Generative AI: autocomplete for everything
sari added
future.com • How Recommendation Algorithms Actually Work | Future - https://future.com/forget-open-source-algorithms-focus-on-experiments-instead
Tom So added
dailynous.com • Philosophers on GPT-3 (
Kasper Jordaens added