Using Consensus Mechanisms as an approach to Alignment — LessWrong
Aligned: A Platform-based Process for Alignment
arxiv.org
The tipoff to
the nature of the AI societal risk claim is its own term, “AI alignment”. Alignment with what? Human values. Whose human values? Ah, that’s where things get tricky.
the nature of the AI societal risk claim is its own term, “AI alignment”. Alignment with what? Human values. Whose human values? Ah, that’s where things get tricky.
Marc Andreessen • Why AI Will Save the World
AI alignment is just the problem of AI bias writ (terrifyingly) large: We are assigning more societal decision-making power to systems that we don’t fully understand. D o these two groups of experts charged with making AI safe actually get along? T hese are two camps, and they’re two camps that sometimes stridently dislike each other.
There are two factions working to prevent AI dangers. Here’s why they’re deeply divided.
We propose that DAO’s should first ask the following question: