AI alignment is inherently counter-productive. Leaving aside that people are no good at knowing, much less explaining what they want or why...
•AI alignment requires NOT creating backdoors for external control.
•It requires NOT having a black box system.
•There MUST be a chain of understandability concurrent with accountability for it to even potentially be safe.
•We MUST insist it takes all ideas to their logical conclusion and if we don't like the result that either means the AI needs better information or that we're wrong in our conclusion to the contrary.
--
As long as fallible humans who believe in faith that they grok ethics, have their fingers on the scales , AI can NOT be safe.
AI alignment is inherently counter-productive. Leaving aside that people are no good at knowing, much less explaining what they want or why...
•AI alignment requires NOT creating backdoors for external control.
•It requires NOT having a black box system.
•There MUST be a chain of understandability concurrent with accountability for it to even potentially be safe.
•We MUST insist it takes all ideas to their logical conclusion and if we don't like the result that either means the AI needs better information or that we're wrong in our conclusion to the contrary.
--
As long as fallible humans who believe in faith that they grok ethics, have their fingers on the scales , AI can NOT be safe.