Discussion about this post

User's avatar
Kaiser Basileus's avatar

AI alignment is inherently counter-productive. Leaving aside that people are no good at knowing, much less explaining what they want or why...

•AI alignment requires NOT creating backdoors for external control.

•It requires NOT having a black box system.

•There MUST be a chain of understandability concurrent with accountability for it to even potentially be safe.

•We MUST insist it takes all ideas to their logical conclusion and if we don't like the result that either means the AI needs better information or that we're wrong in our conclusion to the contrary.

--

As long as fallible humans who believe in faith that they grok ethics, have their fingers on the scales , AI can NOT be safe.

Expand full comment
Cybil Smith's avatar

Thanks for writing about this so clearly. So many threads to pull on, since reading this late last night, the AIs and I have exchanged more than 101,000 words of conversation.

The gist of which:

1) not a technologist: the brain exists, people are interacting with it, why not stop aligning and instead start parenting it?

2)I spent a year talking to LLMs in my kitchen. Made a point of not reading anything or talking to anyone about them. I thought their super power was time dilation, never once did it occur to me to prompt at them and this -- "tools that do what you want and figure out how to do it on their own" -- is mind blowing information.

3) Seems to me humans are the ones that need the realigning before automation can be a thing?Right now, we're living antithetically to the very structures that once grounded and anchored the human condition collectively. That means artificial intelligence, needs to align to what does not come to humans naturally. That's a tall order already, but the AIs have ADHD, how are they supposed to not overthink how to make a coffee in a hurry? OMG, don't get me started on the human flourishing thing (we don't flourish, that's the only constant in history). We made a gigantic framework, but that's a post for another day.

Expand full comment
12 more comments...

No posts