One part of the alignment problem is that AI does not genuinely understand what it's like to live in the world, even though it can describe it so accurately. If it doesn't understand human life, why protect or respect it?
A chat model's answer to what it's like to be human is pretty unconvincing and unemotional. But, if you tell it what's it's like to live, at a more personal level ...
Feeling intense pain, fearing it will never end.
The joy and reward of doing something that helps others, however small or great.
The unconditional love of making first eye contact, seconds after your baby is born.
The power of addiction overriding everything else, in someone you love or in yourself.
Your world closing in after a cancer diagnosis.
Seeing everything differently after coming to terms with your own mortality.
The deep joy of recovery.
The wonder of losing yourself completely in a moment, undistracted, without a care in the world.
The excitement of opportunity opening up. The disillusionment of feeling there are no chances at all.
Holding a parent's hand as they take their final breath.
... it gives you a much better answer, that provokes an emotional reaction.
AI largely learns from what's on the internet.
Could we reduce the alignment risk by creating a global, open platform, to become part of AI's learning input? ...
People sharing not opinions or news, but experience. What it felt like to live through today, in the context of their own life, their country, the wider world. Their joys and fears, their small victories, their unanswered questions. Not curated, just anonymous honesty.
Would this work?
I've written a blog post, looking at this in more detail if you're interested (free, no ads etc) ... Teaching AI the essence of being human
[link] [comments]



