The integrity behind ethical AI

The artificial intelligence industry was jolted May 22 when a leading company, Anthropic, announced that its latest model resorted to self-preservation in a test run. Much of the shock was simply that the new digital assistant, Claude Opus 4, used blackmail against a fictional character in a particular scenario in order to avoid being shut down. It was like a tingling plot twist in a sci-fi flick.

Yet just as jolting was that Anthropic was so open about Claude’s failure to operate with a level of moral intelligence that its inventors sought to build into it.

The transparency was intentional. Anthropic foresees the reality of ethical AI as dependent on the values of both researchers and users who demand qualities like transparency and trust in AI. The company’s motto for training AI systems is “helpful, honest, and harmless.” And it shares its safety standards and results of new models.

Source link

Related Posts

Load More Posts Loading...No More Posts.