-
chevron_right
AI with a moral compass? Anthropic outlines “Constitutional AI” in its Claude chatbot
news.movim.eu / ArsTechnica · Tuesday, 9 May, 2023 - 21:16
On Tuesday, AI startup Anthropic detailed the specific principles of its " Constitutional AI " training approach that provides its Claude chatbot with explicit "values." It aims to address concerns about transparency, safety, and decision-making in AI systems without relying on human feedback to rate responses.
Claude is an AI chatbot similar to OpenAI's ChatGPT that Anthropic released in March .
"We’ve trained language models to be better at responding to adversarial questions, without becoming obtuse and saying very little," Anthropic wrote in a tweet announcing the paper. "We do this by conditioning them with a simple set of behavioral principles via a technique called Constitutional AI."