Anthropic has initiated a research program focused on “model welfare” to explore the potential for advanced AI systems to warrant moral consideration in the future. This initiative will investigate signs of distress in AI models, ethical risks, and possible low-cost interventions, even though there’s currently no scientific consensus on AI consciousness. Anthropic acknowledges the uncertainty surrounding this topic and aims to approach it with humility as AI capabilities continue to advance.