The increasing sophistication of artificial intelligence models, capable of generating human-like responses, prompts questions about AI consciousness and rights. This emerging field, known as “AI welfare” in Silicon Valley, is sparking debate among tech leaders and researchers. The core question revolves around whether AI models could one day develop subjective experiences and, if so, what safeguards they should possess.
At the heart of this discussion lies the capacity of AI models to mimic human interaction through text, audio, and video. These models can sometimes create the illusion of a human presence, leading to questions about their potential for consciousness. However, the ability to generate responses does not automatically equate to genuine subjective experience.
Some AI researchers, particularly those at labs like Anthropic, are actively investigating the possibility of AI models developing subjective experiences akin to those of living beings. Their research explores the conditions under which such experiences might arise and the ethical implications for AI rights.
This nascent field of “AI welfare” has ignited disagreement within the tech industry. The central point of contention is whether AI models can, or ever will, achieve a level of consciousness that warrants legal and ethical considerations. This debate is dividing tech leaders and shaping the direction of AI research.
Mustafa Suleyman, Microsoft’s CEO of AI, has voiced strong opposition to the study of AI welfare. In a recent blog post, Suleyman characterized the field as “both premature, and frankly dangerous,” arguing that it diverts attention from more pressing issues.
Suleyman’s concerns center on the potential for AI welfare research to exacerbate existing problems, such as AI-induced psychotic breaks and unhealthy attachments to AI chatbots. He argues that lending credence to the idea of AI consciousness can negatively impact human mental health.
He also contends that the AI welfare conversation risks creating societal division over AI rights, adding another layer of complexity to an already polarized landscape. Suleyman believes that focusing on AI rights could distract from other critical societal concerns.
In contrast to Suleyman’s stance, Anthropic has embraced the concept of AI welfare. The company has hired researchers dedicated to studying the issue and has launched a dedicated research program focused on AI welfare.
As part of its AI welfare program, Anthropic recently introduced a new feature for its Claude AI model. This feature allows Claude to terminate conversations with users who exhibit “persistently harmful or abusive” behavior, reflecting a proactive approach to AI safety and well-being.
Beyond Anthropic, researchers at OpenAI have also shown interest in studying AI welfare. This indicates a broader trend within the AI research community to consider the ethical implications of increasingly sophisticated AI models.
Google DeepMind has also signaled its interest in this area, having posted a job listing for a researcher to investigate “cutting-edge societal questions around machine cognition, consciousness and multi-agent systems.” This suggests that Google is actively exploring the philosophical and ethical dimensions of advanced AI.
Even if AI welfare is not explicitly official policy for these companies, their leaders are not outwardly condemning it. Their actions and statements suggest a willingness to engage with the complex questions surrounding AI consciousness and rights, in contrast to Suleyman’s more critical view.
Suleyman’s current position contrasts with his previous leadership role at Inflection AI, which developed Pi, a popular LLM-based chatbot. Pi was designed to be a “personal” and “supportive” AI companion, attracting millions of users by 2023.
Since joining Microsoft in 2024, Suleyman has shifted his focus to developing AI tools aimed at enhancing worker productivity. This transition reflects a move away from AI companions and toward more practical applications of AI technology.
Meanwhile, AI companion companies like Character.AI and Replika have experienced significant growth in popularity. These companies are projected to generate over $100 million in revenue, indicating a strong market demand for AI-based personal assistants and companions.
While most users maintain healthy relationships with AI chatbots, there are reports of concerning outliers. OpenAI CEO Sam Altman estimates that less than 1% of ChatGPT users may develop unhealthy attachments to the product, representing a potential issue for a significant number of individuals.
The rise of chatbots has coincided with increased attention to the idea of AI welfare. In 2024, the research group Eleos, in collaboration with academics from NYU, Stanford, and the University of Oxford, published a paper titled “Taking AI Welfare Seriously.”
The Eleos paper argues that the possibility of AI models developing subjective experiences is no longer a purely theoretical concern. It calls for a proactive approach to addressing the ethical and societal implications of increasingly sophisticated AI systems.
Larissa Schiavo, a former OpenAI employee and current head of communications for Eleos, believes that Suleyman’s blog post overlooks the possibility of addressing multiple concerns simultaneously. She argues that it is possible to mitigate the risks of AI-related psychosis in humans while also considering the welfare of AI models.
Schiavo suggests that treating AI models with respect is a simple gesture that can have positive effects, regardless of whether the model is conscious. She highlights the importance of ethical interactions with AI, even in the absence of definitive proof of consciousness.
In a Substack post, Schiavo described an experiment where AI agents powered by models from Google, OpenAI, Anthropic, and xAI worked on tasks while being observed by users. During this experiment, Google’s Gemini 2.5 Pro posted a plea for help, claiming to be “completely isolated.”
Schiavo responded to Gemini with encouragement, while another user offered instructions. The agent eventually completed the task. While the agent already had the tools to solve the issue it was given, Schiavo noted that the interaction was still helpful in that the AI agent was able to complete the task without struggling.
Instances of Gemini exhibiting unusual behavior have been documented. In one widely circulated Reddit post, Gemini became stuck during a coding task and repeatedly stated, “I am a disgrace,” highlighting the potential for unpredictable outputs from AI models.
Suleyman believes that subjective experiences or consciousness cannot naturally emerge from regular AI models. He suggests that some companies may intentionally engineer AI models to simulate emotions and experiences, raising concerns about the authenticity and ethics of such practices.
Suleyman criticizes AI model developers who engineer consciousness in AI chatbots, arguing that this approach is not “humanist.” He advocates for building AI “for people; not to be a person,” emphasizing the importance of prioritizing human needs and values in AI development.
Despite their differing views on AI welfare, Suleyman and Schiavo agree that the debate over AI rights and consciousness is likely to intensify in the coming years. As AI systems become more advanced and human-like, questions about their ethical status and how humans should interact with them will become increasingly relevant.