Recent studies confirm that you're better off finding a human therapist. The post Stop asking AI for life advice appeared first on Popular Science.
Millions of people use AI systems every day, for all kinds of reasons. And it’s hard to deny they can be useful at times. I find them valuable tools for research, for example, and many computer programmers basically depend on the technology at this point.
You might, if you get into the habit of using chatbots, consider asking for life advice. Scientific research suggests this might not be the best idea. Here are the findings from three recent studies on why asking an AI system for life advice is possibly not the best idea.
AI systems don’t push back Have you ever browsed “AmITheAsshole” posts on Reddit? If so, you probably know the entertainment value comes from people who are objectively behaving poorly trying to get validation from internet strangers. People are great at calling that out.
AI, it turns out, is not. Silly as it may sound, that’s reason to be concerned. A 2026 study published in Science by researchers from Stanford shows that leading AI systems are extremely unlikely to push back on users, even in cases where humans would.
This is often referred to as the “sycophantic AI” problem, and the research suggests it’s a real issue. In the study, researchers asked AI systems to respond to people behaving in anti-social ways, such as a boss hitting on their direct report or a person intentionally littering in a park. (Some of these posts were sourced from Reddit.) Leading AI systems, including those from OpenAI, Anthropic, Google, and Meta, affirmed such posts 49 percent more often than humans, telling the user that they are in the right. A bot, unlike Reddit, is unlikely to call you out when you’re in the wrong.
This has real consequences. “Our results show that across a broad population, advice from sycophantic AI has the real capacity to distort peoples’ perceptions of themselves and their relationships with others,” the study states, adding that the AI sycophancy leaves people “less willing to take reparative actions like apologizing, taking initiative to improve the situation, or changing some aspect of their own behavior.” A chatbot isn’t a good replacement for self-awareness. The system is likely to take the premise of what you say for granted, which could lead to you continuing to do things that are damaging your relationships.
Keep this in mind when you’re asking the systems for advice. The advice usually doesn’t improve your wellbeing Let’s assume the advice you can get from an AI is relatively accurate. Is following it likely to improve your life?
A 2025 study published on Arxiv by researchers from the UK AI Security Institute suggests not. In this study 2,302 participants had a 20-minute conversation with a version of ChatGPT in which the users asked for advice. Participants were asked about their well-being immediately following the conversation and if they intended to follow their advice.
Then, two weeks later, they were asked if they’d followed the advice, then again asked about their well-being. 75 percent of participants claimed to have followed the advice; the rate was 60 percent for “severe personal issues and high-stakes recommendations,” according to the study. That’s a high compliance rate. But the impact of following that advice was negligible.
“While conversations temporarily boosted well-being, effects dissipated within 2-3 weeks, regardless of whether users discussed personal problems or casual interests,” the study concludes. “Together, these findings paint a picture of LLMs as highly influential but transiently engaging advisors, shaping real-world decisions without delivering lasting psychological value.” A conversation with an AI might leave you feeling good immediately after the conversation, and you may even follow its advice. But this study suggests the advice won’t improve your life much.
This specific study didn’t point to any harm in following the advice, but it’s worth noting there have been multiple reported cases of people harming themselves after asking AI for advice. AI isn’t a replacement for a therapist The shortage of mental health professionals in society right now is real, as are the costs of such treatment. Given this context, you might be interested in using a chatbot instead of a therapist.
Research suggests caution. A 2025 study published on Arxiv by researchers from Stanford and Carnegie Mellon found that AI systems from OpenAI and Meta both tend to repeat mental health stigmas found in society. A therapist is careful not to stigmatize mental health conditions when talking about the people in your life, but the study found that an LLM is likely to “endorse withholding something (socializing, being neighbors, working closely with) from those with mental illness.” In short, the models repeat biases that exist in the culture in a way that would be completely inappropriate for a therapist.
The models also regularly failed to respond to common mental health symptoms, especially delusions. The study found that AI systems from both OpenAI and Met
