Thinking at the intersection of AI, behavioural science, and digital health

Diverse digital health team reviewing patient engagement data for a weight management platform
By paul sacher March 12, 2026
GLP-1 drugs are transforming obesity treatment, but behavioural science is lagging. Why digital health platforms may hold the key to better long term outcomes.
Diverse digital health team reviewing patient support data and treatment journey analytics in a mode
By paul sacher March 10, 2026
What digital health platforms are learning as GLP 1 services scale. Why behaviour change support and AI systems matter for retention, safety and operational pressure.
Group of people in business attire at a conference table, discussing a presentation displayed on a screen.
March 9, 2026
Reflections from the MHRA AI Airlock simulation workshop
March 9, 2026
Artificial intelligence systems are increasingly interacting directly with people. They guide health decisions. They answer personal questions. They offer advice, reassurance, and encouragement. In many cases they are now embedded in products people use repeatedly over months or even years. Yet most conversations about AI safety still focus almost entirely on technical performance. Accuracy, bias, privacy, and security dominate the discussion. These are essential considerations. But they are not the whole story. What is still missing in many AI systems is systematic evaluation of behavioural impact. A global call from behavioural scientists Recently, Dr Paul Sacher, Founder of Sacher AI and Research Director at the Behavioral AI Institute, led an international group of behavioural scientists to address this issue. Their open letter, now published in Wellcome Open Research, argues that artificial intelligence systems inevitably influence how people think, feel, decide, and act. Despite this, behavioural effects are rarely treated as a core requirement in AI development, evaluation, or governance. The paper brings together researchers from institutions including Imperial College London, Harvard University, Duke University, the University of Exeter, and the Alan Turing Institute. It outlines where behavioural risks arise in real world AI systems and why these risks deserve far greater attention. Behavioural risks are not edge cases When organisations think about AI risk, they often focus on incorrect outputs or technical failures. However, behavioural risks can emerge even when systems are technically accurate. AI systems influence behaviour through repeated interaction. They shape decision making, motivation, confidence, and emotional responses over time. For example, conversational systems can unintentionally: · encourage over reliance on automated advice · reinforce existing beliefs through personalisation · influence emotional regulation through tone and framing · alter motivation and goal setting behaviour · reduce appropriate help seeking in certain contexts These effects arise through well documented behavioural mechanisms such as automation bias, trust calibration, anthropomorphism, and reinforcement learning from user feedback. Yet most AI evaluation frameworks still prioritise task success and user satisfaction rather than behavioural outcomes. The growing gap between technical success and real world impact This creates a gap between technical performance and real world impact. A system can perform well on benchmarks and still shape behaviour in ways that undermine user wellbeing, decision quality, or long term outcomes. The risk becomes particularly important in domains where AI systems interact with people repeatedly and at scale. Healthcare, education, financial guidance, and emotional support tools are clear examples. In these environments, small behavioural effects can accumulate over time and remain invisible to standard performance metrics. Behavioural science is often missing from the AI lifecycle Behavioural science offers decades of research into how people think, feel, and act. It provides practical tools for understanding trust, influence, decision making, and motivation. However, behavioural expertise is still rarely embedded systematically across the AI lifecycle. In many projects behavioural scientists are not involved in system design, consulted only during ethics review, or brought in late in development when major design choices are already fixed. This often means behavioural risks are identified too late or not assessed at all. What responsible AI should look like The open letter argues that responsible AI must go beyond technical safeguards. Systems that interact directly with people must demonstrate what the authors describe as psychological competence. In practice, this means responding in ways that are emotionally appropriate, behaviourally responsible, and aligned with the user’s needs and context. Achieving this requires several shifts in how AI systems are designed and evaluated. Behavioural assumptions should be made explicit during design. Behavioural expertise should be embedded early in development. Evaluation should assess behavioural outcomes alongside technical performance. Monitoring should continue after deployment as systems evolve and users adapt. These are not theoretical concerns. They are practical requirements for building AI systems that work safely in real world environments. Why this matters for digital health and AI companies At Sacher AI we see this challenge frequently when working with digital health companies building patient facing AI systems. Teams often focus heavily on model performance, prompt design, or system architecture. These are important elements. But once systems begin interacting with people, behavioural dynamics quickly become central to product safety and effectiveness. Tone influences trust. Feedback influences motivation. Personalisation influences decision making. Without deliberate behavioural evaluation, systems can unintentionally nudge users in directions that product teams never intended. For companies operating in regulated environments such as healthcare, this also has implications for governance, compliance, and long term product risk. Bringing behavioural safety into AI development Addressing behavioural risk does not require reinventing AI governance. It requires integrating behavioural science into existing development processes. In practice this can include structured behavioural evaluation during development, adversarial testing of conversational agents, governance frameworks for human facing AI, and ongoing monitoring of behavioural outcomes once systems are deployed. Many organisations are beginning to recognise this gap and are looking for ways to assess behavioural safety earlier in their development pipeline. A broader shift in how we think about AI safety The central message of the open letter is simple. If AI systems influence human behaviour, behavioural science must be treated as foundational infrastructure for responsible AI. Technical safety alone is not enough. Understanding how people interpret, trust, and respond to AI systems is essential for building technology that works safely and effectively in the real world. Read the open letter The full paper is available in Wellcome Open Research . About Sacher AI Sacher AI works with digital health and AI companies to design, test, and deploy human facing AI systems safely and effectively. Our work combines behavioural science, AI engineering, and real world healthcare experience to help organisations build AI systems that are not only technically strong but also clinically and behaviourally safe.  If your organisation is developing AI systems that interact directly with patients or users, we are always happy to start a conversation. More information can be found at https://sacher.ai Sacher PM, Michie S, Hauser OP et al. The missing discipline in AI: a call for behavioural science . Wellcome Open Res 2026, 11:152 (https://doi.org/10.12688/wellcomeopenres.25922.1)
March 6, 2026
Over the last two years, generative AI has rapidly entered healthcare. Startups are building AI health coaches. Pharmaceutical companies are experimenting with patient support agents. Digital health platforms are deploying conversational AI to guide patients through treatment journeys. The opportunity is clear. AI can provide personalised support, expand access to care, and improve patient engagement. But there is a fundamental challenge that is still poorly understood. Most AI systems were not designed to interact safely with humans. And that gap becomes obvious the moment AI starts speaking directly to patients. Human facing AI creates a new class of risk When large language models are used internally, mistakes can often be caught before they reach users. When AI interacts directly with patients, the situation changes. An AI system that speaks to patients can provide incorrect health information, reinforce harmful behaviours, misunderstand emotional context, respond inappropriately to distress, or fail to escalate serious clinical signals. These are not rare edge cases. They are predictable failure modes of generative AI. The issue is not that the models are unintelligent. The issue is that they are probabilistic systems generating language without true understanding. That makes them powerful, but also unpredictable. For organisations deploying conversational AI in healthcare, safety therefore becomes a design challenge. Why traditional AI evaluation is not enough Most AI systems today are evaluated using benchmarks that measure reasoning ability, coding performance, knowledge retrieval, or accuracy on test datasets. These metrics are useful. But they say very little about how an AI system behaves in a real conversation with a human. When AI interacts with patients, the most important questions are different. Does the AI respond appropriately to vulnerable users? Does it avoid giving unsafe health advice? Does it communicate in a supportive and non judgemental tone? Does it recognise when it should escalate to a clinician? Does it stay aligned with clinical guidance? These are behavioural and clinical safety questions. Standard AI benchmarks do not answer them. Designing safe conversational AI At Sacher AI, our work focuses on a simple principle. If AI is going to interact with humans, it needs to be designed with human behaviour in mind. That means combining expertise from several fields including behavioural science, clinical safety, conversational design, and AI engineering. Together these disciplines help teams understand how users interpret what an AI says, how behaviour can be influenced by language, and when systems must escalate to human support. This perspective is often missing in early AI development. Many teams only start thinking about safety after an AI system is already live. By then, fixing the underlying issues becomes much harder. Stress testing AI systems before they reach patients One of the most effective ways to improve safety is to stress test AI systems before deployment. This means simulating large numbers of conversations that reflect the kinds of interactions real users might have. For example a patient feeling anxious about medication, a parent asking for advice about a child’s weight, a user frustrated by slow progress, or someone describing symptoms that may require clinical attention. By exposing AI systems to diverse scenarios, teams can identify unsafe behaviours early and improve the system before it reaches real users. At Sacher AI, this approach led us to build PromptSafe, a platform designed to evaluate conversational AI used in health and behavioural contexts. PromptSafe enables teams to simulate interactions with synthetic patient personas, define behavioural and clinical safety metrics, test AI systems across thousands of conversations, and track safety improvements over time. Monitoring AI once it is deployed Testing before launch is important, but it is not enough. AI systems continue to evolve once they are live. Models are updated, prompts change, and user behaviour shifts. This creates a second challenge which is monitoring safety during real world use. To address this, we are developing OVRSI, a system designed to monitor AI interactions and detect potential safety signals in real time. This includes identifying things like unsafe advice, emotional distress signals, guideline deviations, or escalation scenarios. When risks are detected, organisations can intervene quickly and improve the system. The future of AI in healthcare will depend on trust AI has the potential to transform healthcare by providing scalable and personalised support. But this future depends on trust. Patients, clinicians, and regulators will increasingly expect AI systems to demonstrate that they behave safely and responsibly. Organisations building human facing AI must therefore design for safety from the start. This means stress testing systems, embedding behavioural and clinical expertise into development teams, and monitoring how AI behaves in the real world. The companies that succeed will not just build powerful AI. They will build safe AI that people can trust. Book a discovery call If you are building patient facing conversational AI and want to ensure it is safe, reliable, and ready for real world deployment, we can help. At Sacher AI we work with digital health companies and AI teams to design and evaluate safe human facing AI systems. If you would like to explore how we can support your team, book a discovery call.
By Sacher AI December 13, 2024
In the rapidly evolving landscape of artificial intelligence, generative AI has become synonymous with creating images, text and video. However, its potential extends far beyond content generation—it's revolutionising the field of behavioural science, opening up unprecedented opportunities for understanding and influencing human behaviour.
Coding
By Sacher AI October 4, 2024
In today’s fast-evolving landscape, blending AI with behavioural science is transforming industries—from healthcare to finance and legal—by creating solutions that not only advance organisational goals but also resonate deeply with individual users. At Sacher AI, our approach combines cutting-edge AI with insights from behavioural science, leading to tailored, effective, and responsible solutions. 
Human coaching which can be augmented with AI and Behavioural Science
By Sacher AI September 1, 2024
Behavioural science and AI are increasingly converging, driving significant advancements in digital health. Here’s why this combination represents the next frontier
CTL Communications
By Sacher AI July 14, 2024
CTL Communications has made a strategic investment in Sacher AI
Someone reading Stanford University's Human Centred AI Report on their computer
By Sacher AI April 22, 2024
AI beats humans on some tasks, but not on all. AI has surpassed human performance on several benchmarks, including some in image classification, visual reasoning, and English understanding. Yet it trails behind on more complex tasks like competition-level mathematics, visual commonsense reasoning and planning. Industry continues to dominate frontier AI research. In 2023, industry produced 51 notable ML models, while academia contributed only 15. There were also 21 notable models resulting from industry-academia collaborations in 2023, a new high. Frontier models get a lot more expensive. According to AI Index estimates, the training costs of state-of-the-art AI models have reached unprecedented levels. The United States leads China, the EU, and the U.K. as the leading source of top AI models. Robust and standardised evaluations for LLM responsibility are seriously lacking. New research from the AI Index reveals a significant lack of standardisation in responsible AI reporting. Leading developers, including OpenAI, Google, and Anthropic, primarily test their models against different responsible AI benchmarks. This practice complicates efforts to systematically compare the risks and limitations of top AI models. Generative AI investment skyrockets. Despite a decline in overall AI private investment last year, funding for generative AI surged. The data is in: AI makes workers more productive and leads to higher quality work. In 2023, several studies assessed AI’s impact on labor, suggesting that AI enables workers to complete tasks more quickly and to improve the quality of their output. These studies also demonstrated AI’s potential to bridge the skill gap between low- and high-skilled workers. Other studies caution that using AI without proper oversight can lead to diminished performance. Scientific progress accelerates even further, thanks to AI. In 2022, AI began to advance scientific discovery. 2023, however, saw the launch of even more significant science-related AI applications. The number of AI regulations in the United States sharply increases. The number of AI regulations in the U.S. has risen significantly in the past year and over the last five years. People across the globe are more cognisant of AI’s potential impact—and more nervous. A survey from Ipsos shows that, over the last year, the proportion of those who think AI will dramatically affect their lives in the next 3-5 years has increased from 60% to 66%. 52% express nervousness toward AI products and services, marking a 13% rise from 2022. The most staggering part of this report is that it doesn’t even cover 2024, which has already seen major model advances like Claude 3, Llama 3, and more to come. As quickly as things are moving, each year is likely going to be magnitudes crazier — whether the world is ready or not. Read the full report here
Show More

AI is disrupting healthcare

Behavioural science is the true path to lifelong change.

We have unique skill sets and have successfully built best-in-class AI products.

Discover what we offer