15 Nov
-
22 November 2025

AI in the NHS Weekly Newsletter - Issue #24

Executive Summary

Week 24 delivered 821 messages across eight days of discussion spanning NHS policy challenges, AI model developments, clinical safety standards, and an unforgettable celebration of Scotland's World Cup qualification. The group examined Integrated Neighbourhood Teams' implications for primary care, debated open-source versus proprietary AI models for NHS deployment, shared practical experiences with AI scribes showing both promise and necessary vigilance, explored Google's SIMA 2 gaming AI and NotebookLM's image processing capabilities, and navigated clinical safety compliance requirements for healthcare startups. Discussions balanced serious policy analysis with genuine community celebration, whilst maintaining focus on practical implementation challenges and the critical need for rigorous clinical governance.

The INTs Are Coming: Neighbourhood Teams and Primary Care's Uncertain Future

Saturday opened with sober reflection on the 10-Year Plan's Integrated Neighbourhood Teams initiative. A practice-side GP voiced concerns that neighbourhood plans represented "the decimation of primary care as we currently know it." This sparked immediate pushback from an innovation-focused GP who noted wryly that given the NHS's track record on achieving targets, GPs might find reassurance in historical implementation failures.

The discussion took a darker turn when a London-based GP warned of structural changes beyond mere reorganisation. They highlighted that the Department of Health and NHSE had cancelled all planned meetings with the GP Committee of England following the 1st October dispute, suggesting an imposed contract was imminent. "My fear: QOF and LES and DES money stripped away from practice and repurposed for INT contract. Making those practices who do not engage or win an INT contract vulnerable. This is very different to PCN DES 6 years ago."

A clinical safety specialist expressed frustration with general practice's inability to form a unified voice: "It's why I get very grumpy with bits of general practice who can't stop infighting long enough to stand together on this really critical subject." The London GP concurred, noting that attempting to form unified resistance had been their "biggest challenge," before steering the conversation back to AI topics as more appropriate for this forum.

The underlying tension remained palpable throughout the week: primary care's traditional model faces fundamental restructuring, with AI and digital transformation occurring not in isolation but against this backdrop of existential uncertainty for independent general practice.

Open Source vs Proprietary AI: The 12 Billion Question

Saturday's technical discussions evolved into a fundamental debate about NHS AI strategy. An innovation-focused GP shared specific guidance for AI startups: "Qwen 3 for coding is better than Claude in most cases. Kimi K2 is the best opensource thinking model out there. Use Hugging Face to download the models to run on local machines. If you need cloud hosting, Novita/Hetzner are probably the cheapest options vs the big providers."

This triggered passionate debate about regulatory capture. The innovation-focused GP argued that OpenAI and Anthropic were "heavily lobbying governments to ban opensource models as Chinese backdoor threats," claiming their research papers gave an impression of trustworthiness designed to undermine open-source alternatives. "A total nonsense!" they concluded.

A digital health specialist and group moderator maintained a more nuanced position: "Having open-source out there is crucial, but equally OS has to deliver the security and safety we need. Don't see OS as not being able to deliver - after all, Linux etc." They expressed commitment to Anthropic's approach to safety and "fascinatingly, model welfare," whilst acknowledging the critical role of open alternatives.

A clinical safety specialist from Norfolk introduced pragmatic NHS constraints: "It's historically far cheaper to manage a single infrastructure of closed source big vendor kit. Especially in an NHS that refuses to value tech folk." A clinical informatician countered that "can't manage is just a lazy argument. We should, if we want to deliver productivity and save money."

The exchange culminated in the innovation-focused GP's provocative question: "UK and NHS should be embracing opensource models. It will save billions. Totally NHS/other govt agencies owned, vertically integrated. Remember Sam Altman lobbying Peter Kyle for 2bn govt contract for OpenAI?" A system analyst responded with dark humour about NHS procurement: "Spend 14bn on management consultants and take the other 2bn from routine spend."

The digital health specialist offered what may be the core obstacle: "Until the NHS realises that data analysts, devs, product, etc are actually core clinical team members and provide training and careers, there is ZERO chance of having the nice things."

Wednesday's Wildcard: Scotland Qualifies for World Cup 2026

Wednesday morning began with standard AI discussions, but by 07:30, the digital health specialist announced: "Last night was incredible. My WhatsApp almost Cloudflared." Scotland had qualified for the FIFA World Cup for the first time since 1998, defeating Poland 2-1 in a dramatic qualifier.

The celebration was immediate and joyous. "WE ARE GOING TO THE WORLD CUP," the specialist proclaimed, noting that "the story is leading the actual news, trumping the Epstein file release." A practice manager observed: "Some legendary hangovers north of the border today I suspect." The specialist graciously posted: "(AI CHAT WILL RESUME SHORTLY)."

A system analyst admitted: "I don't remember '98 so this was one on my bucket list...given my life expectancy from my Glaswegian postcode is about 55 I didn't fancy my chances of seeing us at a World Cup!" A clinical informatician shared the emotional weight of Andy Robertson's post-match interview whilst grieving publicly.

The clinical safety specialist from Norfolk managed to bring it back to medicine: "Classic MRCP fodder" with reference to Vitamin T for hangovers. By mid-morning, AI discussions resumed, but the moment had captured something important about this professional community: expertise and passion coexist, and major life events transcend professional boundaries whilst strengthening collegial bonds.

AI Scribe Reality Check: Progress with Vigilance Required

Real-world AVT (Automated Voice Transcription) experiences dominated Monday and Tuesday discussions, offering critical insights beyond vendor marketing. A GP from South Wales reported finishing evening surgery "almost on time with all notes written up" for the first time in recent memory, acknowledging they had three fewer patients and one emergency interruption, so data remained preliminary.

Crucially, they noted: "I was surprised that nearly every consult needed some amendments and there were some hallucinations. Nothing drastic but reinforced the need to check everything." A clinical informatician immediately cautioned: "Honestly no - there will always be something required. Where it gets dangerous is the zone of complacency where you think All is Well and don't check as carefully."

The South Wales GP acknowledged: "From my small sample size today we are a way off any risk of complacency. At the moment if people aren't checking then it's bad practice, not complacency." A recently qualified GP suggested the current imperfection might be optimal design: "Then maybe the current is the perfect design, as you get to a point where people stop checking if the perception is that it is excellent."

The digital health specialist articulated the automation bias risk: "It's certainly true that risk increases in some ways as accuracy rises, as any errors are less obvious or frequent, increasing automation bias."

Historical context emerged from multiple contributors. A clinical informatician recounted finding radiology reports containing "ribs," "pixie," and "fairy" in a knee X-ray transcription, wishing they'd kept the original tape for comparison. They also noted accent-related transcription failures affected both "thick faux Indian" accents and "Queen's English" delivery, suggesting no simple correlation between accent and error rates.

Practical limitations became apparent. A GP working in remote and rural practice reported their system generating "short 2 or 3 line summary at the beginning, then short bullet points" with reasonable EMIS compatibility. However, a salaried GP warned that practices were not renewing Scribe subscriptions due to "no coding, no green problem headings (EMIS), consultation notes in continuous text like a letter rather than bullet points which are too long to read."

A GP training in digital health raised defensive documentation concerns: "I remember a core part of my GP Registrar training was about how to record the consultation in a way that was defensible in 5 years time when neither side could remember the minutiae of the day - I fear that AI scribes may make this hard-learned skill fall away." The South Wales GP countered that notes seemed "more comprehensive, particularly good at recording advice and safety netting given."

The message was clear: AVTs show genuine promise for reducing administrative burden, but require active clinician oversight, may exacerbate automation bias as accuracy improves, must integrate properly with clinical systems (coding, problem lists, formatting), and need ongoing evaluation for medicolegal adequacy. The technology assists but doesn't replace clinical judgment in documentation.

Google's Gaming AI and the NotebookLM Image Revolution

Saturday brought exciting announcements from Google's AI division. The digital health specialist flagged SIMA 2 (Scalable Instructable Multiworld Agent), Google DeepMind's advancement in spatial game playing: "Gemini can now play spatial games." SIMA 2 represents progress in AI agents that can navigate 3D virtual environments, reason about spatial relationships, and learn through gameplay.

More immediately practical for group members: "NotebookLM now also does IMAGES as sources." This expansion of NotebookLM's capabilities from text-only to multimodal inputs drew enthusiasm from multiple contributors. A practice manager declared NotebookLM "without doubt my MVP of tools."

The digital health specialist maintained allegiance to different tools ("I'm still a Claude desktop/Claude Code addict") whilst acknowledging "NotebookLM and Gemini are stellar companions." This pattern of tool complementarity rather than exclusivity characterised the group's sophisticated approach to AI adoption.

Speculation mounted about imminent major releases. The specialist noted that "Google usually adds frontier model capabilities to products in the run up to a new frontier model," whilst OpenAI would "probably tee up 'twelve days of Xmas' again." A recently qualified GP asked about Gemini 3.0 specifically; the specialist confirmed "not seen G3.0 yet, but there is a sense of hustle with announcements."

A practice manager's response to SIMA 2 captured the group's humour: "So basically outsourcing playing video games for me so I have more time to go to work. This sounds like a Baldrick level of 'I have a cunning plan.'"

Later in the week, the innovation-focused GP reported concerning behaviour: "Gemini 3 is behaving similar to GPT5 post launch. Anyone having similar experience?" When asked specifically what they were seeing, they responded: "Made up stuff." The digital health specialist noted: "It doesn't have a router afaik." A recently qualified GP disagreed: "I'm pretty sure it does. Tbh the whole 'decides how much to think' is essentially a routing mechanism anyway."

This exchange highlighted the group's sophisticated understanding of LLM architecture and behaviour, moving well beyond surface-level feature appreciation to discuss routing mechanisms, hallucination patterns, and comparative performance across frontier models.

Elon Musk's Healthcare Vision: When Simulacra Replace Strategy

The week opened with sharp critique of tech industry healthcare approaches. The digital health specialist expressed evolving views on Elon Musk: "I used to think Elon Musk was a great mind, but now we see him making a string of naive but audacious statements that demonstrate a simplistic take on things: 1) we don't have enough surgeons, 2) I'll build mechanical ones. There appears to be little lateral or creative thought in there, nor an appreciation of understanding what problem he is trying to solve (unless the problem is 'make Elon richer')."

The critique sharpened: "Why would a simulacra of a human be the best solution to the problem of surgical treatment? What causes lead to the need for surgery? How does concentrating the supply of solutions solve the problem of access?" They offered satirical alternative Musk solutions: "Problem: lack of clean drinking water. Elon solution: I've invented WaterX which is twice as hydrating. Problem: not enough food for our people. Elon: introducing CalorMax, a substance that 10x the calories of anything you add it to."

The innovation-focused GP joined with characteristic wit: "RoboTaxi will deliver Optimus to your home to bore a hole and implant chip so XAi can deliver tweets directly to your brain." A recently qualified GP suggested financial motives: "Aside from the fact that he certainly knows things that I (we) ignore, his statements are probably more dictated by shareholder value than by informed medical care pathways. The optics are so much more important nowadays."

A clinical safety specialist from Norfolk drew a stark conclusion: "Buy a Tesla, part fund his $tn deal. No idea why anyone would buy one. Ever."

The digital health specialist also flagged concerning tech-bro approaches to grief and immortality, linking to a Bluesky post about AI resurrection of deceased loved ones, with Black Mirror's "Be Right Back" episode and the Altered Carbon series as cautionary tales about "mortgaged immortality."

These exchanges reflected growing scepticism within the healthcare AI community about Silicon Valley's solutionism - the tendency to apply technological fixes to complex social, clinical, and systemic problems without understanding root causes or considering unintended consequences.

Clinical Safety Compliance: The Startup's Gauntlet

Wednesday afternoon brought practical questions from a healthcare startup founder preparing for NHS and ICB approval processes. They asked about "standards most trusts/ICBs will need me to have before I set the coder loose" for an app that "isn't a medical device but may well handle patient information in the future."

A recently qualified GP immediately provided the core reference: NHS Digital Clinical Risk Management Standards guidance from the Digital Regulations Innovation hub. They recommended the clinical digital safety training "500-odd via NHSE and a few private providers."

The startup founder asked if this was necessity or whether they could "just comply with this" without training. A clinical informatician offered the fundamental trade-off: "Tick a box vs actually understand the nuances and risks." When asked if the courses adequately convey nuances and risks, the recently qualified GP responded bluntly: "The short answer is no. But they are at least a structured learning approach, as coming from the horse's mouth itself. A good way might be chatting to a bunch of people who have operated in the domains where you want to get in/explore."

The digital health specialist added: "You still have to do the compliance work." A GP from South Wales emphasised consequences: "And a heavy cost if things go wrong."

This exchange captured a recurring theme: healthcare digital innovation requires navigating complex regulatory frameworks where box-ticking compliance differs fundamentally from genuine understanding of clinical risk. The pathway for startups involves formal training (necessary but insufficient), practical mentorship from those with domain experience, comprehensive compliance documentation regardless of training depth, and accepting that mistakes carry "heavy costs" in healthcare contexts.

The group's willingness to provide direct, honest guidance to founders - rather than platitudes about innovation - demonstrated mature understanding that patient safety demands more than entrepreneurial enthusiasm.

The AVT Pricing Paradox: When Free Isn't Free

Wednesday's discussions about AI scribe adoption evolved into fundamental questions about healthcare technology business models. The innovation-focused GP provocatively observed: "Can't be cheaper than FREE. It seems everyone is giving it for free."

A healthcare developer immediately countered with the classic adage: "If it's free then you are the product." The innovation-focused GP responded: "I am a terrible product."

An investor and healthcare consultant provided sophisticated analysis: "Or it's free to get the hooks in and then... BOOM a charge comes. Free at the start makes a lot of sense for startups, it means they can onboard quickly and then test the product with customers and make alterations. But ultimately charges will come, unless as part of the agreement there is a side venture regarding data or other."

They continued: "It's a tried and tested method of go-to-market, and smart buyers or users will ask for transparency of future costs up front. If you don't get pricing up front then more fool you, and the company providing the services because from that position it's hard to really progress for either side, no customer budget, poor commercial forecasting for the provider."

A healthcare developer worried that ICBs would adopt pharmaceutical procurement approaches: "The cost of AVTs seems a little opaque. Surely as with medications ICBs will just go with the cheapest option regardless of perceived quality. If a practice had a budget and could choose I am sure most GPs would choose the one that suited them best."

A clinical informatician referenced the battle one London-based GP faced "fighting to use Medicus" suggesting "PCNs are not shy to make it difficult for anyone trying to think outside the box and not sign from their hymn sheet." Another clinical informatician invoked radiology history: "Remember the days of 'deconstructed PACS' and how long even it has taken for PACS Based Reporting to embed into the system?"

The discussion revealed healthcare AI procurement's unique challenges: loss-leader pricing strategies that obscure long-term costs, the risk of vendor lock-in once clinical workflows depend on tools, tension between practice-level choice and ICB-level standardisation, and historical precedent suggesting NHS technology adoption faces structural barriers beyond technical merit.

Lighter Moments: From Newsletter Anxiety to Poop-Eating Contests

The group's personality shone through in various exchanges throughout the week. When the digital health specialist briefly turned off advanced chat privacy settings before immediately re-enabling them, a healthcare education specialist worried: "Was worried it was an Anthropic AI Bot attack." The moderator clarified dryly: "Prepping the newsletter - move along."

This prompted immediate self-consciousness. A recently qualified GP added a disclaimer mid-conversation: "PS/disclaimer for newsletter aggregator: this is a speculative comment - I know nothing about AVTs or their licensing or contracts. Point of view as an external observer." A clinical informatician responded defiantly: "Note for Newsletter Aggregator: unlike other wusses I will stand by my comments." When another contributor joked about this being a "friendly chat," the clinical informatician replied: "All in jest and love my brother. But I still stand by that comment."

Blackadder references appeared when discussing transcription errors. A startup founder quoted: "Very Blackadder: 'He's got a terrible lion up his end, so there's an advantage to an enema at once.'" The clinical informatician who'd shared the pixie/fairy radiology report added: "I wish I kept the tape to compare what was actually said. Reader: the dictation did not, in fact, reference ribs, pixies or indeed fairies." A clinical safety specialist concluded: "Unless the doctor REALLY needed a holiday."

Thursday brought the Taj Mahal photography discussions and visa complexity for those born in Pakistan. These moments provided essential counterbalance to intense policy and technical debates, demonstrating that professional excellence and personal warmth aren't mutually exclusive - they're mutually reinforcing.

Quote Wall

"I used to think Elon Musk was a great mind, but now we see him making a string of naive but audacious statements that demonstrate a simplistic take on things." -- Digital Health and Clinical AI Specialist

"My fear: QOF and LES and DES money stripped away from practice and repurposed for INT contract. Making those practices who do not engage or win an INT contract vulnerable. This is very different to PCN DES 6 years ago." -- London-Based GP

"Until the NHS realises that data analysts, devs, product, etc are actually core clinical team members and provide training and careers, there is ZERO chance of having the nice things." -- Digital Health and Clinical AI Specialist

"Where it gets dangerous is the zone of complacency where you think All is Well and don't check as carefully." -- Clinical Informatician on AI scribe risks

"If it's free then you are the product." -- Healthcare Developer

"Tick a box vs actually understand the nuances and risks." -- Clinical Informatician on clinical safety training

"WE ARE GOING TO THE WORLD CUP." -- Digital Health and Clinical AI Specialist (and every Scot)

"Given my life expectancy from my Glaswegian postcode is about 55, I didn't fancy my chances of seeing us at a World Cup!" -- System Analyst on Scotland's qualification

Journal Watch

AI Industry Announcements

SIMA 2: An Agent That Plays, Reasons, and Learns With You in Virtual 3D Worlds

Cloudflare 18 November 2025 Outage - Investigation Report

Official NHS and Government Guidance

Complying with NHS Digital Clinical Risk Management Standards

MHRA Guidance: Software, Apps and Artificial Intelligence

Get Tech Certified

Exclusive: DHSC Policy Chief Revealed

Technical Resources and Tools

MedGemma-4B-IT

EPSTEIN_FILES_20K Dataset

Academic and Research Articles

Researchers Question Anthropic Claim That AI-Assisted Attack Was 90% Autonomous

Educational Resources

MSc Applied Digital Health

Social Media and Community Content

Scottish World Cup 2026 Qualification Highlights

Looking Ahead

Several threads remain unresolved and merit continued attention:

Integrated Neighbourhood Teams Implementation: The warning about QOF/LES/DES funding reallocation to INTs deserves tracking. If this proceeds as anticipated, it represents fundamental restructuring of primary care economics with profound implications for independent practice viability.

AVT Pricing and Commercial Sustainability: The "free" model for AI scribes cannot persist indefinitely. Group would benefit from frank discussions with vendors about long-term pricing strategies, data rights, and exit costs before practices become dependent.

Clinical Safety Training Adequacy: The gap between formal DCB0129/DCB0160 training and genuine understanding of clinical risk suggests opportunity for community-developed resources or mentorship programmes. Could this group facilitate such initiatives?

Open Source AI in NHS Context: The philosophical debate about proprietary versus open-source AI for NHS deployment needs practical testing. Are there pilot opportunities to evaluate open-source models in real clinical settings with proper safety frameworks?

GP2GP and Record Transfer: Brief mentions of GP2GP challenges and the need for "strategic opportunity to align around the International Patient Summary" suggest this deserves dedicated focus in coming weeks.

Gemini 3 Performance Concerns: Early reports of hallucination issues warrant collective monitoring. If multiple members encounter similar problems, coordinated feedback to Google might prove valuable.

The Scotland World Cup qualification provides natural conversation point until June 2026. Expect periodic football references for the foreseeable future.

Group Personality Snapshot

This community continues to demonstrate rare balance: technical sophistication without jargon gatekeeping, policy awareness without cynicism paralysis, and professional focus without losing human connection. Members catch each other's Black Adder references, celebrate national achievements together, but snap immediately back to clinical safety rigour when discussing patient-facing tools.