View in Browser

PythRaSh's AI Newsletter

Issue #30  |  Week of May 6, 2026

A sincere apology: We missed the last two editions — April 22 and April 29 — due to ongoing technical difficulties. We are truly sorry for the interruption and deeply appreciate your patience. We're back this week with a packed issue covering the most significant AI + Biology developments from the past two weeks. Thank you for sticking with us!

Hi, there! This fortnight's theme is AI Enters the Life Sciences Mainstream. OpenAI launched its first biology-specific reasoning model named after Rosalind Franklin. A landmark JAMA study revealed that all 21 leading LLMs fail more than 80% of early differential diagnoses. Clinical AI funding hit new highs, and new arXiv papers are proving that bigger models are not automatically safer in clinical contexts. A lot happened across two weeks — let's dive in.

🚀 EVENT OF THE WEEK

OpenAI Launches GPT-Rosalind — The First Frontier AI Model Built for Life Sciences

On April 16, 2026, OpenAI unveiled GPT-Rosalind, a frontier reasoning model purpose-built for biology, drug discovery, genomics, and translational medicine — named after Rosalind Franklin, the British chemist whose X-ray crystallography work revealed the structure of DNA.

Unlike general-purpose models, GPT-Rosalind is optimised for multi-step scientific workflows: target discovery, pathway analysis, protein engineering, literature synthesis, and hypothesis generation. It combines improved tool use with deeper domain understanding across chemistry, protein biology, and genomics. Launch partners include Amgen, Moderna, the Allen Institute, and Thermo Fisher Scientific.

Simultaneously, Novo Nordisk announced it would deploy GPT-Rosalind across its entire R&D pipeline, manufacturing operations, and commercial functions — with full integration scaled by end of 2026, covering the full medicine lifecycle from discovery through to supply chain.

Why this matters: GPT-Rosalind signals that frontier AI labs are building biology into their core platforms — not merely partnering with pharma. For biology researchers, this is a step-change in the kind of AI assistance available for complex, multi-step scientific work.

Key takeaways:

  • First frontier AI model specifically designed for biological reasoning and drug discovery workflows
  • Named after Rosalind Franklin — a meaningful symbolic choice for the life sciences community
  • Novo Nordisk partnership signals immediate real-world industrial deployment at scale

⚡ Quick Updates

  • Aidoc Raises $150M Series E: The clinical AI company behind the CARE foundation model — deployed in nearly 2,000 hospitals and analysing 60M+ patient cases annually — closed a $150M round led by Goldman Sachs Growth Equity (General Catalyst, SoftBank Vision Fund 2, NVIDIA Ventures). Funds expand imaging modalities and add automated draft radiology report generation. Total funding now exceeds $500M. MedTech Dive
  • FDA Issues RFI on AI for Clinical Trials: On April 29, the FDA published a formal Request for Information seeking industry input on a pilot program to assess how AI technologies can improve efficiency, speed, and decision quality in early-phase clinical trials — a key regulatory step toward formalising AI's role in drug development. SBA Office of Advocacy
  • ALP Bio Raises €1.9M for AI-Organoid Platform: The Swiss startup combines human immune organoids with generative AI to predict and reduce antibody immunogenicity risks before expensive late-stage trials — addressing one of the costliest failure modes in biologic drug development. Lambda Biologics
  • UC Berkeley & UCSF Launch Voio — AI Radiology Startup: Researchers from the UC Berkeley/UCSF Joint Program in Computational Precision Health spun out Voio, building AI models that generate draft radiology reports and predict patient risk for cancer, osteoporosis, and heart failure years in advance. Berkeley News
  • FDA Approves First-Ever PROTAC Therapy: Pfizer and Arvinas received early FDA approval for Veppanu (ARV-471), the first-ever approved PROTAC protein degrader, targeting ESR1-mutated metastatic breast cancer — validating the structural biology pipeline where AI tools like AlphaFold are playing a growing role. Lambda Bio

📚 Top Research Papers

Haiku: Linking Spatial Biology and Clinical Histology via Tri-Modal Contrastive Learning

Authors: Yan Cui, Jacob S. Leiby, Wenhui Lei, Dokyoon Kim et al. | Published: April 30, 2026 | arXiv 2605.00925

Haiku is a tri-modal contrastive learning model trained on 26.7 million spatial proteomics patches from 3,218 tissue sections across 1,606 patients and 11 organ types. It links molecular, morphological, and clinical data in a shared embedding space — enabling three-way cross-modal retrieval (Recall@50 = 0.611), survival prediction (C-index 0.737, +7.91% over unimodal), and zero-shot biomarker inference (Pearson correlation 0.718 across 52 biomarkers) from text descriptions alone.

Spatial Proteomics Cancer Prognostics

SaFE-Scale: Safety and Accuracy Follow Different Scaling Laws in Clinical LLMs

Authors: Sebastian Wind, Tri-Thien Nguyen et al. | University Hospital Erlangen / TU Munich | Published: May 5, 2026 | arXiv 2605.04039

SaFE-Scale challenges the assumption that larger, more accurate clinical LLMs are automatically safer. Evaluating 34 LLMs across 6 deployment conditions on the RadSaFE-200 radiology safety benchmark, they found clean verified evidence reduced high-risk errors from 12.0% to 2.6% — while agentic RAG and max-context prompting left dangerous overconfidence elevated. Bigger does not mean safer: evidence quality, not scale, determines clinical AI safety.

Clinical Safety Radiology AI

Dual-Modal Lung Cancer AI: Interpretable Radiology + Microscopy with Clinical Risk Integration

Authors: Baramee Sukumal, Aueaphum Aueawatthanaphisut | Published: April 17, 2026 | arXiv 2604.16104

An interpretable AI system fusing CT radiology with H&E histopathology to classify five lung cancer categories — adenocarcinoma, squamous cell carcinoma, small cell lung cancer, and normal tissue — with 0.87 accuracy and AUROC > 0.97. Six explainability techniques (Grad-CAM, Grad-CAM++, Integrated Gradients, Occlusion, Saliency Maps, SmoothGrad) generate tumour localisations aligned with expert radiologist judgment. Grad-CAM++ showed the highest faithfulness.

Lung Cancer Interpretable AI

MARBLE: Multimodal Alignment for Genomic Biomarker Prediction in Computational Pathology

Authors: Multiple institutions (Stanford lead) | Published: March 2026 | arXiv 2603.00193

MARBLE aligns H&E histopathology representations with genomic biomarkers generated by LLMs and protein language models, learning features that generalise across cancer types and institutions. At inference time, no genomic sequencing is required — molecular biomarkers are inferred directly from routine pathology slides. Significant improvements in cross-cohort biomarker prediction over unimodal image-only baselines.

Precision Oncology Computational Pathology

💻 Top GitHub Repos of the Week

chemprop

⭐ ~2,400 stars | MIT | Python

The gold-standard neural network framework for predicting molecular properties (ADME, toxicity, bioactivity) using message-passing networks on molecular graphs. Widely adopted in pharma for early-stage compound screening and structure-activity relationship analysis. Actively maintained by the MIT/Chemprop team.

graphein

⭐ ~1,200 stars | Python

Builds graph representations of protein structures from PDB and AlphaFold outputs — enabling GNN-based analysis of protein-protein interactions, binding pockets, and drug-protein docking. An essential tool for structure-based drug design and protein engineering pipelines.

Therapeutics Data Commons (TDC)

⭐ ~1,200 stars | Harvard MIMS | Python

66+ curated ML-ready benchmark datasets spanning drug-target interaction, ADMET property prediction, molecular generation, and clinical outcome modelling. The standard infrastructure for fairly benchmarking drug discovery AI models across the research community.

torchxrayvision

⭐ ~1,100 stars | Python | Clinical AI

Unified library providing access to 9 major chest X-ray datasets (CheXpert, NIH, MIMIC-CXR, PadChest) plus pre-trained models for classification, segmentation, and feature extraction. Reduces clinical AI setup from weeks to hours — essential for pneumonia, atelectasis, and cardiomegaly detection research.

MedResearcher-R1

⭐ ~501 stars | Python | Agentic Medical AI

An agentic deep research model trained with knowledge-informed trajectory synthesis for medical scenarios. Enables autonomous synthesis of biomedical literature, clinical guidelines, and patient case data — supporting clinical decision support, medical education, and hypothesis generation at scale.

🛠️ Top AI Products of the Week

Open Wearables

624 upvotes | Open Source & Self-Hosted | Health Infrastructure

A fully open-source, MIT-licensed platform providing a single API for every wearable device (Whoop, Oura, Apple Watch, Garmin). Includes open health scoring algorithms and AI-ready structured context for building personalised health products — metabolic tracking, recovery analytics, clinical monitoring — with full self-hosting for data privacy compliance.

Plurai

729 upvotes | AI Agent Reliability | Developer Tools

Vibe-training for AI agent reliability: describe what your agent should and should not do, and Plurai auto-generates training data, validates it, and deploys a custom guard model in minutes — no annotation pipeline needed. Built on BARRED research; 8× lower cost than GPT-as-judge with 43% fewer failures. Critical for healthcare AI teams needing fast, compliant guardrails.

GPT-5.5 by OpenAI

469 upvotes | General-Purpose AI | Research & Analysis

OpenAI's most advanced model to date, excelling at coding, research, data analysis, and multi-step autonomous task execution with minimal guidance. For biology researchers: a powerful assistant for literature synthesis, data wrangling, protocol design, and scientific writing — and the general reasoning backbone powering GPT-Rosalind (this week's main story).

DeepSeek-V4

396 upvotes | Open-Source | 1M Context LLM

Open-source MoE model (V4-Pro: 1.6T params, V4-Flash: 284B params) with a 1 million token context window via novel hybrid attention — enabling analysis of entire genomic datasets, full clinical trial corpora, or complete scientific literature in a single pass. Open-source and self-hostable, critical for sensitive clinical and genomic data environments.

⚠️ AI Criticism & Concerns

Critical Perspectives on AI Ethics and Safety

As AI rapidly integrates into clinical practice and drug discovery, critical examination of risks and ethical implications remains essential.

JAMA Study: All 21 Leading LLMs Fail 80%+ of Early Differential Diagnoses

A Mass General Brigham study published in JAMA Network Open found that every major LLM tested — including frontier models from OpenAI, Anthropic, Google, xAI, and DeepSeek — failed more than 80% of early differential diagnoses when patient data was incomplete. While the same models achieved 91% final diagnosis accuracy with complete records, the critical early triage stage showed consistent, severe failure across all 21 models. "Marketing LLMs as diagnostic agents risks fostering false confidence precisely where they are least reliable," the researchers concluded.

Read More — The Register

Americans' Trust in AI Healthcare Drops to 42% — Down from 52% in 2024

A 2026 U.S. News survey found only 42% of Americans are open to AI being used as part of their healthcare, down from 52% just two years ago. Fewer also believe AI can make healthcare more efficient (55%, down from 64%). The erosion is driven by growing coverage of AI diagnostic failures and hallucinations in clinical settings. As AI healthcare adoption accelerates, this trust gap risks creating serious patient compliance and adoption barriers — particularly in communities already distrustful of the medical system.

Read More — U.S. News

AI Recommends Different Treatments Based on Race, Gender, and Income — Not Just Health

A study analysing over 1.7 million AI-generated clinical vignette responses found that race, gender, income, and housing status influenced AI treatment recommendations — even when patients had identical health conditions. Researchers tested nine AI programs with 1,000 ER cases, keeping medical symptoms constant while varying demographics, and found recommendations changed based on personal characteristics rather than clinical need. Deployed at scale, such systems could systematically amplify existing healthcare disparities.

Read More — UCSF CoDEx

EU AI Act High-Risk Healthcare Requirements Take Effect August 2026 — Is the Industry Ready?

Starting August 1, 2026, the EU AI Act's core requirements for high-risk AI systems — including medical devices and clinical decision support — become enforceable. Requirements include conformity assessments, technical documentation, post-market monitoring, human oversight obligations, and transparency mandates. Industry analysts warn that most European healthcare AI deployments are not yet compliant, and penalties of up to €30M or 6% of global turnover could force emergency changes to AI validation across EU health systems.

Read More — JMIR

Closing Note

Two weeks is a long time in AI — and this fortnight proved it. We went from OpenAI naming a life sciences model after a crystallography pioneer to a landmark JAMA study exposing systematic failure at the clinical front line. The acceleration is real, and so are the gaps. As GPT-Rosalind ships to pharma partners, PROTAC therapies reach patients, and the EU AI Act approaches enforcement, the convergence of capability and accountability is becoming unavoidable.

Bigger models are not automatically safer. Faster deployment is not automatically better. The researchers behind SaFE-Scale and the JAMA differential diagnosis study are doing the hard work of establishing that principle empirically. Their work matters as much as the breakthroughs.

Thank you for your continued patience through our technical difficulties, and for reading PythRaSh's AI Newsletter. If you found this week's insights valuable, please share them with colleagues at the intersection of AI and biology.

Have feedback or suggestions? Reply to this email — I read every message!

See you next week!

Md Rasheduzzaman

Share This Newsletter

Found this valuable? Share it with your network!

Share on Twitter Share on LinkedIn Share on Facebook Forward via Email
Unsubscribe | Update Preferences | View in Browser

PythRaSh's AI Newsletter

PythRaSh's AI Newsletter, Fleischerwiese 4, Greifswald-17489, Germany

Visit our website | Connect on LinkedIn | GitHub

Questions? Reply to this email or contact Md Rasheduzzaman
Email: md.rasheduzzaman.ugoe@gmail.com