Metadata
Title
News & Events
Category
international
UUID
1079831396df49d394f1aa9d6da8b690
Source URL
https://global.sjtu.edu.cn/en/news-events/news/2412
Parent URL
https://global.sjtu.edu.cn/en/news-events/meet-sjtu/research
Crawl Time
2026-03-13T07:08:58+00:00
Rendered Raw Markdown

News & Events

Source: https://global.sjtu.edu.cn/en/news-events/news/2412 Parent: https://global.sjtu.edu.cn/en/news-events/meet-sjtu/research

Back

Study@SJTU

Prospective International Students

Back

Prospective International Students

News Notices Degree Programs Non-Degree Programs Virtual Opportunities Scholarships Application

Current International Students

Back

Current International Students

News Notices Administration Directory University Policies Campus Events Services Job Information Downloads Useful Links FAQ - Research

Back

Research

Overview

Back

Overview

Academic Strengths

Back

Academic Strengths

Research

Back

Research

Facilities & Platforms

Back

Facilities & Platforms

Institutes & Centers Think Tank Institutions Journals

Contact US

Back

Contact US - Work@SJTU

Back

Work@SJTU

Overview

Back

Overview

Faculty & Staff

Back

Faculty & Staff

Gather@SJTU

Back

Gather@SJTU

International Faculty Club Events

Join Us

Back

Join Us

Employment Opportunities Explore Careers Discover SJTU About Shanghai

Services

Back

Services

Visa & Work Permit Overseas Expert Invitation Campus Life Your Child's Education Entry Services Travel and Accommodations - Global SJTU

Back

Global SJTU

Global Network

Back

Global Network

Global Cooperation

Back

Global Cooperation

Seed Funds

Back

Seed Funds

Contact Us

Back

Contact Us - News & Events

Back

News & Events

SJTU in the News

Back

SJTU in the News

Noticeboard

Back

Noticeboard

Upcoming Events

Back

Upcoming Events

Meet SJTU

Back

Meet SJTU

Research Faculty & Staff Students & Campus Alumni Global Cooperation

Newsletters

Back

Newsletters

SJTU in the media

Back

SJTU in the media - About Us

Back

About Us

SJTU in 2030

Back

SJTU in 2030

Donation

Our Story

Back

Our Story

Our History Schools & Colleges Institutes & Centers SJTU Museums Highlights Discover SJTU

Leadership & Governance

Back

Leadership & Governance

University Leadership University Administration

About Shanghai

Back

About Shanghai

Download

Back

Download

SJTU School of Artificial Intelligence Proposes Text-Driven 3D Medical Image Segmentation Foundation Model

November 21, 2025 Page views: 3103

npj Digital Medicine (CAS Q1 Top journal, 5-year Impact Factor ≈16.6) recently published the latest research titled “Large-vocabulary segmentation for medical images with text prompts,” conducted by the Smart Healthcare Team of the SJTU School of Artificial Intelligence.

In this work, the team introduced SAT (Segment Anything in Radiology Scans, driven by Text prompts) — the first 3D medical image segmentation foundation model that incorporates multimodal anatomical knowledge and is driven by text prompts. Its effectiveness and generalization capability were validated across a large-scale benchmark comprising 72 datasets.

This work (i.e., SAT) has been selected as an official baseline for the CVPR 2025 BiomedSegFM Challenge.

Medical image segmentation is vital for diagnosis, surgical planning, and disease monitoring. Traditional approaches rely on “specialized models” trained separately for each task, lacking generalizability across organs and imaging modalities.

Meanwhile, as large language models evolve rapidly in healthcare, there is an urgent need for segmentation tools that integrate language understanding with spatial localization.

To address these challenges, the SJTU team proposed SAT(Segment Anything in radiology scans, driven by Text prompts) — the first 3D segmentation foundation model enhanced by anatomical knowledge and guided by medical text prompts.

Figure 1: Differences in paradigm between SAT and existing specialized models (e.g., nnU-Net) and interactive models (e.g., MedSAM).

In terms of data, the research team first constructed the field’s first multimodal knowledge graph covering more than 6,000 human anatomical concepts, drawing on UMLS and several other sources. The graph includes detailed textual descriptions of each concept, multiple types of semantic relationships among concepts, as well as selected visual examples and segmentation annotations.

The team also integrated 72 public datasets to build SAT-DS, a large-scale 3D medical image segmentation dataset containing 22,186 3D scans and 302,033 segmentation annotations. SAT-DS spans three imaging modalities—CT, MRI, and PET—and covers 497 segmentation targets across eight major anatomical regions.

Figure 2: As one of the largest 3D medical image segmentation datasets in the field, SAT-DS covers 497 segmentation categories across eight major anatomical regions.

Methodologically, the research team proposed a text-prompt-driven segmentation framework that adopts a two-stage process of “knowledge injection+segmentation training.” First, multimodal anatomical knowledge from more than 6,000 concepts is injected into the text encoder through contrastive learning to obtain more accurate and visually aligned concept representations. Then, the encoded text prompts are combined with 3D image features to decode and produce segmentation predictions.

Figure 3: The construction of SAT consists of two stages: knowledge injection and segmentation training.

The research team trained two variants of the model on SAT-DS: SAT-Pro (447M parameters) and SAT-Nano (110M parameters). Across all 497 categories, SAT-Pro achieves performance comparable to 72 individually trained nnU-Net models, while reducing the parameter count by more than 80%, significantly lowering usage and deployment costs.

Compared with the interactive segmentation model MedSAM, SAT-Pro demonstrates substantial advantages in organ segmentation and can achieve fully automatic segmentation through text prompts.

Relative to contemporary work, BiomedParse (first released in May 2024), SAT-Pro delivers notably superior overall performance and supports a much larger vocabulary of segmentation categories.

The team also conducted external validation experiments, confirming SAT’s strong generalization ability across different datasets. More experimental details are provided in the published paper.

Figure 4: In-domain comparison of SAT-Pro and SAT-Nano with specialized models including nnU-Net, U-Mamba, and SwinUNETR.

As the first text-prompt-based 3D medical image segmentation foundation model, SAT offers value in multiple dimensions:

This study was conducted by the Smart Healthcare Team of the SJTU School of Artificial Intelligence. The first author is Ziheng Zhao, PhD student at Shanghai Jiao Tong University. The corresponding authors are Prof. Yanfeng Wang and Assoc. Prof. Weidi Xie of the SJTU School of Artificial Intelligence. The work was supported by the Science and Technology Innovation 2030 “New Generation Artificial Intelligence (2030)” Major Project (2022ZD0160702).

All code, models, and datasets from this study have been open-sourced:

Code: https://github.com/zhaoziheng/SAT/

Paper: https://www.nature.com/articles/s41746-025-01964-w

Dataset: https://github.com/zhaoziheng/SAT-DS/

Author: SJTU School of Artificial Intelligence

Contributing Unit: SJTU School of Artificial Intelligence

Translated by: Denise

Proof read: Rui Su

[March 05, 2026

A gift from SJTU to the world 🥰 DeepRare has arrived!

Learn more](https://global.sjtu.edu.cn/en/news-events/news/2479)

[February 27, 2026

A Strong Start to the Year of the Horse! SJTU Publishes Two Nature Papers on the Third Day of the Chinese New Year!

Learn more](https://global.sjtu.edu.cn/en/news-events/news/2465)

[December 31, 2025

New optical chip can help advance generative AI

Learn more](https://global.sjtu.edu.cn/en/news-events/news/2458)