SMMILe enables accurate spatial quantification in digital pathology using multiple-instance learning


Publication: Nature Cancer

19 November 2025


Zeyu Gao, Anyu Mao, Yuxing Dong, Hannah Clayton, Jialun Wu, Jiashuai Liu, ChunBao Wang, Kai He, Tieliang Gong, Chen Li &  Mireia Crispin-Ortuzar

Abstract:

Spatial quantification is a critical step in most computational pathology tasks, from guiding pathologists to areas of clinical interest to discovering tissue phenotypes behind novel biomarkers. To circumvent the need for manual annotations, modern computational pathology methods have favored multiple-instance learning approaches that can accurately predict whole-slide image labels, albeit at the expense of losing their spatial awareness. Here we prove mathematically that a model using instance-level aggregation could achieve superior spatial quantification without compromising on whole-slide image prediction performance. We then introduce a superpatch-based measurable multiple-instance learning method, SMMILe, and evaluate it across 6 cancer types, 3 highly diverse classification tasks and 8 datasets involving 3,850 whole-slide images. We benchmark SMMILe against nine existing methods using two different encoders—an ImageNet pretrained and a pathology-specific foundation model—and show that in all cases SMMILe matches or exceeds state-of-the-art whole-slide image classification performance while simultaneously achieving outstanding spatial quantification.

View Publication 

© Copyright - NIHR Cambridge Biomedical Research Centre 2025