[GitHub - updating soon] [Demo - updating soon] [Paper - updating soon] [Poster - updating soon]
While deep learning models have achieved expert-level performance in computational pathology, their clinical translation is hindered by a critical lack of generalizability. Models trained on data from one source often fail dramatically when applied to unseen data from a different domain. This project frames the model as a scientific instrument to rigorously investigate and quantify this "domain shift" phenomenon. The core objective is not perfect generalization, but to measure its absence and analyze the reasons for failure, providing evidence-based insights for building robust AI tools for pathology.
The central challenge is domain shift. A model can identify cancer nuclei in one hospital’s dataset but may collapse on images from another hospital or organ type. Subtle variations in tissue preparation, staining, and scanners create this domain gap. Quantifying this gap is the first step toward developing reliable models for clinical use.
The thesis is structured as a two-phase experiment:
The project uses HoVer-Net/U-Net architectures trained on the CoNSeP dataset and evaluated on MoNuSeg for cross-domain performance. Metrics include Panoptic Quality (PQ) for segmentation and macro F1-score for classification.