Tissue stained with H&E.
Developed for automatic segmentation of Nuclei in Brightfield images. Identification and segmentation of individual nuclei is of interest in many applications. Nuclei can be difficult to detect accurately and precisely across different images using traditional image analysis with feature engineering. This APP utilizes artificial intelligence (AI) for automatic nuclei detection in H&E and IHC stained tissue of various types including breast, prostate, liver, stomach and kidney. It consists of a pre-trained deep neural network and the APP is ready for use without additional training.
The APP function as a template where the user can easily configure it to fulfill their needs using the Visiopharm Author™ module. The APP currently has three configuration options which can be switched on and off as needed:
It is also possible for the user to customize the APP by adding other post-processing steps and/or output variables as needed.
Tissue stained with H&E.
Nuclei segmentation in H&E stained tissue. The APP can be set to count the total number of nuclei.
Nuclei segmentation in ER stained tissue. The APP can be set to count the total number of nuclei.
The APP is configured to divide the segmented nuclei into positive and negative based on the intensity of the DAB staining. The APP can be set to output the positive percentage.
The APP is configured to divide the segmented nuclei into negative, 1+, 2+, and 3+ based on the intensity of the DAB staining. The APP can be set to output the H-score.
Quantitative Output variables
The output variables obtained from this protocol can be:
Step 1: Load an image
Step 2: Load and run the APP “10167 – Nuclei Detection, AI (Brightfield)”
The APP was trained using 54,000 annotated nuclei from H&E and IHC stained tissue of various types including breast, prostate, liver, stomach and kidney. The architectural structure of the network is a U-Net which is popular for medical image segmentation. The neural network uses a cascade of layers of nonlinear processing units for feature extraction and transformation, with each successive layer using the output from the previous layers as input. U-Net uses an encoder-decoder structure with a contracting path and an expansive path. For more information of the network architecture, see .
There is no staining protocol available.
To run the APP, a NVIDIA GPU with minimum 4 GB RAM is required.
The APP utilizes the Visiopharm Engine™, Engine™ AI and Viewer software modules, where Engine™ and Engine™ AI offer an execution platform to expand processing capability and speed of image analysis. The Viewer allows a fast review together with several types of image adjustment properties e.g. outlining of regions, annotations and direct measures of distance, curve length, radius, etc.
By adding the Author™ and/or Author™ AI modules the APP can be customized to fit other purposes. These modules offer a comprehensive and dedicated set of tools for creating new fit-for-purpose analysis APPs, and no programming experience is required.
Nuclei Detection, Nuclei Segmentation, IHC, H&E, Brightfield, Image Analysis, Deep Learning, Artificial Intelligence, AI, U-Net
1. Ronneberger, O. et al. U-Net: Convolutional Networks for Biomedical Image Segmentation. International Conference on Medical Image Computing and Computer-Assisted Intervention, 2015, 234-241, DOI.