Skip to main content

AI Segmentation User Guide

warning

Disclaimer: The AI Segmentation functionality is intended for research purposes only and is not approved for clinical or diagnostic use by any regulatory body (e.g., FDA, CE). Results must always be reviewed and validated by qualified professionals.

Overview

info

Volvicon integrates four AI segmentation frameworks for automated image analysis:

AI Segmentation - nnU-Net

  • TotalSegmentator: Whole-body CT segmentation covering 117 anatomical structures.
  • nnU-Net: Self-configuring deep learning framework for medical image segmentation.
  • MONAI: Medical Open Network for AI, providing access to pre-trained models.
  • Cellpose-SAM: General-purpose instance segmentation for cells, nuclei, and densely packed objects.

This guide details the installation, configuration, and execution of these frameworks within Volvicon.


Installation

Prerequisites

tip

Before you begin, ensure you have:

  • Volvicon application installed
  • Administrator privileges for initial setup
  • Stable internet connection for framework downloads

Installation Methods

tip

This is the easiest and most reliable way to install AI frameworks in Volvicon.

  1. Run Volvicon as Administrator

    • Right-click on Volvicon executable
    • Select "Run as administrator"
  2. Access AI Segmentation Settings

    • Navigate to the AI segmentation module
    • Select the framework you want to install
    • Click the "Install" button
  3. Follow Installation Prompts

    • The application will automatically download and configure the selected framework
    • Installation progress will be displayed
    • Restart Volvicon after installation completes

Method 2: Manual Batch File Installation

note

If application-based installation fails or for advanced users, use the manual method below:

  1. Navigate to Scripts Directory

    <Volvicon Installation Directory>\scripts\Python\
  2. Run Installation Batch Files as Administrator

    • For TotalSegmentator: totalsegmentator_venv_windows_setup.bat
    • For nnU-Net: nnunet_venv_windows_setup.bat
    • For MONAI: monai_venv_windows_setup.bat
    • For Cellpose-SAM: cellpose_venv_windows_setup.bat
  3. Wait for Completion

    • Installation may take 10-30 minutes depending on internet speed
    • Do not close the command window during installation

Troubleshooting Installation

Common Issues

warning

Installation Fails or Corrupted Environment

  • Solution: Delete the virtual environment and reinstall
  • Location: <Volvicon Installation Directory>\ai\frameworks\venvs\
  • Delete the specific framework folder (e.g., totalsegmentator-env, nnunet-env, monai-env, cellpose-env)
  • Run installation again as administrator
warning

Permission Errors

  • Ensure Volvicon is running as administrator
  • Check that Windows User Account Control (UAC) allows the installation
warning

Network/Download Errors

  • Verify internet connection
  • Check firewall settings allow Python package downloads
  • Retry installation after resolving network issues

TotalSegmentator

Description

TotalSegmentator provides automatic segmentation of 117 anatomical structures in whole-body CT scans. It's particularly effective for comprehensive anatomical analysis.

AI Segmentation - TotalSegmentator

Supported Input

  • Image Format: NIfTI (.nii, .nii.gz)
  • Modality: CT scans
  • Anatomy: Whole-body, thorax, or specific anatomical regions

Usage

  1. Load CT Image

    • Open your CT scan in Volvicon
    • Ensure image is in NIfTI format
  2. Select TotalSegmentator

    • Navigate to AI Segmentation module
    • Choose "TotalSegmentator" from framework dropdown
  3. Configure Settings

    • Task: Select appropriate segmentation task (e.g., "total", "lung_vessels", "body")
    • Device: Choose CPU or GPU (if CUDA available)
    • Fast Mode: Enable for quicker processing with slightly reduced accuracy
  4. Run Segmentation

    • Click "Run Segmentation"
    • Monitor progress in the output log
    • Results will automatically load as segmentation masks

Example Tasks

  • total - Complete 117-structure segmentation
  • lung_vessels - Detailed lung vessel segmentation
  • body - Body outline segmentation
  • cerebral_bleed - Cerebral hemorrhage detection
  • hip_implant - Hip implant segmentation
  • coronary_arteries - Coronary artery segmentation

Performance Notes

tip
  • CPU processing: 5-15 minutes depending on image size
  • GPU processing: 2-5 minutes (requires CUDA-compatible GPU)
  • Memory requirement: 8-16 GB RAM recommended

nnU-Net

Description

nnU-Net is a self-configuring deep learning framework that automatically adapts to different medical segmentation tasks. It requires pre-trained models specific to your segmentation task.

Model Requirements

Model Directory Structure

nnU-Net models must be organized in the following structure:

<model_directory>/
├── nnUNetTrainer__nnUNetPlans__3d_fullres/
│ ├── fold_0/
│ │ ├── checkpoint_best.pth
│ │ └── checkpoint_final.pth
│ ├── fold_1/
│ ├── fold_2/
│ ├── fold_3/
│ ├── fold_4/
│ ├── dataset.json
│ ├── plans.json
│ └── predict_from_raw_data_args.json

Obtaining Models

  • Download pre-trained models from nnU-Net model repository
  • Train custom models using nnU-Net framework
  • Ensure model compatibility with your data type

Usage

  1. Prepare Model Directory

    • Place nnU-Net model in accessible directory
    • Verify directory structure matches requirements
  2. Load Image

    • Open medical image in Volvicon
    • Supported formats: NIfTI (.nii, .nii.gz)
  3. Configure nnU-Net Settings

    • Model Directory: Browse to your nnU-Net model folder
    • Configuration: Usually "3d_fullres" for 3D images
    • Device: Select CPU or GPU
    • Folds: Choose specific fold or "all" for ensemble
    • Preprocessing/Segmentation processes (npp/nps): Defaults to 0/0 (sequential mode). Increasing these values enables parallel workers but can cause multiprocessing IPC failures on Windows — keep at 0/0 unless you have confirmed stability with the specific model.
  4. Run Segmentation

    • Click "Run Segmentation"
    • Processing time varies by model complexity
    • Results load automatically as segmentation overlay

Model Configuration Types

  • 2d - For 2D slice-based processing
  • 3d_fullres - Full resolution 3D processing
  • 3d_lowres - Lower resolution 3D for faster processing
  • 3d_lowres_high - A model-defined custom configuration used by certain pre-trained models (for example, ToothFairy dental segmentation). Not a standard nnU-Net configuration; availability depends on the installed model.
  • 3d_cascade_fullres - Two-stage processing for improved accuracy

nnU-Net Best Practices

tip

npp=0 and nps=0 (sequential mode) are the default in Volvicon and the recommended setting on Windows. Higher values enable parallel preprocessing and export workers, but can produce intermittent multiprocessing IPC failures — particularly with models such as ToothFairy. Increase only when you have confirmed stability on the target system.

  • Validate that dataset, trainer, plans, and configuration belong to the same model folder.
  • Keep fold selection simple (fold 0) unless you specifically need multi-fold ensembling.
  • Prefer GPU for inference speed, but expect fallback to CPU arrays when GPU memory is insufficient.

Troubleshooting nnU-Net

warning

Model Loading Errors

  • Verify model directory structure
  • Check that plans.json and dataset.json exist
  • Ensure model was trained with compatible nnU-Net version
warning

Memory Issues

  • Use 3d_lowres configuration for large images
  • Reduce batch size in model configuration
  • Consider using CPU if GPU memory insufficient
warning

Windows Export Worker Errors (WinError 87, SpawnPoolWorker)

  • Symptom: Inference computation finishes, but export fails in multiprocessing workers.
  • Mitigation: Run with npp=0 and nps=0 (sequential mode).
  • In Volvicon’s nnU-Net runner, this fallback is applied automatically when the known error signature is detected.

See also AI-Powered Segmentation Tutorial and Batch Processing and Automation for scripted and multi-dataset workflows.


MONAI

Description

MONAI (Medical Open Network for AI) provides access to a comprehensive collection of pre-trained medical AI models. It supports various segmentation tasks across different anatomical regions and imaging modalities.

AI Segmentation - MONAI

Supported Models

MONAI bundles are pre-packaged models that include:

  • Pre-trained neural networks
  • Preprocessing pipelines
  • Model metadata and configuration

✅ Example Supported (Single-Channel) Bundles:

  • wholeBody_ct_segmentation - Multi-organ CT segmentation
  • wholeBrainSeg_Large_UNEST_segmentation - Brain tissue segmentation
  • prostate_mri_anatomy - Prostate MRI segmentation

❌ Example Unsupported Bundle Types:

  • lung_nodule_ct_detection - Detection models not supported
  • ventricular_short_axis_3label - 2D models not supported
  • BraTS bundles - Multi-channel models requiring multiple MRI sequences
  • Any bundles requiring >1 input channel

Bundle Requirements

Bundle Directory Structure

<bundle_directory>/
├── configs/
│ ├── inference.json
│ ├── metadata.json
│ └── train.json
├── models/
│ └── model.pt
├── scripts/
│ └── (custom Python modules if any)
└── docs/
└── README.md

Obtaining Bundles

  • Download from MONAI Model Zoo
  • Use MONAI Bundle CLI tools
  • Custom bundles following MONAI specifications

Usage

  1. Prepare MONAI Bundle

    • Download or prepare MONAI bundle
    • Verify bundle structure and completeness
  2. Load Medical Image

    • Open image in Volvicon
    • Supported formats: NIfTI (.nii, .nii.gz)
  3. Configure MONAI Settings

    • Bundle Directory: Select path to MONAI bundle
    • Config File: Usually "configs/inference.json"
    • Model File: Usually "models/model.pt"
    • Device: Choose CPU or GPU
    • ROI Size: Inference patch size (auto-detected from bundle)
  4. Advanced Options

    • Sliding Window Batch Size: Adjust for memory constraints
    • Overlap: Control inference overlap (0.0-1.0)
    • Mode: Inference blending mode ("gaussian" or "constant")
  5. Run Inference

    • Click "Run Segmentation"
    • Monitor progress and debug output
    • Results appear as segmentation overlay

Model Compatibility

Supported Model Types

  • 3D Segmentation Models: ✅ Fully supported

    • UNet, SegResNet, SwinUNETR, etc. with 3D spatial dimensions
    • Multi-class and binary segmentation for volumetric data
    • CT and MRI volume models
  • 2D Models: ❌ Not supported

    • Models designed for slice-by-slice processing
    • 2D UNet, 2D SegResNet variants
    • These models expect 2D input but Volvicon processes 3D volumes
  • Multi-Channel Models: ❌ Not supported

    • BraTS bundles requiring multiple MRI sequences
    • Multi-modal imaging bundles
    • Bundles requiring >1 input channel
  • Detection Models: ❌ Not supported

    • RetinaNet, YOLO, R-CNN variants
    • These models output bounding boxes, not segmentation masks

Custom Networks

MONAI runner supports bundles with custom network architectures:

  • Custom Python modules in bundle's scripts/ directory
  • Automatic path resolution for custom imports
  • SwinUNETR and other advanced architectures

Troubleshooting MONAI

warning

Bundle Loading Errors

  • Verify bundle directory structure
  • Check inference.json and metadata.json exist
  • Ensure model file model.pt is present
warning

Network Instantiation Failures

  • Check bundle's Python dependencies
  • Verify custom network modules are properly structured
  • Review error logs for missing parameters
warning

Memory Issues

  • Reduce ROI size for large images
  • Lower sliding window batch size
  • Use CPU if GPU memory insufficient
note

Detection Model Errors

  • MONAI runner only supports segmentation models
  • Detection models (bounding box outputs) are rejected
  • Use segmentation-specific bundles instead
warning

2D Model Errors

  • Error: "2D segmentation model detected"
  • Cause: Bundle contains a 2D model designed for slice-by-slice processing, but Volvicon requires 3D volumetric models
  • Solutions:
    • Use 3D volumetric segmentation bundles instead
    • Look for bundles with spatial_dims: 3 in their configuration
    • For cardiac segmentation: Use 3D cardiac bundles instead of 2D short-axis models
    • Check bundle metadata for spatial dimensions before selection
warning

Channel Compatibility Errors

  • Error: "Multi-channel bundle detected" or "requires X-channel input"
  • Cause: Bundle requires multiple input channels, but Volvicon only supports single-channel images
  • Solutions:
    • Use single-channel compatible bundles instead
    • For brain segmentation: Use wholeBrainSeg_Large_UNEST_segmentation instead of BraTS bundles
    • For multi-organ: Use wholeBody_ct_segmentation for CT scans
    • Check bundle compatibility before selection

Cellpose-SAM

Description

Cellpose-SAM is a general-purpose instance segmentation model combining the Cellpose architecture with the Segment Anything Model (SAM) backbone. Unlike semantic segmentation frameworks that assign class labels to each voxel, Cellpose-SAM performs instance segmentation—identifying and separating individual objects by assigning a unique integer label to each detected instance.

Cellpose-SAM AI Segmentation

This approach is highly effective for:

  • Cell and nuclei segmentation in microscopy images (fluorescence, brightfield, phase contrast).
  • Densely packed objects where adjacent instances must be individually distinguished.
  • Organoid and particle analysis requiring object counting and morphometric measurement.
  • General-purpose object detection in biomedical and industrial imaging contexts.

Cellpose-SAM operates in two modes: 2D mode for single-slice segmentation and 3D mode for volumetric instance segmentation, featuring optional anisotropy correction and slice stitching.

tip

Cellpose-SAM automatically downloads the required model weights (~1.14 GB) upon first use. The model is cached in ~/.cellpose/models/ and requires no manual preparation.

If needed, the model weights can be manually downloaded from cpsam file on Hugging Face. Place the downloaded file in your home directory under C:\Users\UserName\.cellpose\models to enable local use without automatic download.

Supported Input

  • Image Format: NIfTI (.nii, .nii.gz)
  • Modality: Modality-agnostic — works with CT, MRI, microscopy, and other imaging types
  • Dimensionality: 2D slices or 3D volumes
  • Channels: Single-channel (grayscale) images. Multichannel (RGB) images are auto-detected and treated as 2D

Usage

2D Mode

  1. Load Image

    • Open the image in Volvicon.
    • Navigate to the target slice.
  2. Select Cellpose-SAM

    • Navigate to the AI Segmentation module.
    • Select "Cellpose-SAM" from the framework dropdown.
  3. Configure 2D Settings

    • Mode: Select "2D".
    • Slice Index: Set to -1 to process the currently active slice, or specify an exact slice number.
    • Device: Choose CPU or GPU (CUDA).
    • Adjust detection thresholds as necessary (refer to Parameters below).
  4. Run Segmentation

    • Click "Run Segmentation".
    • The output mask will contain labeled instances exclusively on the selected slice. All other slices remain empty (value 0).

3D Mode

  1. Load Volumetric Image

    • Open the 3D volume in Volvicon.
  2. Select Cellpose-SAM

    • Navigate to the AI Segmentation module.
    • Select "Cellpose-SAM" from the framework dropdown.
  3. Configure 3D Settings

    • Mode: Select "3D".
    • Device: Choose CPU or GPU (CUDA).
    • Anisotropy: Define the Z-to-XY spacing ratio for non-isotropic volumes (e.g., 2.5 if Z-spacing is 2.5× the XY-spacing).
    • Stitch Threshold: Set to 0.0 for native 3D segmentation, or approximately 0.9 for 2D-to-3D slice stitching.
    • Adjust detection thresholds as necessary (refer to Parameters below).
  4. Run Segmentation

    • Click "Run Segmentation".
    • The result is a volumetric instance mask where each detected object is assigned a unique label.
note

Native 3D vs. Stitch Mode: When Stitch Threshold is 0.0, Cellpose utilizes native 3D convolutions for volumetric segmentation. When set to a nonzero value (e.g., 0.9), Cellpose segments each Z-slice independently in 2D and subsequently stitches masks across slices using Intersection over Union (IoU) matching. Stitch mode is computationally faster and requires less memory, though it may yield less spatially coherent results.

Parameters

ParameterDefaultRangeDescription
Mode3D2D, 3DSegmentation dimensionality. Use 2D for single-slice segmentation or 3D for volumetric processing.
Devicecpucpu, gpuProcessing device. GPU execution requires a CUDA-compatible NVIDIA GPU.
Flow Threshold0.40.0–5.0Controls mask quality filtering. Lower values (0.1–0.3) accept only high-confidence masks; higher values (0.5–1.0) include uncertain detections.
Cell Probability Threshold0.0−6.0–6.0Minimum probability for a pixel to be considered part of an object. Negative values detect larger/more objects; positive values detect smaller/fewer objects.
Diameter0 (auto)≥ 0Expected object diameter in pixels. Set to 0 for automatic size estimation. Manual specification improves accuracy when object size is known.
Min Size15≥ −1Minimum object area in pixels. Objects smaller than this threshold are discarded. Set to -1 to disable size filtering.
Exclude on Edgesfalsetrue, falseWhen enabled, discards any mask touching the image boundary. Useful for excluding incomplete objects at the periphery.
Batch Size8≥ 1Number of 256×256 image tiles processed simultaneously on the GPU. Higher values increase throughput but require more VRAM.
Anisotropy1.0> 0Z-to-XY voxel spacing ratio for 3D mode. Set to the ratio of Z-spacing to XY-spacing (e.g., 2.0 if Z is 2× coarser). Applicable only in 3D mode.
Stitch Threshold0.00.0–1.0IoU threshold for stitching 2D masks into 3D objects. 0.0 utilizes native 3D segmentation; values around 0.9 enable slice stitching. Applicable only in 3D mode.
Slice Index-1≥ −1Target slice for 2D mode. -1 processes the currently active slice. Applicable only in 2D mode.

Parameter Tuning Guide

tip

Insufficient object detection?

  • Lower Cell Probability Threshold (e.g., from 0.0 to -2.0).
  • Increase Flow Threshold (e.g., from 0.4 to 0.8).
  • Reduce Min Size to retain smaller objects.

Excessive false positives?

  • Raise Cell Probability Threshold (e.g., from 0.0 to 2.0).
  • Lower Flow Threshold (e.g., from 0.4 to 0.2).
  • Increase Min Size to filter out spurious detections.

Objects merging together?

  • Specify Diameter manually if auto-estimation overestimates size.
  • Lower Flow Threshold to enforce stricter mask boundaries.

Incomplete objects at boundaries?

  • Enable Exclude on Edges to discard truncated objects at the image periphery.

3D segmentation performance or memory issues?

  • Utilize Stitch Mode (Stitch Threshold ≈ 0.9) instead of native 3D processing.
  • Reduce Batch Size to lower GPU memory consumption.
  • Switch to CPU processing if GPU memory is insufficient.

Performance Notes

tip
  • CPU Processing (3D): 5–30 minutes, depending on volume size and object density.
  • GPU Processing (3D): 1–10 minutes with a CUDA-compatible GPU.
  • 2D Processing: Typically completes within seconds.
  • System Memory: 8–16 GB RAM recommended; large 3D volumes may require 32 GB.
  • GPU Memory: 4–8 GB VRAM recommended; reduce Batch Size if out-of-memory errors occur.
  • Initial Execution: The first run downloads the model (~1.14 GB), requiring additional time based on network speed.

Troubleshooting Cellpose-SAM

warning

No Masks Produced

  • Lower Cell Probability Threshold (e.g., −2.0) to increase detection sensitivity.
  • Increase Flow Threshold to accept lower-confidence masks.
  • Verify that the input image contains visible objects with sufficient contrast.
  • Ensure Diameter is appropriate for the target objects (or set to 0 for auto-detection).
warning

All Masks Removed by Edge Exclusion

  • Error: "Segmentation produced masks, but all were removed by the exclude-on-edges filter."
  • Cause: Exclude on Edges is enabled, and every detected object touches the image boundary.
  • Solution: Disable Exclude on Edges, or crop the image to ensure complete objects are within the field of view.
warning

CUDA / GPU Errors

  • Verify the presence of a CUDA-compatible NVIDIA GPU.
  • Update GPU drivers to the latest version.
  • Switch to CPU processing if a compatible GPU is unavailable.
  • Reduce Batch Size if encountering GPU out-of-memory (OOM) errors.
warning

Memory Issues in 3D Mode

  • Enable Stitch Mode (Stitch Threshold ≈ 0.9) to reduce peak memory usage.
  • Lower Batch Size during GPU processing.
  • Process a smaller region of interest (ROI) or utilize 2D mode for individual slices.
  • Switch to CPU processing to leverage system RAM instead of GPU VRAM.
note

Model Download on First Run

  • Cellpose-SAM automatically downloads the cpsam model during its initial execution.
  • The model is cached in ~/.cellpose/models/ and reused for subsequent runs.
  • Ensure stable internet connectivity for the initial download.

General Usage Guidelines

Input Requirements

Image Format

  • Primary: NIfTI format (.nii, .nii.gz)
  • Quality: Medical-grade DICOM converted to NIfTI
  • Orientation: Standard radiological orientation preferred

Image Properties

  • Resolution: Appropriate for chosen model
  • Contrast: Adequate tissue contrast for segmentation
  • Artifacts: Minimal motion or metal artifacts
  • Dimensions: 3D volumetric data required (except Cellpose-SAM, which also supports 2D)
    • Supported: 3D CT volumes, 3D MRI volumes, 2D slices (Cellpose-SAM only)
    • Unsupported: 2D slices for TotalSegmentator, nnU-Net, and MONAI
  • Channels: Single-channel (grayscale) images only
    • Supported: CT scans, individual MRI sequences (T1, T2, FLAIR, etc.)
    • Unsupported: Multi-channel bundles requiring multiple imaging sequences

Performance Optimization

Hardware Recommendations

tip
  • CPU: Multi-core processor (8+ cores recommended)
  • RAM: 16-32 GB for complex segmentations
  • GPU: CUDA-compatible GPU with 8+ GB VRAM (optional but recommended)
  • Storage: SSD recommended for faster I/O

Processing Tips

tip
  • GPU Acceleration: Use GPU when available for significantly faster processing
  • Batch Processing: Process multiple similar images with same settings
  • Memory Management: Close unnecessary applications during processing

Output and Results

Segmentation Output

  • Results appear as overlay on original image
  • Multiple tissue classes shown in different colors
  • Segmentation masks saved in project workspace

Result Validation

note
  • Review segmentation accuracy visually
  • Adjust visualization settings for better inspection
  • Manual correction tools available for refinement

Support and Troubleshooting

Common Issues

Installation Problems

warning
  1. Run as Administrator: Most installation issues resolve with admin privileges
  2. Clean Installation: Delete virtual environment and reinstall
  3. Network Issues: Verify internet connectivity and firewall settings

Runtime Errors

warning
  1. Memory Errors: Reduce processing parameters or use CPU
  2. Model Errors: Verify model/bundle integrity and compatibility
  3. Format Errors: Ensure input images are in NIfTI format
  4. 2D Model Errors: Use 3D volumetric models instead of 2D slice-based models
  5. Detection Model Errors: Use segmentation models instead of detection/classification models
  6. No Masks (Cellpose): Adjust flow threshold and cell probability threshold for sensitivity
  7. Edge Exclusion Removed All Masks (Cellpose): Disable exclude-on-edges or ensure objects are fully within the field of view

Performance Issues

tip
  1. Slow Processing: Enable GPU acceleration if available
  2. High Memory Usage: Reduce batch sizes and ROI dimensions
  3. Long Wait Times: Consider using fast modes where available

Empty or poor segmentation output

warning

Actions:

  • Verify input modality/model compatibility.
  • Confirm model resources are valid (bundle/model folder/task).
  • For Cellpose-SAM, tune cellprob_threshold, flow_threshold, and diameter.

Operational recommendations

tip
  • Keep framework versions and project workflows stable within a release cycle.
  • Validate each new model/task combination on reference cases before batch processing.
  • Save parameter presets for repeatability across teams.

Getting Help

Log Files

  • Check Volvicon application logs for detailed error information
  • AI framework logs provide specific model execution details
  • Save relevant log sections when reporting issues

Documentation Resources

Technical Support

  • Contact Volvicon technical support with specific error messages
  • Include system specifications and framework versions
  • Provide sample data and model information when possible

Version Information

This documentation covers:

  • Volvicon AI Segmentation Module 2026.2.1
  • TotalSegmentator v2.11+
  • nnU-Net v2.6.2+
  • MONAI v1.5+
  • Cellpose-SAM v4.0.8+

For the latest updates and version compatibility, refer to the Volvicon release notes and framework documentation.


References and Citation

TotalSegmentator

Wasserthal, J., Breit, H.-C., Meyer, M. T., Pradella, M., Hinck, D., Sauter, A. W., Heye, T., Boll, D. T., Cyriac, J., Yang, S., Bach, M., & Segeroth, M. (2023). TotalSegmentator: Robust segmentation of 104 anatomic structures in CT images. Radiology: Artificial Intelligence, 5(5). https://doi.org/10.1148/ryai.230024

Authors:

Publication Date: July 5, 2023
Access: Free access — Original Research


nnU-Net

Isensee, F., Jaeger, P. F., Kohl, S. A. A., Petersen, J., & Maier-Hein, K. H. (2021). nnU-Net: A self-configuring method for deep learning-based biomedical image segmentation. Nature Methods, 18(2), 203–211.
https://doi.org/10.1038/s41592-020-01008-z


MONAI

Cardoso, M. J., Li, W., Brown, R., Ma, N., Kerfoot, E., Wang, Y., Murrey, B., Myronenko, A., Zhao, C., Yang, D., Nath, V., He, Y., Xu, Z., Hatamizadeh, A., Zhu, W., Liu, Y., Zheng, M., Tang, Y., Yang, I., ... Zhou, S. K. (2022). MONAI: An open-source framework for deep learning in healthcare. arXiv Preprint.
https://doi.org/10.48550/arXiv.2211.02701

Citation Key: arXiv:2211.02701 [cs.LG]
Submission Date: November 4, 2022
Categories: Machine Learning (cs.LG), Artificial Intelligence (cs.AI), Computer Vision and Pattern Recognition (cs.CV)


Cellpose-SAM

Pachitariu, M., Rariden, M., & Stringer, C. (2025). Cellpose-SAM: Superhuman generalization for cellular segmentation. BioRxiv.
https://doi.org/10.1101/2025.04.28.651001


See Also