Skip to contentSkip to frontmatterSkip to Backmatter

An Active Learning Plugin In Napari To Fine-Tune Models For Large-scale Bioimage Analysis

Abstract

The “napari-activelearning” plugin provides a framework to fine-tune deep learning models for large-scale bioimage analysis, such as digital pathology Whole Slide Images. The development of this plugin was motivated by easing the integration of deep learning tools into bioimage analysis workflows. This plugin implements the concept of Active Learning for reducing the time spent on labeling samples when fine-tuning models. Because this plugin is integrated into napari and leverages the use of Next Generation File Formats (e.g. Zarr), it is suitable for fine-tuning deep learning models on large-scale images with little image preparation.

1Introduction

Adoption of deep learning methods for bioimage analysis has grown exponentially in recent years. Part of such success is thanks to transfer learning that enables using models that were trained on large volumes of data from diverse domains, such as the ImageNet Deng et al., 2009 and Segment Anything 1 Billion (SA-1B) Kirillov et al., 2023 datasets, into tasks where annotated data is scarce. An example is cell segmentation in biological microscopy images Greenwald et al., 2021, which requires human annotation of cell structures in images of different modalities and scales. Such images can be up to tens or hundreds of thousands of pixels per side (e.g. WSI used in histopathology), depending on the acquisition magnification and imaging modality. Moreover, the research community has made considerable efforts during the last years to curate databases for training models relevant to the bioimage analysis field. These databases include LIVECell Edlund et al., 2021, TissueNet Greenwald et al., 2021, and CellSeg Lee et al., 2022, among many others. Such databases can be used for de-novo training or be used for fine-tuning models that were trained with image datasets from general domains.

Segmentation of structures in biological image data is a recurrent task in bioimage analysis that serves as an intermediate step for downstream applications. Some of the most common downstream applications for cell segmentation involve cell type classification, counting specific types of cells, and measuring morphological properties of cells in an image. These applications require high quality cell segmentations to obtain accurate information that is used in further bioimage understanding and research. The most relevant deep learning segmentation methods include Cellpose Stringer et al., 2021, Stardist Weigert & Schmidt, 2022, and Micro-SAM Archit et al., 2025. These methods offer pre-trained models for carrying out segmentation of biological structures in multiple imaging modalities, and tools for fine-tuning those same models to new data with user-defined annotations. Training deep learning models for bioimage analysis involves: 1) extracting several rectangular sections of a specified size (i.e. image tiles) from the original image files, 2) annotating each of those tiles according to the tasks being learned, and 3) storing the pairs of tiles and annotations in separate containers or folders as training and testing datasets. However, this approach involves some technical difficulties such as the costs and time associated with transferring training data between researchers for sharing and reproducibility purposes, costs for storing duplicated data from the image tiles already present in the original images, and lack of context and coordinates from where each tile was originally extracted.

The field of Active Learning studies human-in-the-loop strategies in deep learning that can reduce the time and effort required for de-novo training or fine-tuning models. Applications of active learning have also been developed to address computer vision tasks Gal et al., 2017. This is relevant to biological image analysis where data annotation is one of the most time-consuming tasks Budd et al., 2021. The main reason is that due the scale of the images in some modalities (e.g. WSI), the number of samples that can be extracted for labeling can overwhelm the annotator. On the other hand, for relatively smaller images acquired through other modalities, the annotator would require to annotate large amounts of samples to generate sufficient training data. An active learning sampling strategy, based on acquisition functions, can be used to prevent this by presenting a limited number of samples at a time. Moreover, there are acquisition functions such as BALD Houlsby et al., 2011 that focus on obtaining only cases that could improve a model’s performance when correctly labeled.

The bioimage analysis community has been steering towards efficient creation and sharing of repositories containing large volumes of data. NGFFs Moore et al., 2021 are one of the most preferred options for large-scale image storage and management, such as the OME-Zarr data format Moore et al., 2023. The Zarr data format establishes a standard structure for compressed, chunked-based, n-dimensional data storage that can be efficiently stored either in local or cloud repositories @Moore2021-we, Moore et al., 2023. Multiple projects have been created on top of Zarr, including APIs such as the zarr-python library Miles et al., 2020. Some image processing softwares have adopted this data format, like Fiji Schindelin et al., 2012 with the MobIE plugin Pape et al., 2022, and QuPath Bankhead et al., 2017 with its own Zarr data loader. Visualization tools such as napari Sofroniew et al., 2025, neuroglancer Maitin-Shepard et al., 2021, viv Manz et al., 2022, and webKnossos Boergens et al., 2017 have also added support to this data format. Additionally, image data stored as Zarr can be used for training deep learning models without duplicating data. That can be achieved by accessing chunked image data instead of extracting image tiles and storing them separately as is commonly needed in standard deep learning training.

In this work, the “napari-activelearning” plugin for napari is introduced as a tool for easing fine-tuning existing models on large-scale images. This plugin leverages napari’s user-friendly features for annotating new samples following an active learning workflow. Additionally, the “napari-activelearning” plugin uses NGFF to store annotations in a storage-efficient manner, ready for deep learning training.

2Methods

The “napari-activelearning” plugin relies on three components to provide a user-friendly framework to train and fine-tune deep learning models for bioimage analysis: 1) NGFF, for efficient storage of the data generated during the active learning workflow, 2) napari, for visualization and graphical interface with the user, and 3) Active Learning, for assisting users with selection of samples of interest that require annotation for improving the deep learning model performance.

2.1NGFFs

NGFFs Moore et al., 2021, such as OME-Zarr Moore et al., 2023, have been increasingly adopted by the bioimage analysis community thanks to its computation and storage advantages. Zarr format stores large-scale image data as independent n-dimensional tiles, also called chunks, either on local disk or cloud storage. By using chunks as units of storage, the amount of data required to be loaded into memory when accessing specific regions of the image is reduced. Accessing image chunks is parallel-safe, which enables acceleration of image processing through parallel computing. This is useful when applying a model for inference in larger-than-memory image data, where deep learning inference can be applied to regions of the image separately, and applying a subsequent reduction operation on the results. This reduction operation could be an accumulation function for whole image classification, or a stitching algorithm for segmentation tasks.

2.2napari Visualization Tool

napari Sofroniew et al., 2025 is a user-friendly n-dimensional data viewer with extensible capabilities via plugins. This visualizer already offers tools for data annotation, and it is compatible with Zarr, enabling visualization of large-scale image data efficiently. napari has been used to develop deep learning applications for bioimaging analysis such as napari-cellpose Stringer et al., 2021, and Micro-SAM Archit et al., 2025 plugins. These plugins can be used to segment biological structures in data acquired with multiple imaging modalities. However, these plugins are intended to be used with tiles of images that have already been extracted and stored, in contrast to the plugin presented in this work, which can be applied on regions defined by the user directly on original images without storing duplicate data from image tiles.

2.3Active Learning framework

To reduce the amount of data presented to the user for annotation, concepts from the Active Learning framework are implemented in this plugin. This field studies methods for human-in-the-loop learning workflows that avoid overwhelming the annotator with samples to labeling for training. This is achieved through a computation of Acquisition Functions that determine what samples require labeling to subsequently improve the performance of a specific model.

Some of the most used acquisition functions involve Random acquisition, which takes a random unlabeled sample from the dataset with an uniform distribution; Maximize mean standard deviation, that computes the average standard deviation of the probabilities of each class predicted by the model; Maximize variation ratios, which uses the probability of the predicted class of each sample; Maximize the predictive entropy, that is computed from the predicted probabilities for each sample; and Maximize the mutual information between the model parameters and the predicted classes, which is also known as BALD Houlsby et al., 2011. In this plugin, the BALD Houlsby et al., 2011 acquisition function is implemented to score and sort a set of image patches sampled from an image. This acquisition function was selected because it was demonstrated that can be used efficiently with bioimage data for deep learning training when compared with other functions such as Random acquisition Gal et al., 2017. Such efficiency on selecting the most promising samples is highly desirable for increasing the amount of samples presented to the annotator that could improve the model’s performance while reducing redundant samples. In this plugin, the score assigned to a sample extracted at random from the input image is computed following Houlsby et al. (2011):

a(x,M)=I[y,θx,D],a(x, M) = \mathbb{I}[y, \theta| x, D] ,

where a(x,M)a(x, M) is the acquisition function score for sample xx for model MM, the mutual information I\mathbb{I} is computed from the prediction yy made by model MM for sample xx in the dataset DD with the current state of parameters θ\theta. Following Gal et al. (2017), the mutual information can be computed through Monte Carlo integration to compute the BALD function as follows:

I[y,θx,D]=c(1Ttpc^t)log(1Ttpc^t)+1Tc,tpc^tlog pc^t,\mathbb{I}[y, \theta| x, D] = -\sum_{c}\left(\frac{1}{T} \sum_{t} \hat{p_c}^t\right) log \left(\frac{1}{T}\sum_{t}\hat{p_c}^t\right) + \frac{1}{T}\sum_{c,t}\hat{p_c}^t log~\hat{p_c}^t ,

where TT is the total number of steps in the Monte Carlo integration, pc^t\hat{p_c}^t is the component of the prediction made by model MM for sample xx at class cc. To apply the Monte Carlo integration, the prediction made with model MM needs to be converted into a gaussian process. In this plugin, this is done by introducing a Dropout operation after each activation layer found in the deep learning model MM Gal & Ghahramani, 2016. Such introduction of Dropout operations are handled automatically by this plugin as follows: 1) Dropout layers are enabled in evaluation time for models that already have Dropout layers inside their architecture, or 2) Dropout layers are introduced with a default probability of 0.05 after each activation layer found in the model’s module layers.

The active learning process implemented in this plugin is illustrated in Figure 1.

Flowchart of the active learning framework implemented in the “napari-activelearning” plugin.
The process is the following: 1) the user selects a model from the models registry in the plugin,
2) the plugin integrates Dropout layers into a bayesian version of the model (cyto*), 3) prediction on a sample image patch is repeated T times with the bayesian model (cyto3*) in a Monte Carlo simulation process; additionally, the baseline model (cyto3) is used to evaluate the same sample image patch once to obtain the objects labels, 4) the BALD acquisition function is computed from the results of the Monte Carlo simulation process; the resulting BALD acquisition function and objects labels are added to napari’s layers list separately, 5) the annotator selects the next patch with the highest BALD score and corrects its labels using napari’s builtin annotation tools, 6) the user sets up the hyper-parameters of the fine-tuning process corresponding to the baseline model (cyto3) in the plugin, 7) the wrapped fine-tuning process from the selected model is executed using the corrected labels as ground-truth, 8) the sample image patches are predicted again with the fine-tuned model (cyto3**) for comparison.
Note that steps 3 and 4 are repeated for a number of times defined by the annotator to extract sufficient sample patches from the images.
Additionally, the annotator can decide how many samples to correct in step 5, as well as skip labels with low BALD score.

Figure 1:Flowchart of the active learning framework implemented in the “napari-activelearning” plugin. The process is the following: 1) the user selects a model from the models registry in the plugin, 2) the plugin integrates Dropout layers into a bayesian version of the model (cyto*), 3) prediction on a sample image patch is repeated TT times with the bayesian model (cyto3*) in a Monte Carlo simulation process; additionally, the baseline model (cyto3) is used to evaluate the same sample image patch once to obtain the objects labels, 4) the BALD acquisition function is computed from the results of the Monte Carlo simulation process; the resulting BALD acquisition function and objects labels are added to napari’s layers list separately, 5) the annotator selects the next patch with the highest BALD score and corrects its labels using napari’s builtin annotation tools, 6) the user sets up the hyper-parameters of the fine-tuning process corresponding to the baseline model (cyto3) in the plugin, 7) the wrapped fine-tuning process from the selected model is executed using the corrected labels as ground-truth, 8) the sample image patches are predicted again with the fine-tuned model (cyto3**) for comparison. Note that steps 3 and 4 are repeated for a number of times defined by the annotator to extract sufficient sample patches from the images. Additionally, the annotator can decide how many samples to correct in step 5, as well as skip labels with low BALD score.

2.4Implementation details

The “napari-activelearning” plugin is an open-source project implemented in the Python programming language that relies on the napari plugin architecture to offer a user-friendly interface. This plugin consists of three main component widgets that allow the execution of an end-to-end fine-tuning process from within a napari window. Such components are 1) an image groups manager widget, 2) an acquisition function configuration widget, and 3) a labels manager widget.

2.4.1Image Groups Manager

This component is used to gather the metadata of the images used for training and fine-tuning into a single data structure called “image group”. The purpose of this structure, shown in Figure 2, is to define how each layer shown on the napari window will be used in the active learning workflow, such as input data, labels or annotations, and sampling masks. The gathered metadata is used to connect the input data with the selected deep learning model. This is done in this plugin by streaming the pixel data from the input image in the structure and format that the model expects it. The reformating and structuring process commonly involve reordering the spatial axes and color channels of the image data, and even casting the data into a different data type required by the model. The metadata also contains the origin of the image data, which can be a path to an image file on disk or a location in a remote storage. The origin is used by this plugin to access the pixel data directly avoiding duplicating the data when possible. Moreover, sampling masks allow the user to define the regions in the image from where the plugin can sample image patches to be used through the fine-tuning process.

Image groups manager widget used to manage metadata of napari’s layers to be used for inference and fine-tuning along with the Acquisition Function Manager component.
This widget is composed of four main components: a) a set of buttons to create and add image groups, layer groups, and layer channels, b) an editor that allows to define metadata related to each of the layers in the image groups, c) an editor to create sampling masks that allow to focus the active learning process on specified regions of an image defined by the user, and d) a list of the layers added to each image group along with their specific use, such as input image, labels for training, and sampling masks.

Figure 2:Image groups manager widget used to manage metadata of napari’s layers to be used for inference and fine-tuning along with the Acquisition Function Manager component. This widget is composed of four main components: a) a set of buttons to create and add image groups, layer groups, and layer channels, b) an editor that allows to define metadata related to each of the layers in the image groups, c) an editor to create sampling masks that allow to focus the active learning process on specified regions of an image defined by the user, and d) a list of the layers added to each image group along with their specific use, such as input image, labels for training, and sampling masks.

2.4.2Acquisition Function Manager

In this component, the model used for active learning can be selected from a list of registered models, and their hyper-parameters can be configured according to the user’s needs. This component uses NumPy Harris et al., 2020 and PyTorch Paszke et al., 2019 libraries to implement the BALD Houlsby et al., 2011 acquisition function and the Dropout operation insertion Gal & Ghahramani, 2016. The outputs generated in the intermediate steps of the active learning framework are transformed using the Sci-kit Image library Walt et al., 2014 and stored using the OME-Zarr Moore et al., 2023 specification to make them shareable and reproducible under FAIR Wilkinson et al., 2016 guidelines. This component computes the acquisition function score for a set of image patches sampled from the input image defined in the Image Groups Manager component. The number of samples extracted and the total steps for the Monte Carlo integration process can be defined by the user in this widget, as shown in Figure 3. For each sampled patch, the inference and acquisition function are computed and presented to the user in a descending list on the Labels Manager widget. The “napari-activelearning” uses Cellpose Stringer et al., 2021 as default deep learning framework for inference and fine-tuning with its pre-trained models. Moreover, the code that implements this component is intended for its extension to other deep learning models with PyTorch backend Paszke et al., 2019. This can be achieved through the inheritance of a segmentation method class (TunableMethod) that is used to execute pure inference, probability computations, and even fine-tuning of model weights. As an example, the integration of Cellpose models into the napari-activelearning plugin is illustrated in Figure 4. The hyper-parameters that appear at the right in Figure 3 d) had been defined in the fine-tuning function (Figure 4 c) when integrating the Cellpose model into this plugin.

Acquisition function manager widget used mainly to execute model inference and fine-tuning, along with configuration of model parameters and the active learning sampling parameters.
The acquisition function manager widget is composed of six main components that allow a) select the size of the image patches extracted from the input images, b) determine the number of patches that are randomly extracted from the image and the number of repetitions in the Monte Carlo simulation for estimating the acquisition function value, c) configure the inference parameters of the selected deep learning model, d) configure the hyper-parameters for the fine-tuning process, e) execute the inference step on a single image, or all the images added to the image groups manager, as well as execute the fine-tuning process once the labels have been corrected, and f) visualize the progress of the inference step as it is applied to the images added to the image groups manager.

Figure 3:Acquisition function manager widget used mainly to execute model inference and fine-tuning, along with configuration of model parameters and the active learning sampling parameters. The acquisition function manager widget is composed of six main components that allow a) select the size of the image patches extracted from the input images, b) determine the number of patches that are randomly extracted from the image and the number of repetitions in the Monte Carlo simulation for estimating the acquisition function value, c) configure the inference parameters of the selected deep learning model, d) configure the hyper-parameters for the fine-tuning process, e) execute the inference step on a single image, or all the images added to the image groups manager, as well as execute the fine-tuning process once the labels have been corrected, and f) visualize the progress of the inference step as it is applied to the images added to the image groups manager.

Visual representation of the integration of Cellpose models as tunable methods in the napari-activelearning plugin through inheritance of the TunableMethod class.
The derived class (CellposeModel) is required to implement: a) the _run_pred function that predicts the probability of each instance being assigned to the distinct domain classes; in the case of Cellpose that is the probability of each individual pixel being assigned to the foreground or background class, b) the _run_eval function that executes the normal evaluation process of the Cellpose model, and c) the _fine_tune function that wraps the existing fine-tuning process implemented in the Cellpose Python package .

Figure 4:Visual representation of the integration of Cellpose models as tunable methods in the napari-activelearning plugin through inheritance of the TunableMethod class. The derived class (CellposeModel) is required to implement: a) the _run_pred function that predicts the probability of each instance being assigned to the distinct domain classes; in the case of Cellpose that is the probability of each individual pixel being assigned to the foreground or background class, b) the _run_eval function that executes the normal evaluation process of the Cellpose model, and c) the _fine_tune function that wraps the existing fine-tuning process implemented in the Cellpose Python package Stringer et al., 2021.

2.4.3Labels Manager

After the samples are extracted and their acquisition function scores and inferences are computed by the Acquisition Function Manager, these can be reviewed by the user. The labels predicted by the selected model can be corrected using built-in napari annotation tools. These corrected annotations can be used as new targets for fine-tuning the selected model. Additionally, this component provides a simple navigation system to move between inferred labels in space as presented in Figure 5. This system is ideal for reviewing the output generated by the selected model across the image, which can be time consuming in large-scale images, such as WSI or three-dimensional image data.

Label groups manager widget used for handling sampled patches for labels correction following an active learning workflow.
This widget is composed of three main components: a) the list of “label groups” that displays information of the location of the labeled image patches and their corresponding acquisition function value, b) a set of navigation buttons to move between labeled image patches, and c) a set of buttons to edit, commit changes, and remove labels and labels groups.
Note that the currently selected label (at coordinates Z=39, Y=128, and X=0 as displayed in the “Sampling top-left” column of this widget) is in edit mode, which allows to modify the labels of the objects inside the active region without altering the labels of any surrounding image patch.

Figure 5:Label groups manager widget used for handling sampled patches for labels correction following an active learning workflow. This widget is composed of three main components: a) the list of “label groups” that displays information of the location of the labeled image patches and their corresponding acquisition function value, b) a set of navigation buttons to move between labeled image patches, and c) a set of buttons to edit, commit changes, and remove labels and labels groups. Note that the currently selected label (at coordinates Z=39Z=39, Y=128Y=128, and X=0X=0 as displayed in the “Sampling top-left” column of this widget) is in edit mode, which allows to modify the labels of the objects inside the active region without altering the labels of any surrounding image patch.

3Results

The “napari-activelearning” source code is open source and available to the research community at the http://github.com/thejacksonlaboratory/activelearning repository. This package is distributed through PyPI (https://pypi.org/project/napari-activelearning) and can also be installed using the napari plugin manager.

3.1Fine-tuning a Cellpose model

A comparative analysis was carried out to show the benefit of using the active learning framework implemented in this plugin for fine-tuning a deep learning model for bioimage analysis. The experiments consisted of fine-tuning the Cellpose’s Cyto3 model for cell segmentation using training sets of different sizes: 5, 10, 15, 20, 25, and 30 images. The images used in the experiments are part of the Cellpose dataset (https://www.cellpose.org/dataset) Stringer et al., 2021. This dataset was created with images from multiple domains and imaging modalities; however, only a set of 89 images for training and 11 images for testing were used in these experiments. The subset of images selected share the same imaging modality and domain, and it serves as an use case of fine-tuning a model to carry out cell segmentation on a specific imaging modality. The training sets with different sizes were created using the images with the top 5, 10, 15, 20, 25, and 30 BALD score values computed for every image in the pool of 89 training images. The segmentation performance of the different variants of the Cyto3 model were measured in terms of the F1-score computed on the testing set of 11 images. Additionally, a paired Wilcoxon rank test was performed to determine the statistical meaningfulness of the difference between the baseline and fine-tuned models’ segmentation performance. The results of the computational experiments are presented in Table 1.

Table 1:Comparison of the segmentation performance, in terms of F1-score, between the baseline Cyto3 model from Cellpose and fine-tuned version of the same model that were trained with 5, 10, 15, 20, 25, and 30 images taken from a pool of 89 training images. The images used to fine-tune the Cyto3 model were selected based on the highest BALD score computed with this plugin.

ModelVersionTraining sizeTraining F1-scoreTraining Wilcoxon rank test (p-value)Testing F1-scoreTesting Wilcoxon rank test (p-value)
Cyto3baseline-0.8561-0.8447-
Cyto3fine-tuned50.82680.92930.82170.6036
Cyto3fine-tuned100.88380.00320.88000.1620
Cyto3fine-tuned150.89128.18e5e^{-5}0.87850.2057
Cyto3fine-tuned200.89373.06e5e^{-5}0.90130.0654
Cyto3fine-tuned250.89712.17e5e^{-5}0.88540.1185
Cyto3fine-tuned300.89821.07e5e^{-5}0.89080.0741

The comparative analysis results showed the plugin implementation of the active learning framework enables improving the performance of the Cyto3 model through fine-tuning. The results show that the cell segmentation performance of the fine-tuned model can achieve a F1-score of up to 0.9013 on the selected imaging modality. This is significantly greater than the F1-score of 0.8447 obtained by the baseline Cyto3 model (p-value of 0.0654). Moreover, the computational experiments suggest that the active learning plugin can reduce the amount of samples required to be labeled in order to improve the baseline model performance. For this use case, labeling only 10 of the 89 images in the training pool is sufficient to outperform the baseline model performance in the test set (F1-score of 0.8800).

4Conclusion

In this work, the “napari-activelearning” plugin has been introduced for carrying out fine-tuning of deep learning models for large-scale bioimage data. The napari visualization tool was used to host the plugin and serve as a graphical user interface. Additionally, by leveraging NGFF, the annotated data is stored efficiently by creating only image chunks that contain any labels information and avoiding writing empty chunks. Management of the data created through the active learning workflow is annotated following the OME-Zarr v2 specification making it ready for sharing and reproduction. Whereas this plugin was developed to facilitate adoption of deep learning models in bioimage analysis, it is not restricted to these imaging modalities and can be extended to be used with any image stored in the Zarr format. Moreover, a comparative analysis showed that this plugin can be used to improve the performance of a baseline cell segmentation model, reaching a F1-score of 0.9013 while requiring to annotate only 20 images from a pool of 89 images to achive this performance. Finally, the plugin can be used for transfer learning or simply as an interface for deep learning methods that lack graphical user interfaces for inference. A tutorial on how to use this plugin for fine-tuning a Cellpose model to carry out nuclei segmentation on the Cells 3D+2 Channels image from Sci-kit Image (included as a sample image in napari) can be found in the supporting documents section of this paper and on the plugin’s documentation website at https://thejacksonlaboratory.github.io/activelearning/tutorials.html.

License

Copyright © 2025 Cervantes-Sanchez. This is an open-access article distributed under the terms of the Creative Commons Attribution 4.0 International license, which enables reusers to distribute, remix, adapt, and build upon the material in any medium or format, so long as attribution is given to the creator.

Abbreviations
API
Application Programming Interfaces
BALD
Bayesian Active Learning by Disagreement
FAIR
Findable, Accessible, Interoperable, and Reusable
NGFF
Next Generation File Formats
OME
Open Microscopy Environment
WSI
Whole Slide Image
References
  1. Deng, J., Dong, W., Socher, R., Li, L.-J., Li, K., & Fei-Fei, L. (2009). ImageNet: A large-scale hierarchical image database. 2009 IEEE Conference on Computer Vision and Pattern Recognition, 248–255. https://doi.org/10.1109/CVPR.2009.5206848
  2. Kirillov, A., Mintun, E., Ravi, N., Mao, H., Rolland, C., Gustafson, L., Xiao, T., Whitehead, S., Berg, A. C., Lo, W.-Y., Dollár, P., & Girshick, R. (2023). Segment Anything. 2023 IEEE/CVF International Conference on Computer Vision (ICCV), 3992–4003. https://doi.org/10.1109/ICCV51070.2023.00371
  3. Greenwald, N. F., Miller, G., Moen, E., Kong, A., Kagel, A., Dougherty, T., Fullaway, C. C., McIntosh, B. J., Leow, K. X., Schwartz, M. S., Pavelchek, C., Cui, S., Camplisson, I., Bar-Tal, O., Singh, J., Fong, M., Chaudhry, G., Abraham, Z., Moseley, J., … Van Valen, D. (2021). Whole-cell segmentation of tissue images with human-level performance using large-scale data annotation and deep learning. Nat Biotechnol, 40(4), 555–565. https://doi.org/10.1038/s41587-021-01094-0
  4. Edlund, C., Jackson, T. R., Khalid, N., Bevan, N., Dale, T., Dengel, A., Ahmed, S., Trygg, J., & Sjögren, R. (2021). LIVECell—A large-scale dataset for label-free live cell segmentation. Nature Methods, 18(9), 1038–1045. https://doi.org/10.1038/s41592-021-01249-6
  5. Lee, M. Y., Bedia, J. S., Bhate, S. S., Barlow, G. L., Phillips, D., Fantl, W. J., Nolan, G. P., & Schürch, C. M. (2022). CellSeg: a robust, pre-trained nucleus segmentation and pixel quantification software for highly multiplexed fluorescence images. BMC Bioinformatics, 23(1), 46. https://doi.org/10.1186/s12859-022-04570-9
  6. Stringer, C., Wang, T., Michaelos, M., & Pachitariu, M. (2021). Cellpose: a generalist algorithm for cellular segmentation. Nature Methods, 18(1), 100–106. https://doi.org/10.1038/s41592-020-01018-x
  7. Weigert, M., & Schmidt, U. (2022). Nuclei Instance Segmentation and Classification in Histopathology Images with Stardist. The IEEE International Symposium on Biomedical Imaging Challenges (ISBIC). https://doi.org/10.1109/ISBIC56247.2022.9854534
  8. Archit, A., Freckmann, L., Nair, S., Khalid, N., Hilt, P., Rajashekar, V., Freitag, M., Teuber, C., Buckley, G., von Haaren, S., Gupta, S., Dengel, A., Ahmed, S., & Pape, C. (2025). Segment Anything for Microscopy. Nature Methods, 22(3), 579–591. https://doi.org/10.1038/s41592-024-02580-4
  9. Gal, Y., Islam, R., & Ghahramani, Z. (2017). Deep Bayesian Active Learning with Image Data. https://doi.org/10.48550/arXiv.1703.02910
  10. Budd, S., Robinson, E. C., & Kainz, B. (2021). A survey on active learning and human-in-the-loop deep learning for medical image analysis. Medical Image Analysis, 71, 102062. https://doi.org/10.1016/j.media.2021.102062
  11. Houlsby, N., Huszár, F., Ghahramani, Z., & Lengyel, M. (2011). Bayesian Active Learning for Classification and Preference Learning. ArXiv, abs/1112.5745. https://doi.org/10.48550/arXiv.1112.5745
  12. Moore, J., Allan, C., Besson, S., Burel, J.-M., Diel, E., Gault, D., Kozlowski, K., Lindner, D., Linkert, M., Manz, T., Moore, W., Pape, C., Tischer, C., & Swedlow, J. R. (2021). OME-NGFF: a next-generation file format for expanding bioimaging data-access strategies. Nature Methods, 18(12), 1496–1498. https://doi.org/10.1038/s41592-021-01326-w
  13. Moore, J., Basurto-Lozada, D., Besson, S., Bogovic, J., Bragantini, J., Brown, E. M., Burel, J.-M., Casas Moreno, X., de Medeiros, G., Diel, E. E., Gault, D., Ghosh, S. S., Gold, I., Halchenko, Y. O., Hartley, M., Horsfall, D., Keller, M. S., Kittisopikul, M., Kovacs, G., … Swedlow, J. R. (2023). OME-Zarr: a cloud-optimized bioimaging file format with international community support. Histochemistry and Cell Biology, 160(3), 223–251. https://doi.org/10.1007/s00418-023-02209-1
  14. Miles, A., Kirkham, J., Durant, M., Bourbeau, J., Onalan, T., Hamman, J., Patel, Z., shikharsg, Rocklin, M., raphael dussin, Schut, V., de Andrade, E. S., Abernathey, R., Noyes, C., sbalmer, pyup.io bot, Tran, T., Saalfeld, S., Swaney, J., … Banihirwe, A. (2020). zarr-developers/zarr-python: v2.4.0 (v2.4.0) [Computer software]. Zenodo. https://doi.org/10.5281/zenodo.3773450
  15. Schindelin, J., Arganda-Carreras, I., Frise, E., Kaynig, V., Longair, M., Pietzsch, T., Preibisch, S., Rueden, C., Saalfeld, S., Schmid, B., Tinevez, J.-Y., White, D. J., Hartenstein, V., Eliceiri, K., Tomancak, P., & Cardona, A. (2012). Fiji: an open-source platform for biological-image analysis. Nature Methods, 9(7), 676–682. https://doi.org/10.1038/nmeth.2019