Hostname: page-component-77f85d65b8-2tv5m Total loading time: 0 Render date: 2026-03-29T14:02:19.690Z Has data issue: false hasContentIssue false

Deep-learning-based image compression for microscopy images: An empirical study

Published online by Cambridge University Press:  20 December 2024

Yu Zhou
Affiliation:
Department of Biospectroscopy, Leibniz-Institut für Analytische Wissenschaften – ISAS – e.V., Dortmund, Germany Faculty of Computer Science, Ruhr University Bochum, Bochum, Germany
Jan Sollmann
Affiliation:
Department of Biospectroscopy, Leibniz-Institut für Analytische Wissenschaften – ISAS – e.V., Dortmund, Germany Faculty of Computer Science, Ruhr University Bochum, Bochum, Germany
Jianxu Chen*
Affiliation:
Department of Biospectroscopy, Leibniz-Institut für Analytische Wissenschaften – ISAS – e.V., Dortmund, Germany
*
Corresponding author: Jianxu Chen; Email: jianxu.chen@isas.de
Rights & Permissions [Opens in a new window]

Abstract

With the fast development of modern microscopes and bioimaging techniques, an unprecedentedly large amount of imaging data is being generated, stored, analyzed, and shared through networks. The size of the data poses great challenges for current data infrastructure. One common way to reduce the data size is by image compression. This study analyzes multiple classic and deep-learning-based image compression methods, as well as an empirical study on their impact on downstream deep-learning-based image processing models. We used deep-learning-based label-free prediction models (i.e., predicting fluorescent images from bright-field images) as an example downstream task for the comparison and analysis of the impact of image compression. Different compression techniques are compared in compression ratio, image similarity, and, most importantly, the prediction accuracy of label-free models on original and compressed images. We found that artificial intelligence (AI)-based compression techniques largely outperform the classic ones with minimal influence on the downstream 2D label-free tasks. In the end, we hope this study could shed light on the potential of deep-learning-based image compression and raise the awareness of the potential impacts of image compression on downstream deep-learning models for analysis.

Information

Type
Perspective
Creative Commons
Creative Common License - CCCreative Common License - BY
This is an Open Access article, distributed under the terms of the Creative Commons Attribution licence (http://creativecommons.org/licenses/by/4.0), which permits unrestricted re-use, distribution and reproduction, provided the original article is properly cited.
Copyright
© The Author(s), 2024. Published by Cambridge University Press
Figure 0

Figure 1. The workflow of a typical learning-based lossy image compression. The raw image $ \mathbf{x} $ is fed into the encoder $ f $ and obtain the low-dimensional latent representation $ \mathbf{y} $. Then, the lossless entropy coder can further exploit the information redundency: $ \mathbf{y} $ will be firstly quantized to $ \mathbf{z}\in {\mathrm{\mathbb{Z}}}^n $, and then compressed to the bitstream $ \mathbf{b} $ by the entropy encoder $ {f}_e $. This bitstream can be stored for transmission or further decompression. The corresponding entropy decoder $ {g}_e $ is responsible for the decompression and yield the reconstructed latent representation $ \hat{\mathbf{y}} $. Lastly, $ \hat{\mathbf{y}} $ is transmitted to the neural decoder $ g $, yielding the reconstructed image $ \hat{\mathbf{x}} $. The loss function of the system is composed of 2 parts: distortion $ \mathcal{D} $ and rate $ \mathrm{\mathcal{R}} $. Distortion represents the reconstruction quality (e.g., Structural Similarity Index Measure [SSIM] between $ \mathbf{x} $ and $ \hat{\mathbf{x}} $) while rate focuses more on the compression ability. $ \lambda $ acts as the hyper-parameter to balance the rate–distortion trade-off.

Figure 1

Figure 2. Overview of our proposed evaluation pipeline. The objective is to fully estimate the compression performance of different compression algorithms (denoted as $ g\hskip0.3em \circ \hskip0.3em f $) in the bioimage field and investigate their influence to the downstream AI-based bioimage analysis tasks (e.g., label-free task in this study, denoted as $ {f}_l $). The solid line represents data flow while the dash line means evaluation. The bright-field raw image $ x $ will be compressed and decompressed: $ \hat{x}=\left(g\hskip0.3em \circ \hskip0.3em f\right)(x)=g\left(\hskip0.3em f(x)\right) $. Then, we feed the reconstructed $ \hat{x} $ to the label-free model $ {f}_l $ to get the estimated fluorescent image $ \hat{y} $: $ \hat{y}={f}_l\left(\hat{x}\right) $. Meanwhile, normal prediction $ y $ is also made by $ {f}_l $ from the raw image $ x $: $ y={f}_l(x) $. Regarding the evaluation, ①\② exhibits the rate–distortion ability of the compression algorithm, ③\④\⑤ represents their influence to the downstream task $ {f}_l $. Specifically, ① measures the reconstruction ability of the compression method while ② records the bit-rate and can reflect the compression ratio ability. ③ and ④ represents the prediction accuracy of the $ {f}_l $ model using the raw image $ x $ and the reconstructed image $ \hat{x} $ as input, respectively. ⑤ measures the similarity between these two predictions.

Figure 2

Table 1. Evaluation of the average 2D bright-field image quality for the different compression methods compared to the original image, to test the reconstruction ability

Figure 3

Figure 3. Compression ratio versus image reconstruction quality (SSIM) for different compression methods. It is evident that there is a trade-off between the compression ratio and the image reconstruction quality. Note that JPEGXR and JPEG-2000-LOSSY are invisible due to the low quality.

Figure 4

Figure 4. Visualization of 2D bright-field image compression result (first row, model: mbt2018 (mse)) + downstream label-free model prediction (second row). The upper right compression result is visually plausible compared to the input, and the compressed prediction (bottom left) using the label-free model is very close to the original prediction (bottom middle), which suggests the minimal influence of the selected deep-learning-based compression to the downstream task.

Figure 5

Figure 5. Visualization of 3D compression result based on the bmshj2018-factorized model.

Figure 6

Table 2. Evaluation of the average prediction quality for the different compression methods compared to the ground truth, to test the impact of the compression methods to the label-free task

Figure 7

Table 3. Evaluation of the average prediction quality for the different compression methods compared to the original prediction

Figure 8

Table 4. 3D compression results using the bmshj2018-factorized model

Figure 9

Figure 6. The prediction result of the downstream label-free models trained with raw/lossy compressed images, respectively. The input is the lossy compressed bright-field images using mbt2018 (mse) model. (a) Prediction from a label-free model trained with raw uncompressed images,(26) (b) Prediction from a label-free model trained with images compressed with mbt2018 (mse) model, (c) The ground truth. The label-free model trained on uncompressed data fails to produce accurate results when applied to lossy compressed images, as evidenced by the visible artifacts. This highlights the incompatibility between the model trained on original data and the application of lossy compression.

Supplementary material: File

Zhou et al. supplementary material

Zhou et al. supplementary material
Download Zhou et al. supplementary material(File)
File 1.5 MB