Your browser doesn't support javascript.
loading
Show: 20 | 50 | 100
Results 1 - 20 de 96
Filter
1.
Entropy (Basel) ; 26(6)2024 May 29.
Article in English | MEDLINE | ID: mdl-38920476

ABSTRACT

Block compressed sensing (BCS) is a promising method for resource-constrained image/video coding applications. However, the quantization of BCS measurements has posed a challenge, leading to significant quantization errors and encoding redundancy. In this paper, we propose a quantization method for BCS measurements using convolutional neural networks (CNN). The quantization process maps measurements to quantized data that follow a uniform distribution based on the measurements' distribution, which aims to maximize the amount of information carried by the quantized data. The dequantization process restores the quantized data to data that conform to the measurements' distribution. The restored data are then modified by the correlation information of the measurements drawn from the quantized data, with the goal of minimizing the quantization errors. The proposed method uses CNNs to construct quantization and dequantization processes, and the networks are trained jointly. The distribution parameters of each block are used as side information, which is quantized with 1 bit by the same method. Extensive experiments on four public datasets showed that, compared with uniform quantization and entropy coding, the proposed method can improve the PSNR by an average of 0.48 dB without using entropy coding when the compression bit rate is 0.1 bpp.

2.
Res Sq ; 2024 Apr 22.
Article in English | MEDLINE | ID: mdl-38746384

ABSTRACT

This paper presents a study on the computational complexity of coding for machines, with a focus on image coding for classification. We first conduct a comprehensive set of experiments to analyze the size of the encoder (which encodes images to bitstreams), the size of the decoder (which decodes bitstreams and predicts class labels), and their impact on the rate-accuracy trade-off in compression for classification. Through empirical investigation, we demonstrate a complementary relationship between the encoder size and the decoder size, i.e., it is better to employ a large encoder with a small decoder and vice versa. Motivated by this relationship, we introduce a feature compression-based method for efficient image compression for classification. By compressing features at various layers of a neural network-based image classification model, our method achieves adjustable rate, accuracy, and encoder (or decoder) size using a single model. Experimental results on ImageNet classification show that our method achieves competitive results with existing methods while being much more flexible. The code will be made publicly available.

3.
J Imaging ; 10(5)2024 May 20.
Article in English | MEDLINE | ID: mdl-38786578

ABSTRACT

Vector quantization (VQ) is a block coding method that is famous for its high compression ratio and simple encoder and decoder implementation. Linde-Buzo-Gray (LBG) is a renowned technique for VQ that uses a clustering-based approach for finding the optimum codebook. Numerous algorithms, such as Particle Swarm Optimization (PSO), the Cuckoo search algorithm (CS), bat algorithm, and firefly algorithm (FA), are used for codebook design. These algorithms are primarily focused on improving the image quality in terms of the PSNR and SSIM but use exhaustive searching to find the optimum codebook, which causes the computational time to be very high. In our study, our algorithm enhances LBG by minimizing the computational complexity by reducing the total number of comparisons among the codebook and training vectors using a match function. The input image is taken as a training vector at the encoder side, which is initialized with the random selection of the vectors from the input image. Rescaling using bilinear interpolation through the nearest neighborhood method is performed to reduce the comparison of the codebook with the training vector. The compressed image is first downsized by the encoder, which is then upscaled at the decoder side during decompression. Based on the results, it is demonstrated that the proposed method reduces the computational complexity by 50.2% compared to LBG and above 97% compared to the other LBG-based algorithms. Moreover, a 20% reduction in the memory size is also obtained, with no significant loss in the image quality compared to the LBG algorithm.

4.
Entropy (Basel) ; 26(5)2024 Apr 24.
Article in English | MEDLINE | ID: mdl-38785606

ABSTRACT

End-to-end learned image compression codecs have notably emerged in recent years. These codecs have demonstrated superiority over conventional methods, showcasing remarkable flexibility and adaptability across diverse data domains while supporting new distortion losses. Despite challenges such as computational complexity, learned image compression methods inherently align with learning-based data processing and analytic pipelines due to their well-suited internal representations. The concept of Video Coding for Machines has garnered significant attention from both academic researchers and industry practitioners. This concept reflects the growing need to integrate data compression with computer vision applications. In light of these developments, we present a comprehensive survey and review of lossy image compression methods. Additionally, we provide a concise overview of two prominent international standards, MPEG Video Coding for Machines and JPEG AI. These standards are designed to bridge the gap between data compression and computer vision, catering to practical industry use cases.

5.
Sensors (Basel) ; 24(8)2024 Apr 19.
Article in English | MEDLINE | ID: mdl-38676223

ABSTRACT

Vector Quantization (VQ) is a technique with a wide range of applications. For example, it can be used for image compression. The codebook design for VQ has great significance in the quality of the quantized signals and can benefit from the use of swarm intelligence. Initialization of the Linde-Buzo-Gray (LBG) algorithm, which is the most popular VQ codebook design algorithm, is a step that directly influences VQ performance, as the convergence speed and codebook quality depend on the initial codebook. A widely used initialization alternative is random initialization, in which the initial set of codevectors is drawn randomly from the training set. Other initialization methods can lead to a better quality of the designed codebooks. The present work evaluates the impacts of initialization strategies on swarm intelligence algorithms for codebook design in terms of the quality of the designed codebooks, assessed by the quality of the reconstructed images, and in terms of the convergence speed, evaluated by the number of iterations. Initialization strategies consist of a combination of codebooks obtained by initialization algorithms from the literature with codebooks composed of vectors randomly selected from the training set. The possibility of combining different initialization techniques provides new perspectives in the search for the quality of the VQ codebooks. Nine initialization strategies are presented, which are compared with random initialization. Initialization strategies are evaluated on the following algorithms for codebook design based on swarm clustering: modified firefly algorithm-Linde-Buzo-Gray (M-FA-LBG), modified particle swarm optimization-Linde-Buzo-Gray (M-PSO-LBG), modified fish school search-Linde-Buzo-Gray (M-FSS-LBG) and their accelerated versions (M-FA-LBGa, M-PSO-LBGa and M-FSS-LBGa) which are obtained by replacing the LBG with the accelerated LBG algorithm. The simulation results point out to the benefits of the proposed initialization strategies. The results show gains up to 4.43 dB in terms of PSNR for image Clock with M-PSO-LBG codebooks of size 512 and codebook design time savings up to 67.05% for image Clock, with M-FF-LBGa codebooks with size N=512, by using initialization strategies in substitution to Random initialization.

6.
Entropy (Basel) ; 26(4)2024 Apr 05.
Article in English | MEDLINE | ID: mdl-38667870

ABSTRACT

Rapid and continuous advancements in remote sensing technology have resulted in finer resolutions and higher acquisition rates of hyperspectral images (HSIs). These developments have triggered a need for new processing techniques brought about by the confined power and constrained hardware resources aboard satellites. This article proposes two novel lossless and near-lossless compression methods, employing our recent seed generation and quadrature-based square rooting algorithms, respectively. The main advantage of the former method lies in its acceptable complexity utilizing simple arithmetic operations, making it suitable for real-time onboard compression. In addition, this near-lossless compressor could be incorporated for hard-to-compress images offering a stabilized reduction at nearly 40% with a maximum relative error of 0.33 and a maximum absolute error of 30. Our results also show that a lossless compression performance, in terms of compression ratio, of up to 2.6 is achieved when testing with hyperspectral images from the Corpus dataset. Further, an improvement in the compression rate over the state-of-the-art k2-raster technique is realized for most of these HSIs by all four variations of our proposed lossless compression method. In particular, a data reduction enhancement of up to 29.89% is realized when comparing their respective geometric mean values.

7.
Sensors (Basel) ; 24(4)2024 Feb 14.
Article in English | MEDLINE | ID: mdl-38400366

ABSTRACT

This paper discusses optimizing desktop image quality and bandwidth consumption in remote IoT GUI desktop scenarios. Remote desktop tools, which are crucial for work efficiency, typically employ image compression techniques to manage bandwidth. Although JPEG is widely used for its efficiency in eliminating redundancy, it can introduce quality loss with increased compression. Recently, deep learning-based compression techniques have emerged, challenging traditional methods like JPEG. This study introduces an optimized RFB (Remote Frame Buffer) protocol based on a convolutional neural network (CNN) image compression algorithm, focusing on human visual perception in desktop image processing. The improved RFB protocol proposed in this paper, compared to the unoptimized RFB protocol, can save 30-80% of bandwidth consumption and enhances remote desktop image quality, as evidenced by improved PSNR and MS-SSIM values between the remote desktop image and the original image, thus providing superior desktop image transmission quality.

8.
Sensors (Basel) ; 24(3)2024 Jan 25.
Article in English | MEDLINE | ID: mdl-38339507

ABSTRACT

Image compression is a vital component for domains in which the computational resources are usually scarce such as automotive or telemedicine fields. Also, when discussing real-time systems, the large amount of data that must flow through the system can represent a bottleneck. Therefore, the storage of images, alongside the compression, transmission, and decompression procedures, becomes vital. In recent years, many compression techniques that only preserve the quality of the region of interest of an image have been developed, the other parts being either discarded or compressed with major quality loss. This paper proposes a study of relevant papers from the last decade which are focused on the selection of a region of interest of an image and on the compression techniques that can be applied to that area. To better highlight the novelty of the hybrid methods, classical state-of-the-art approaches are also analyzed. The current work will provide an overview of classical and hybrid compression methods alongside a categorization based on compression ratio and other quality factors such as mean-square error and peak signal-to-noise ratio, structural similarity index measure, and so on. This overview can help researchers to develop a better idea of what compression algorithms are used in certain domains and to find out if the presented performance parameters are of interest for the intended purpose.

9.
Indian J Nucl Med ; 38(3): 231-238, 2023.
Article in English | MEDLINE | ID: mdl-38046967

ABSTRACT

Aim and Objective: The objective of this study was to optimize the threshold for discrete cosine transform (DCT) coefficients for near-lossless compression of Tc-99 m Dimercaptosuccinic acid (DMSA) scan images using discrete cosine transformation. Materials and Methods: Two nuclear medicine (NM) Physicians after reviewing several Tc-99 m DMSA scan images provided 242 Tc-99 m DMSA scan images that had scar. These Digital imaging and communication in medicine (DICOM) images were converted in the Portable Network Graphics (PNG) format. DCT was applied on these PNG images, which resulted in DCT coefficients corresponding to each pixel of the image. Four different thresholds equal to 5, 10, 15, and 20 were applied and then inverse discrete cosine transformation was applied to get the compressed Tc-99 m DMSA scan images. Compression factor was calculated as the ratio of the number of nonzero elements after thresholding DCT coefficients to the number of nonzero elements before thresholding DCT coefficients. Two NM physicians who had provided the input images visually compared the compressed images with its input image, and categorized the compressed images as either acceptable or unacceptable. The quality of compressed images was also assessed objectively using the following eight image quality metrics: perception-based image quality evaluator, structural similarity index measure (SSIM), multiSSIM, feature similarity indexing method, blur, global contrast factor, contrast per pixel, and brightness. Pairwise Wilcoxon signed-rank sum tests were applied to find the statistically significant difference between the value of image quality metrics of the compressed images obtained at different thresholds and the value of the image quality metrics of its input images at the level of significance = 0.05. Results: At threshold 5, (1) all compressed images (242 out of 242 Tc-99 m DMSA scan images) were acceptable to both the NM Physicians, (2) Compressed image looks identical to its original image and no loss of clinical details was noticed in compressed images, (3) Up to 96.65% compression (average compression: 82.92%) was observed, and (4) Result of objective assessment supported the visual assessment. The quality of compressed images at thresholds 10, 15, and 20 was significantly better than that of input images at P < 0.0001. However, the number of unacceptable compressed images at thresholds 10, 15, and 20 was 6, 38, and 70, respectively. Conclusions: Up to 96.65%, near-losses compression of Tc-99 m DMSA images was found using DCT by thresholding DCT coefficients at a threshold value equal to 5.

10.
Sensors (Basel) ; 23(21)2023 Oct 31.
Article in English | MEDLINE | ID: mdl-37960560

ABSTRACT

JPEG is the international standard for still image encoding and is the most widely used compression algorithm because of its simple encoding process and low computational complexity. Recently, many methods have been developed to improve the quality of JPEG images by using deep learning. However, these methods require the use of high-performance devices since they need to perform neural network computation for decoding images. In this paper, we propose a method to generate high-quality images using deep learning without changing the decoding algorithm. The key idea is to reduce and smooth colors and gradient regions in the original images before JPEG compression. The reduction and smoothing can suppress red block noise and pseudo-contour in the compressed images. Furthermore, high-performance devices are unnecessary for decoding. The proposed method consists of two components: a color transformation network using deep learning and a pseudo-contour suppression model using signal processing. The experimental results showed that the proposed method outperforms standard JPEG in quality measurements correlated with human perception.

11.
Heliyon ; 9(9): e20191, 2023 Sep.
Article in English | MEDLINE | ID: mdl-37809752

ABSTRACT

Fluorescence images enhancement is important as it can provide more information for medical diagnosis. In this work, we design three simple yet useful filters based on the combinations of mathematical functions, which are proved to be effective in strengthening the images acquired from the fluorescence microscope. Using these filters, detailed objects can be found in the dark sections of the fluorescence images. In addition, these filters can be used to enhance the low-light image, which provide satisfactory visual information and marginal profile for the blurred objects in the image. Moreover, these filters have been used to enhance the image with high degradation by the Gaussian noise, where clear edge profile can be extracted. Finally, we have shown that these filters can be utilized for the image compression. Compression ratio can be obtained to be 0.9688. This study shows the making of the filters with dual functions for the image enhancement and the image compression. Our designed filters are showing the potentials in the field of biomedical imaging and pattern identification.

12.
Entropy (Basel) ; 25(10)2023 Sep 25.
Article in English | MEDLINE | ID: mdl-37895504

ABSTRACT

Of late, image compression has become crucial due to the rising need for faster encoding and decoding. To achieve this objective, the present study proposes the use of canonical Huffman coding (CHC) as an entropy coder, which entails a lower decoding time compared to binary Huffman codes. For image compression, discrete wavelet transform (DWT) and CHC with principal component analysis (PCA) were combined. The lossy method was introduced by using PCA, followed by DWT and CHC to enhance compression efficiency. By using DWT and CHC instead of PCA alone, the reconstructed images have a better peak signal-to-noise ratio (PSNR). In this study, we also developed a hybrid compression model combining the advantages of DWT, CHC and PCA. With the increasing use of image data, better image compression techniques are necessary for the efficient use of storage space. The proposed technique achieved up to 60% compression while maintaining high visual quality. This method also outperformed the currently available techniques in terms of both PSNR (in dB) and bit-per-pixel (bpp) scores. This approach was tested on various color images, including Peppers 512 × 512 × 3 and Couple 256 × 256 × 3, showing improvements by 17 dB and 22 dB, respectively, while reducing the bpp by 0.56 and 0.10, respectively. For grayscale images as well, i.e., Lena 512 × 512 and Boat 256 × 256, the proposed method showed improvements by 5 dB and 8 dB, respectively, with a decrease of 0.02 bpp in both cases.

13.
IEEE Trans Circuits Syst Video Technol ; 33(8): 4108-4121, 2023 Aug.
Article in English | MEDLINE | ID: mdl-37547669

ABSTRACT

Advances in both lossy image compression and semantic content understanding have been greatly fueled by deep learning techniques, yet these two tasks have been developed separately for the past decades. In this work, we address the problem of directly executing semantic inference from quantized latent features in the deep compressed domain without pixel reconstruction. Although different methods have been proposed for this problem setting, they either are restrictive to a specific architecture, or are sub-optimal in terms of compressed domain task accuracy. In contrast, we propose a lightweight, plug-and-play solution which is generally compliant with popular learned image coders and deep vision models, making it attractive to vast applications. Our method adapts prevalent pixel domain neural models that are deployed for various vision tasks to directly accept quantized latent features (other than pixels). We further suggest training the compressed domain model by transferring knowledge from its corresponding pixel domain counterpart. Experiments show that our method is compliant with popular learned image coders and vision task models. Under fair comparison, our approach outperforms a baseline method by a) more than 3% top-1 accuracy for compressed domain classification, and b) more than 7% mIoU for compressed domain semantic segmentation, at various data rates.

14.
Indian J Nucl Med ; 38(2): 103-109, 2023.
Article in English | MEDLINE | ID: mdl-37456182

ABSTRACT

Introduction: The objective of the study was to compress 99m-Tc TRODAT single-photon emission computerized tomography (SPECT) scan image using Singular Value Decomposition (SVD) into an acceptable compressed image and then calculate the compression factor. Materials and Methods: The SVD of every image from the image dataset of 2256 images (of forty-eight 99m-Tc TRODAT SPECT studies [48 studies X 47 trans-axial images = 2256 trans-axial images]) was computed and after truncating singular values smaller than a threshold, the compressed image was reconstructed. The SVD computation time and percentage compression achieved were calculated for each image. Two nuclear medicine physicians visually compared compressed image with its original image, and labeled it as either acceptable or unacceptable. Compressed image having loss of clinical details or presence of compression artifact was labeled unacceptable. The quality of compressed image was also assessed objectively using the following image quality metrics: Error, structural similarity (SSIM), brightness, global contrast factor (GCF), contrast per pixel (CPP), and blur. We also compared the TRODAT uptake in basal ganglia estimated from the compressed image and original image. Results: Nuclear Medicine Physician labeled each image acceptable, as they found compressed image identical to its original image. The values of brightness, GCF, CPP, and blur metrics show that compressed images are less noisy, brighter, and sharper than its original image. The median values of error (0.0006) and SSIM (0.93) indicate that the compressed images were approximately identical to its original image. In 39 out of 48 studies, the percentage difference in TRODAT uptake (in basal ganglia from compressed and original image) was negligible (approximately equal to zero). In remaining 9 studies, the maximum percentage difference was 13%. The SVD computation time and percentage compression achieved for a TRODAT study were 0.17398 s and up to 54.61%, respectively. Conclusions: The compression factor up to 54.61% was achieved during 99m-Tc TRODAT SPECT scan image compression using SVD, for an acceptable compressed image.

15.
Diagn Interv Imaging ; 2023 Jun 14.
Article in English | MEDLINE | ID: mdl-37328394

ABSTRACT

PURPOSE: The purpose of this study was to compare a new free-breathing compressed sensing cine (FB-CS) cardiac magnetic resonance imaging (CMR) to the standard reference multi-breath-hold segmented cine (BH-SEG) CMR in an unselected population. MATERIALS AND METHODS: From January to April 2021, 52 consecutive adult patients who underwent both conventional BH-SEG CMR and new FB-CS CMR with fully automated respiratory motion correction were retrospectively enrolled. There were 29 men and 23 women with a mean age of 57.7 ± 18.9 (standard deviation [SD]) years (age range: 19.0-90.0 years) and a mean cardiac rate of 74.6 ± 17.9 (SD) bpm. For each patient, short-axis stacks were acquired with similar parameters providing a spatial resolution of 1.8 × 1.8 × 8.0 mm3 and 25 cardiac frames. Acquisition and reconstruction times, image quality (Likert scale from 1 to 4), left and right ventricular volumes and ejection fractions, left ventricular mass, and global circumferential strain were assessed for each sequence. RESULTS: FB-CS CMR acquisition time was significantly shorter (123.8 ± 28.4 [SD] s vs. 267.2 ± 39.3 [SD] s for BH-SEG CMR; P < 0.0001) at the penalty of a longer reconstruction time (271.4 ± 68.7 [SD] s vs. 9.9 ± 2.1 [SD] s for BH-SEG CMR; P < 0.0001). In patients without arrhythmia or dyspnea, FB-CS CMR provided subjective image quality that was not different from that of BH-SEG CMR (P = 0.13). FB-CS CMR improved image quality in patients with arrhythmia (n = 18; P = 0.002) or dyspnea (n = 7; P = 0.02), and the edge sharpness was improved at end-systole and end-diastole (P = 0.0001). No differences were observed between the two techniques in ventricular volumes and ejection fractions, left ventricular mass or global circumferential strain in patients in sinus rhythm or with cardiac arrhythmia. CONCLUSION: This new FB-CS CMR addresses respiratory motion and arrhythmia-related artifacts without compromising the reliability of ventricular functional assessment.

16.
Curr Med Imaging ; 2023 Mar 30.
Article in English | MEDLINE | ID: mdl-36999185

ABSTRACT

Wavelets are defined as mathematical functions that segment the data into different frequency levels. We can easily capture the fine and coarse details of an image or signal referred to as a subband. And it also helps in subband thresholding to achieve good compression performance. In recent days in telemedicine services, the handling of medical images is prominently increasing and it leads to the demand for medical image compression. While compressing the medical images, we have to concentrate on the data that holds important information, and at the same time, it must retain the image quality. Near-Lossless compression plays an essential role to achieve a better compression ratio than lossy compression and provides better quality than lossless compression. In this paper, we analyzed the sub-banding of Discrete Wavelet Transform (DWT) using different types of wavelets and made an optimal selection of wavelets for subband thresholding to attain a good compression performance with an application to medical images. We used Set Partitioning In Hierarchical Trees (SPIHT) compression scheme to test the compression performance of different wavelets. The Peak Signal to Noise Ratio (PSNR), Bits Per Pixel (BPP), Compression Ratio, and percentage of number of zeros are used as metrics to assess the performance of all the selected wavelets. And to find out its efficiency in possessing the essential information of medical images, the subband of the selected wavelets is further utilized to devise the near-lossless compression scheme for medical images.

17.
Sensors (Basel) ; 23(3)2023 Jan 23.
Article in English | MEDLINE | ID: mdl-36772337

ABSTRACT

The usage of media such as images and videos has been extensively increased in recent years. It has become impractical to store images and videos acquired by camera sensors in their raw form due to their huge storage size. Generally, image data is compressed with a compression algorithm and then stored or transmitted to another platform. Thus, image compression helps to reduce the storage size and transmission cost of the images and videos. However, image compression might cause visual artifacts, depending on the compression level. In this regard, performance evaluation of the compression algorithms is an essential task needed to reconstruct images with visually or near-visually lossless quality in case of lossy compression. The performance of the compression algorithms is assessed by both subjective and objective image quality assessment (IQA) methodologies. In this paper, subjective and objective IQA methods are integrated to evaluate the range of the image quality metrics (IQMs) values that guarantee the visually or near-visually lossless compression performed by the JPEG 1 standard (ISO/IEC 10918). A novel "Flicker Test Software" is developed for conducting the proposed subjective and objective evaluation study. In the flicker test, the selected test images are subjectively analyzed by subjects at different compression levels. The IQMs are calculated at the previous compression level, when the images were visually lossless for each subject. The results analysis shows that the objective IQMs with more closely packed values having the least standard deviation that guaranteed the visually lossless compression of the images with JPEG 1 are the feature similarity index measure (FSIM), the multiscale structural similarity index measure (MS-SSIM), and the information content weighted SSIM (IW-SSIM), with average values of 0.9997, 0.9970, and 0.9970 respectively.

18.
Sensors (Basel) ; 23(3)2023 Feb 01.
Article in English | MEDLINE | ID: mdl-36772613

ABSTRACT

It has always been a major issue for a hospital to acquire real-time information about a patient in emergency situations. Because of this, this research presents a novel high-compression-ratio and real-time-process image compression very-large-scale integration (VLSI) design for image sensors in the Internet of Things (IoT). The design consists of a YEF transform, color sampling, block truncation coding (BTC), threshold optimization, sub-sampling, prediction, quantization, and Golomb-Rice coding. By using machine learning, different BTC parameters are trained to achieve the optimal solution given the parameters. Two optimal reconstruction values and bitmaps for each 4 × 4 block are achieved. An image is divided into 4 × 4 blocks by BTC for numerical conversion and removing inter-pixel redundancy. The sub-sampling, prediction, and quantization steps are performed to reduce redundant information. Finally, the value with a high probability will be coded using Golomb-Rice coding. The proposed algorithm has a higher compression ratio than traditional BTC-based image compression algorithms. Moreover, this research also proposes a real-time image compression chip design based on low-complexity and pipelined architecture by using TSMC 0.18 µm CMOS technology. The operating frequency of the chip can achieve 100 MHz. The core area and the number of logic gates are 598,880 µm2 and 56.3 K, respectively. In addition, this design achieves 50 frames per second, which is suitable for real-time CMOS image sensor compression.

19.
Sensors (Basel) ; 23(2)2023 Jan 09.
Article in English | MEDLINE | ID: mdl-36679527

ABSTRACT

Small satellites empower different applications for an affordable price. By dealing with a limited capacity for using instruments with high power consumption or high data-rate requirements, small satellite missions usually focus on specific monitoring and observation tasks. Considering that multispectral and hyperspectral sensors generate a significant amount of data subjected to communication channel impairments, bandwidth constraint is an important challenge in data transmission. That issue is addressed mainly by source and channel coding techniques aiming at an effective transmission. This paper targets a significant further bandwidth reduction by proposing an on-the-fly analysis on the satellite to decide which information is effectively useful before coding and transmitting. The images are tiled and classified using a set of detection algorithms after defining the least relevant content for general remote sensing applications. The methodology makes use of the red-band, green-band, blue-band, and near-infrared-band measurements to perform the classification of the content by managing a cloud detection algorithm, a change detection algorithm, and a vessel detection algorithm. Experiments for a set of typical scenarios of summer and winter days in Stockholm, Sweden, were conducted, and the results show that non-important content can be identified and discarded without compromising the predefined useful information for water and dry-land regions. For the evaluated images, only 22.3% of the information would need to be transmitted to the ground station to ensure the acquisition of all the important content, which illustrates the merits of the proposed method. Furthermore, the embedded platform's constraints regarding processing time were analyzed by running the detection algorithms on Unibap's iX10-100 space cloud platform.


Subject(s)
Data Compression , Algorithms , Seasons , Telemetry , Sweden
20.
Asian J Endosc Surg ; 16(2): 255-261, 2023 Apr.
Article in English | MEDLINE | ID: mdl-36479621

ABSTRACT

INTRODUCTION: Telerobotic surgery relies on communication lines, causing delays, and video information requires pre-transmission compression. Such delays and video degradation will continue to be unavoidable making communication conditions verification essential. Understanding the network specification values required for telerobotic surgery entails determining acceptable levels of delay and degradation due to the video compression and restoration processes during surgery. METHODS: The hinotori™ surgical robot from Medicaroid was used. Eight surgeons, skilled in robotic surgery, performed gastrectomy or rectal resection on pigs. Image compression (bitrate: 120, 60, 30, 20, 10 Mbps) was random, changing encoder settings during surgery, and delay times (30, 50, 100, 150 milliseconds) were pseudo-randomly inserted, changing emulator settings. Acceptable video levels were evaluated. Subjective evaluations by surgeons and evaluators regarding image degradation and operability, and objective evaluations of image degradation and operability were given five-point ratings. RESULTS: Regarding delay time, 30 and 50 millisecond periods garnered average ratings of 3.6 and 4.0, respectively, signifying that surgery was feasible. However, at 100 and 150 millisecond, average ratings were 2.9 and 2.3, respectively, indicating surgery was not feasible for the most part in these cases. The average rating for image compression was 4.0 or higher for bitrates of 20, 30, 60, and 120 Mbps, suggesting that surgery is possible even at bitrates as low as 10 Mbps, with an average rating of 4.0. CONCLUSION: In remote robotic surgery using the hinotori™, image compression and delay time are largely acceptable, so surgery can be safely performed.


Subject(s)
Data Compression , Robotic Surgical Procedures , Robotics , Surgeons , Telemedicine , Humans , Animals , Swine , Telemedicine/methods , Robotics/methods
SELECTION OF CITATIONS
SEARCH DETAIL
...