ISSN: 2074-9074 (Print)
ISSN: 2074-9082 (Online)
DOI: https://doi.org/10.5815/ijigsp
Website: https://www.mecs-press.org/ijigsp
Published By: MECS Press
Frequency: 6 issues per year
Number(s) Available: 133
IJIGSP is committed to bridge the theory and practice of images, graphics, and signal processing. From innovative ideas to specific algorithms and full system implementations, IJIGSP publishes original, peer-reviewed, and high quality articles in the areas of images, graphics, and signal processing. IJIGSP is a well-indexed scholarly journal and is indispensable reading and references for people working at the cutting edge of images, graphics, and signal processing applications.
IJIGSP has been abstracted or indexed by several world class databases: Scopus, Google Scholar, Microsoft Academic Search, CrossRef, Baidu Wenku, IndexCopernicus, IET Inspec, EBSCO, JournalSeek, ULRICH's Periodicals Directory, WorldCat, Scirus, Academic Journals Database, Stanford University Libraries, Cornell University Library, UniSA Library, CNKI Scholar, ProQuest, J-Gate, ZDB, BASE, OhioLINK, iThenticate, Open Access Articles, Open Science Directory, National Science Library of Chinese Academy of Sciences, The HKU Scholars Hub, etc..
IJIGSP Vol. 16, No. 5, Oct. 2024
REGULAR PAPERS
Low-light scenes are characterized by the loss of illumination, the noise, the color distortion and serious information degradation. The low-light image enhancement is a significant part of computer vision technology. The low-light image enhancement methods aim to an image recover to a normal-light image from dark one, a noise-free image from a noisy one, a clear image from distorting one. In this paper, the low-light image enhancement technology based on Retinex-based deep network combined with the image processing-based module is proposed. The proposed technology combines the use of traditional and deep learning methodologies, designed within a simple yet efficient architectural framework that focuses on essential feature extraction. The proposed preprocessing module of low-light image enhancement is centered on the unique knowledge and features of an image. The choice of a color model and a technique of an image transformation depends on an image dynamic range to ensure high results in terms of transfer a color, detail integrity and overall visual quality. The proposed Retinex-based deep network has been trained and tested on transformed images by means of preprocessing module that leads to an effective supervised approach to low-light image enhancement and provide superior performance. The proposed preprocessing module is implemented as an independent image enhancement module in a computer system of an image analysis and as the component module in a neural network system of an image analysis. Experimental results on the low light paired dataset show that the proposed method can reduce noise and artifacts in low-light images, and can improve contrast and brightness, demonstrating its advantages. The proposed approach injects new ideas into low light image enhancement, providing practical applications in challenging low-light scenarios.
[...] Read more.Deep learning based speech enhancement approaches provides better perceptual quality and better intelligibility. But most of the speech enhancement methods available in literature estimates enhanced speech using processed amplitude, energy, MFCC spectrum, etc along with noisy phase. Because of difficult in estimating clean speech phase from noisy speech the noisy phase is still using in reconstruction of enhanced speech. Some methods are developed for estimating clean speech phase and it is observed that it is complex for estimation. To avoid difficulty and for better performance rather than using Discrete Fourier Transform (DFT) the Discrete Cosine Transform (DCT) and Discrete Sine Transform (DST) based convolution neural networks are proposed for better intelligibility and improved performance. However, the algorithms work either features of time domain or features of frequency domain. To have advantage of both time domain and frequency domain here the fusion of DCT and time domain approach is proposed. In this work DCT Dense Convolutional Recurrent Network (DCTDCRN), DST Convolutional Gated Recurrent Neural Network (DSTCGRU), DST Convolution Long Short term Memory (DSTCLSTM) and DST Convolutional Gated Recurrent Neural Network (DSTDCRN) are proposed for speech enhancement. These methods are providing superior performance and less processing difficulty when compared to the state of art methods. The proposed DCT based methods are used further in developing joint time and magnitude based speech enhancement method. Simulation results show superior performance than baseline methods for joint time and frequency based processing. Also results are analyzed using objective performance measures like Signal to Noise Ratio (SNR), Perceptual Evaluation of Speech Quality (PESQ) and Short-Time Objective Intelligibility (STOI).
[...] Read more.The research article presents a robust solution to detect surgical masks using a combination of deep learning techniques. The proposed method utilizes the SAM to detect the presence of masks in images, while EfficientNet is employed for feature extraction and classification of mask type. The compound scaling method is used to distinguish between surgical and normal masks in the data set of 2000 facial photos, divided into 60% training, 20% validation, and 20% testing sets. The machine learning model is trained on the data set to learn the discriminative characteristics of each class and achieve high accuracy in mask detection. To handle the variability of mask types, the study applies various versions of EfficientNet, and the highest accuracy of 97.5% is achieved using EfficientNetV2L, demonstrating the effectiveness of the proposed method in detecting masks of different complexities and designs.
[...] Read more.The article is devoted to the modified multidimensional Kalman filter with Chebyshev points development to solve the task of diagnosing and parring off failures in the measurement channels of complex dynamic objects automatic control system, which will provide a more accurate and reliable assessment of system state in the presence of outliers in the data. An implementation of the proposed modified multidimensional Kalman filter with Chebyshev points is proposed in the form of a modified recurrent neural network containing a failure diagnostics layer, a failure parry layer, a filtering and smoothing layer, and a results aggregation layer. This structure of the modified recurrent neural network made it possible to solve the main problems of the method of diagnosing and parring off failures of the measuring channels of complex dynamic objects automatic control system, such as diagnosing failures with an accuracy of 0.99802, fending off failures with an accuracy of 0.99796, and assessing the state of the system with an accuracy of 0.99798. It is proposed to use a modified loss function of a recurrent neural network as a general loss function for diagnostics, fault restoring and system state assessment, which makes it possible to avoid retraining when there are a large number of parameters or insufficient data. It has been experimentally proven that the loss function remains stable on both the training and validation data sets for 1000 training epochs and does not go beyond –2.5 % to +2.5 %, which indicates a low-risk overtraining or undertraining of the model. It has been experimentally confirmed that the use of a modified recurrent neural network in solving the task of diagnosing and parring off failures of the measuring channels of complex dynamic objects automatic control system is appropriate in comparison with a radial basis functions neural network and a multidimensional Kalman filter without a neural network implementation, based on metrics such as the root mean square deviation, mean absolute error, mean absolute percentage error, coefficient of determination for the accuracy of reproducing previous data, and coefficient of determination for the accuracy of predicting future values. For example, the value of the standard deviation of the modified recurrent neural network is 0.00226, which is 1.65 times less than the radial basis function neural network and 2.20 times less than the multidimensional Kalman filter without a neural network implementation.
[...] Read more.Image enhancement in the pre-processing stage of biometric systems is a crucial task in image analysis. Image degradation significantly impacts the biometric system’s performance, which occurs during biometric image capturing, and demands an appropriate enhancement technique. Generally, biometric images are mixed with full of noise and deformation due to the image capturing process, pressure with sensor surface, and photometric transformations. Therefore, these systems highly demand pure discriminative features for identification, and the system’s performance heavily depends on such quality features. Hence, enhancement techniques are typically applied in captured images before go into the feature extraction stage in any biometrics recognition pipeline. In palmprint biometrics, contact-based palmprints consist of several ridges, creases, skin wrinkles, and palm lines, leading to several spurious minutiae during feature extraction. Therefore, selecting an appropriate enhancement technique to make them smooth becomes a significant task. The feature extraction process necessitates a completely pre-processed image to locate key features, which significantly influences the identification performance. Thus, the palmprint system’s performance can be enhanced by exploiting competent enhancement filters. Palmprints have reported a lack of novelty in enhancement techniques rather than more centering on feature encoding and matching techniques. Some enhancement techniques in fingerprints were adopted for palmprints in the past. However, there is no clear evidence of their impact on image quality, and to what extent they affect the quality in specific applications. Further, frequency level filters such as the Gabor and Fourier transforms exploited in fingerprints would not be practically feasible for palmprints due to the computational cost for a larger surface area. Thus, it opens an investigation for utilising enhancement techniques in degraded palmprints in a different direction. This work delves into a preliminary investigation of the usage of existing enhancement techniques utilised for pre-processing of contact fingerprint images and biomedical images. Several enhancement filters were experimented on severely degraded palmprints, and the image quality was measured using image quality metrics. The High-boost filter comparatively performed better peak-signal-to-noise ratio, while other filters affected the image quality. The experiment is further extended to compare the identification performance of degraded palmprints in the presence and absence of enhanced images. The results reveal that the enhanced images with the filter that has the highest peak signal-to-noise ratio (High boost filter) only show an increased genuine accept rate compared to the ground truth value. The High-boost filter slightly decreases the system’s equal error rate, indicating the potential of exploiting a pre-enhancement technique on degraded prints with an appropriate filter without compromising the raw image quality. Optimised enhancement techniques could be another initiative for addressing the severity of image degradation in contact handprints. Doing so they could be successfully exploited in civilian applications like access control along with other applications. Further, utilising appropriate enhancement filters for degraded palmprints can enhance the existing palmprint system’s performance in forensics, and make it more reliable for legal outcomes.
[...] Read more.Cyclones, with their high-speed winds and enormous quantities of rainfall, represent severe threats to global coastal regions. The ability to quickly and accurately identify cyclonic cloud formations is critical for the effective deployment of disaster preparedness measures. Our study focuses on a unique technique for precise delineation of cyclonic cloud regions in satellite imagery, concentrating on images from the Indian weather satellite INSAT-3D. This novel approach manages to achieve considerable improvements in cyclone monitoring by leveraging the image capture capabilities of INSAT-3D. It introduces a refined image processing continuum that extracts cloud attributes from infrared imaging in a comprehensive manner. This includes transformations and normalization techniques, further augmenting the pursuit of accuracy. A key feature of the study's methodology is the use of an adaptive threshold to correct complications related to luminosity and contrast; this enhances the detection accuracy of the cyclonic cloud formations substantially. The study further improves the preciseness of cloud detection by employing a modified contour detection algorithm that operates based on predefined criteria. The methodology has been designed to be both flexible and adaptable, making it highly effective while dealing with a wide array of environmental conditions. The utilization of INSAT-3D satellite images maximizes the performing capability of the technique in various situational contexts.
[...] Read more.Static weather conditions like fog, haze, and mist in hilly and urban areas cause reduced road visibility. Due to different weather conditions, autonomous vehicles cannot identify objects, traffic signs, and signals. So, this leads to many accidents, endangering living beings’ lives. The significance of this work lies in its aim to develop a model that can provide clear visibility for autonomous vehicles during bad weather conditions. Image restoration is one of the important issues in the image processing field as the images may be of low contrast and quality due to restricted visibility and, the development of a model that reduces the halos and artifacts produced in the image using the Median Channel based Image Restoration (MCIR) technique has significant research value. In this technique, the image restoration is done by calculating the atmospheric light and the transmission map using the MCIR technique and patching the pixels for different patch sizes. The Dark Channel Prior (DCP) method and the MCIR technique are compared for different patch sizes by evaluating the output images using the PSNR, SSIM, and MSE metrics. The results show that MCIR technique provides better Peak Signal-to-Noise Ratio (PSNR), Mean Square Error (MSE), and Structural Similarity Index Measure (SSIM) values than the DCP method with reduced halos and artifacts. This result highlights the effectiveness of the MCIR technique for image restoration. The software model developed can be applied to autonomous vehicles and surveillance cameras for the restoration of the images, which can improve their performance and safety.
[...] Read more.In this paper, we developed a new approach to solve the problem of infective endocarditis (IE) diagnostics based on intelligent analysis of patients’ echocardiography images. The approach is based on echocardiography segmentation results and detection of valvular anomalies (namely vegetations). In this article for the first time investigates CNNs and Visual Transformers (ViT) based segmentation methods within the framework of the vegetation segmentation task on echocardiography images. Additionally, ensemble methods for combining segmentation models using a new method of models competition for data points were proposed. Furthermore, we investigated methods for aggregating the results of the ensemble based on a new meta-model, pointwise weighted aggregation, which weighs the results of each model pixel by pixel. The last proposed step was to automatically calculate the volume of segmented vegetation to determine the degree of disease and the need for urgent surgical intervention. For the studied and proposed methods, the following ensemble segmentation accuracy was achieved on the test dataset: iou 0.7822, dice score 0.886. The proposed empirical algorithm for calculating the volume of vegetations provided the basis for further improvements of the studied approach. The results obtained indicate the great potential of the developed approaches in clinical practice.
[...] Read more.Denoising is a vital aspect of image preprocessing, often explored to eliminate noise in an image to restore its proper characteristic formation and clarity. Unfortunately, noise often degrades the quality of valuable images, making them meaningless for practical applications. Several methods have been deployed to address this problem, but the quality of the recovered images still requires enhancement for efficient applications in practice. In this paper, a wavelet-based universal thresholding technique that possesses the capacity to optimally denoise highly degraded noisy images with both uniform and non-uniform variations in illumination and contrast is proposed. The proposed method, herein referred to as the modified wavelet-based universal thresholding (MWUT), compared to three state-of-the-art denoising techniques, was employed to denoise five noisy images. In order to appraise the qualities of the images obtained, seven performance indicators comprising the Root Mean Square Error (RMSE), Mean Absolute Error (MAE), Structural Content (SC), Peak Signal to Noise Ratio (PSNR), Structural Similarity Index Method (SSIM), Signal-to-Reconstruction-Error Ratio (SRER), Blind Spatial Quality Evaluator (NIQE), and Blind/Referenceless Image Spatial Quality Evaluator (BRISQUE) were employed. The first five indicators – RMSE, MAE, SC, PSNR, SSIM, and SRER- are reference indicators, while the remaining two – NIQE and BRISQUE- are referenceless. For the superior performance of the proposed wavelet threshold algorithm, the SC, PSNR, SSIM, and SRER must be higher, while lower values of NIQE, BRISQUE, RMSE, and MAE are preferred. A higher and better value of PSNR, SSIM, and SRER in the final results shows the superior performance of our proposed MWUT denoising technique over the preliminaries. Lower NIQE, BRISQUE, RMSE, and MAE values also indicate higher and better image quality results using the proposed modified wavelet-based universal thresholding technique over the existing schemes. The modified wavelet-based universal thresholding technique would find practical applications in digital image processing and enhancement.
[...] Read more.Image Processing is the art of examining, identifying and judging the significances of the Images. Image enhancement refers to attenuation, or sharpening, of image features such as edgels, boundaries, or contrast to make the processed image more useful for analysis. Image enhancement procedures utilize the computers to provide good and improved images for study by the human interpreters. In this paper we proposed a novel method that uses the Genetic Algorithm with Multi-objective criteria to find more enhance version of images. The proposed method has been verified with benchmark images in Image Enhancement. The simple Genetic Algorithm may not explore much enough to find out more enhanced image. In the proposed method three objectives are taken in to consideration. They are intensity, entropy and number of edgels. Proposed algorithm achieved automatic image enhancement criteria by incorporating the objectives (intensity, entropy, edges). We review some of the existing Image Enhancement technique. We also compared the results of our algorithms with another Genetic Algorithm based techniques. We expect that further improvements can be achieved by incorporating linear relationship between some other techniques.
[...] Read more.Image analysis belongs to the area of computer vision and pattern recognition. These areas are also a part of digital image processing, where researchers have a great attention in the area of content retrieval information from various types of images having complex background, low contrast background or multi-spectral background etc. These contents may be found in any form like texture data, shape, and objects. Text Region Extraction as a content from an mage is a class of problems in Digital Image Processing Applications that aims to provides necessary information which are widely used in many fields medical imaging, pattern recognition, Robotics, Artificial intelligent Transport systems etc. To extract the text data information has becomes a challenging task. Since, Text extraction are very useful for identifying and analysis the whole information about image, Therefore, In this paper, we propose a unified framework by combining morphological operations and Genetic Algorithms for extracting and analyzing the text data region which may be embedded in an image by means of variety of texts: font, size, skew angle, distortion by slant and tilt, shape of the object which texts are on, etc. We have established our proposed methods on gray level image sets and make qualitative and quantitative comparisons with other existing methods and concluded that proposed method is better than others.
[...] Read more.In the field of medical image analysis, supervised deep learning strategies have achieved significant development, while these methods rely on large labeled datasets. Self-Supervised learning (SSL) provides a new strategy to pre-train a neural network with unlabeled data. This is a new unsupervised learning paradigm that has achieved significant breakthroughs in recent years. So, more and more researchers are trying to utilize SSL methods for medical image analysis, to meet the challenge of assembling large medical datasets. To our knowledge, so far there still a shortage of reviews of self-supervised learning methods in the field of medical image analysis, our work of this article aims to fill this gap and comprehensively review the application of self-supervised learning in the medical field. This article provides the latest and most detailed overview of self-supervised learning in the medical field and promotes the development of unsupervised learning in the field of medical imaging. These methods are divided into three categories: context-based, generation-based, and contrast-based, and then show the pros and cons of each category and evaluates their performance in downstream tasks. Finally, we conclude with the limitations of the current methods and discussed the future direction.
[...] Read more.This article proposes a receiving device in which arbitrary input signals are subject to pre-detector processing for the subsequent implementation of the idea of compressing broadband modulated pulses with a matched filter to increase the signal-to-noise ratio and improve resolution. For this purpose, a model of a dispersive delay line is developed based on series-connected high-frequency time delay lines with taps in the form of bandpass filters, and analysis of this model is performed as a part of the radio receiving device with chirp signal compression. The article presents the mathematical description of the processes of formation and compression of chirp signals based on their matched filtering using the developed model and proposes the block diagram of a radio receiving device using the principle of compression of received signals. The proposed model can be implemented in devices for receiving unknown signals, in particular in passive radar. It also can be used for studying signal compression processes based on linear frequency modulation in traditional radar systems.
[...] Read more.Mushrooms are the most familiar delicious food which is cholesterol free as well as rich in vitamins and minerals. Though nearly 45,000 species of mushrooms have been known throughout the world, most of them are poisonous and few are lethally poisonous. Identifying edible or poisonous mushroom through the naked eye is quite difficult. Even there is no easy rule for edibility identification using machine learning methods that work for all types of data. Our aim is to find a robust method for identifying mushrooms edibility with better performance than existing works. In this paper, three ensemble methods are used to detect the edibility of mushrooms: Bagging, Boosting, and random forest. By using the most significant features, five feature sets are made for making five base models of each ensemble method. The accuracy is measured for ensemble methods using five both fixed feature set-based models and randomly selected feature set based models, for two types of test sets. The result shows that better performance is obtained for methods made of fixed feature sets-based models than randomly selected feature set-based models. The highest accuracy is obtained for the proposed model-based random forest for both test sets.
[...] Read more.Breast cancer can be detected by mammograms, but not all of them are of high enough quality to be diagnosed by physicians or radiologists. Therefore, denoising and contrast enhancement in the image are issues that need to be addressed. There are numerous techniques to reduce noise and enhance contrast; the most popular of which incorporate spatial filters and histogram equalization. However, these techniques occasionally result in image blurring, particularly around the edges. The purpose of this article is to propose a technique that uses wavelet denoising in conjunction with top-hat and bottom-hat morphological transforms in the wavelet domain to reduce noise and image quality without distorting the image. Use five wavelet functions to test the proposed method: Haar, Daubechies (db3), Coiflet (coif3), Symlet (sym3), and Biorthogonal (bior1.3); each wavelet function employs levels 1 through 4 with four types of wavelet shrinkage: Bayer, Visu, SURE, and Normal. Three flat structuring elements in the shapes of a disk, a square, and a diamond with sizes 2, 5, 10, 15, 20, and 30 are utilized for top-hat and bottom-hat morphological transforms. To determine optimal parameters, the proposed method is applied to mdb001 mammogram (mini MIAS database) contaminated with Gaussian noise with SD, ? = 20. Based on the quality assessment quantities, the Symlet wavelet (sym3) at level 3, with Visu shrinkage and diamond structuring element size 5 produced the best results (MSE = 50.020, PSNR = 31.140, SSIM = 0.407, and SC = 1.008). The results demonstrate the efficacy of the proposed method.
[...] Read more.E-healthcare systems (EHSD), medical communications, digital imaging (DICOM) things have gained popularity over the past decade as they have become the top contenders for interoperability and adoption as a global standard for transmitting and communicating medical data. Security is a growing issue as EHSD and DICOM have grown more usable on any-to-any devices. The goal of this research is to create a privacy-preserving encryption technique for EHSD rapid communication with minimal storage. A new 2D logistic-sine chaotic map (2DLSCM) is used to design the proposed encryption method, which has been developed specifically for peer-to-peer communications via unique keys. Through the 3D Lorenz map which feeds the initial values to it, the 2DLSCM is able to provide a unique keyspace of 2544 bits (2^544bits) in each go of peer-to-peer paired transmission. Permutation-diffusion design is used in the encryption process, and 2DLSCM with 3DLorenz system are used to generate unique initial values for the keys. Without interfering with real-time medical transmission, the approach can quickly encrypt any EHSD image and DICOM objects. To assess the method, five distinct EHSD images of different kinds, sizes, and quality are selected. The findings indicate strong protection, speed, and scalability when compared to existing similar methods in literature.
[...] Read more.During past few years, brain tumor segmentation in magnetic resonance imaging (MRI) has become an emergent research area in the ?eld of medical imaging system. Brain tumor detection helps in finding the exact size and location of tumor. An efficient algorithm is proposed in this paper for tumor detection based on segmentation and morphological operators. Firstly quality of scanned image is enhanced and then morphological operators are applied to detect the tumor in the scanned image.
[...] Read more.Ultrasound based breast screening is gaining attention recently especially for dense breast. The technological advancement, cancer awareness, and cost-safety-availability benefits lead rapid rise of breast ultrasound market. The irregular shape, intensity variation, and additional blood vessels of malignant cancer are distinguishable in ultrasound images from the benign phase. However, classification of breast cancer using ultrasound images is a difficult process owing to speckle noise and complex textures of breast. In this paper, a breast cancer classification method is presented using VGG16 model based transfer learning approach. We have used median filter to despeckle the images. The layers for convolution process of the pretrained VGG16 model along with the maxpooling layers have been used as feature extractor and a proposed fully connected two layers deep neural network has been designed as classifier. Adam optimizer is used with learning rate of 0.001 and binary cross-entropy is chosen as the loss function for model optimization. Dropout of hidden layers is used to avoid overfitting. Breast Ultrasound images from two databases (total 897 images) have been combined to train, validate and test the performance and generalization strength of the classifier. Experimental results showed the training accuracy as 98.2% and testing accuracy as 91% for blind testing data with a reduced of computational complexity. Gradient class activation mapping (Grad-CAM) technique has been used to visualize and check the targeted regions localization effort at the final convolutional layer and found as noteworthy. The outcomes of this work might be useful for the clinical applications of breast cancer diagnosis.
[...] Read more.Image Processing is the art of examining, identifying and judging the significances of the Images. Image enhancement refers to attenuation, or sharpening, of image features such as edgels, boundaries, or contrast to make the processed image more useful for analysis. Image enhancement procedures utilize the computers to provide good and improved images for study by the human interpreters. In this paper we proposed a novel method that uses the Genetic Algorithm with Multi-objective criteria to find more enhance version of images. The proposed method has been verified with benchmark images in Image Enhancement. The simple Genetic Algorithm may not explore much enough to find out more enhanced image. In the proposed method three objectives are taken in to consideration. They are intensity, entropy and number of edgels. Proposed algorithm achieved automatic image enhancement criteria by incorporating the objectives (intensity, entropy, edges). We review some of the existing Image Enhancement technique. We also compared the results of our algorithms with another Genetic Algorithm based techniques. We expect that further improvements can be achieved by incorporating linear relationship between some other techniques.
[...] Read more.Denoising is a vital aspect of image preprocessing, often explored to eliminate noise in an image to restore its proper characteristic formation and clarity. Unfortunately, noise often degrades the quality of valuable images, making them meaningless for practical applications. Several methods have been deployed to address this problem, but the quality of the recovered images still requires enhancement for efficient applications in practice. In this paper, a wavelet-based universal thresholding technique that possesses the capacity to optimally denoise highly degraded noisy images with both uniform and non-uniform variations in illumination and contrast is proposed. The proposed method, herein referred to as the modified wavelet-based universal thresholding (MWUT), compared to three state-of-the-art denoising techniques, was employed to denoise five noisy images. In order to appraise the qualities of the images obtained, seven performance indicators comprising the Root Mean Square Error (RMSE), Mean Absolute Error (MAE), Structural Content (SC), Peak Signal to Noise Ratio (PSNR), Structural Similarity Index Method (SSIM), Signal-to-Reconstruction-Error Ratio (SRER), Blind Spatial Quality Evaluator (NIQE), and Blind/Referenceless Image Spatial Quality Evaluator (BRISQUE) were employed. The first five indicators – RMSE, MAE, SC, PSNR, SSIM, and SRER- are reference indicators, while the remaining two – NIQE and BRISQUE- are referenceless. For the superior performance of the proposed wavelet threshold algorithm, the SC, PSNR, SSIM, and SRER must be higher, while lower values of NIQE, BRISQUE, RMSE, and MAE are preferred. A higher and better value of PSNR, SSIM, and SRER in the final results shows the superior performance of our proposed MWUT denoising technique over the preliminaries. Lower NIQE, BRISQUE, RMSE, and MAE values also indicate higher and better image quality results using the proposed modified wavelet-based universal thresholding technique over the existing schemes. The modified wavelet-based universal thresholding technique would find practical applications in digital image processing and enhancement.
[...] Read more.In the field of medical image analysis, supervised deep learning strategies have achieved significant development, while these methods rely on large labeled datasets. Self-Supervised learning (SSL) provides a new strategy to pre-train a neural network with unlabeled data. This is a new unsupervised learning paradigm that has achieved significant breakthroughs in recent years. So, more and more researchers are trying to utilize SSL methods for medical image analysis, to meet the challenge of assembling large medical datasets. To our knowledge, so far there still a shortage of reviews of self-supervised learning methods in the field of medical image analysis, our work of this article aims to fill this gap and comprehensively review the application of self-supervised learning in the medical field. This article provides the latest and most detailed overview of self-supervised learning in the medical field and promotes the development of unsupervised learning in the field of medical imaging. These methods are divided into three categories: context-based, generation-based, and contrast-based, and then show the pros and cons of each category and evaluates their performance in downstream tasks. Finally, we conclude with the limitations of the current methods and discussed the future direction.
[...] Read more.Ultrasound based breast screening is gaining attention recently especially for dense breast. The technological advancement, cancer awareness, and cost-safety-availability benefits lead rapid rise of breast ultrasound market. The irregular shape, intensity variation, and additional blood vessels of malignant cancer are distinguishable in ultrasound images from the benign phase. However, classification of breast cancer using ultrasound images is a difficult process owing to speckle noise and complex textures of breast. In this paper, a breast cancer classification method is presented using VGG16 model based transfer learning approach. We have used median filter to despeckle the images. The layers for convolution process of the pretrained VGG16 model along with the maxpooling layers have been used as feature extractor and a proposed fully connected two layers deep neural network has been designed as classifier. Adam optimizer is used with learning rate of 0.001 and binary cross-entropy is chosen as the loss function for model optimization. Dropout of hidden layers is used to avoid overfitting. Breast Ultrasound images from two databases (total 897 images) have been combined to train, validate and test the performance and generalization strength of the classifier. Experimental results showed the training accuracy as 98.2% and testing accuracy as 91% for blind testing data with a reduced of computational complexity. Gradient class activation mapping (Grad-CAM) technique has been used to visualize and check the targeted regions localization effort at the final convolutional layer and found as noteworthy. The outcomes of this work might be useful for the clinical applications of breast cancer diagnosis.
[...] Read more.Image analysis belongs to the area of computer vision and pattern recognition. These areas are also a part of digital image processing, where researchers have a great attention in the area of content retrieval information from various types of images having complex background, low contrast background or multi-spectral background etc. These contents may be found in any form like texture data, shape, and objects. Text Region Extraction as a content from an mage is a class of problems in Digital Image Processing Applications that aims to provides necessary information which are widely used in many fields medical imaging, pattern recognition, Robotics, Artificial intelligent Transport systems etc. To extract the text data information has becomes a challenging task. Since, Text extraction are very useful for identifying and analysis the whole information about image, Therefore, In this paper, we propose a unified framework by combining morphological operations and Genetic Algorithms for extracting and analyzing the text data region which may be embedded in an image by means of variety of texts: font, size, skew angle, distortion by slant and tilt, shape of the object which texts are on, etc. We have established our proposed methods on gray level image sets and make qualitative and quantitative comparisons with other existing methods and concluded that proposed method is better than others.
[...] Read more.Mushrooms are the most familiar delicious food which is cholesterol free as well as rich in vitamins and minerals. Though nearly 45,000 species of mushrooms have been known throughout the world, most of them are poisonous and few are lethally poisonous. Identifying edible or poisonous mushroom through the naked eye is quite difficult. Even there is no easy rule for edibility identification using machine learning methods that work for all types of data. Our aim is to find a robust method for identifying mushrooms edibility with better performance than existing works. In this paper, three ensemble methods are used to detect the edibility of mushrooms: Bagging, Boosting, and random forest. By using the most significant features, five feature sets are made for making five base models of each ensemble method. The accuracy is measured for ensemble methods using five both fixed feature set-based models and randomly selected feature set based models, for two types of test sets. The result shows that better performance is obtained for methods made of fixed feature sets-based models than randomly selected feature set-based models. The highest accuracy is obtained for the proposed model-based random forest for both test sets.
[...] Read more.Diabetic retinopathy is one of the most serious eye diseases and can lead to permanent blindness if not diagnosed early. The main cause of this is diabetes. Not every diabetic will develop diabetic retinopathy, but the risk of developing diabetes is undeniable. This requires the early diagnosis of Diabetic retinopathy. Segmentation is one of the approaches which is useful for detecting the blood vessels in the retinal image. This paper proposed the three models based on a deep learning approach for recognizing blood vessels from retinal images using region-based segmentation techniques. The proposed model consists of four steps preprocessing, Augmentation, Model training, and Performance measure. The augmented retinal images are fed to the three models for training and finally, get the segmented image. The proposed three models are applied on publically available data set of DRIVE, STARE, and HRF. It is observed that more thin blood vessels are segmented on the retinal image in the HRF dataset using model-3. The performance of proposed three models is compare with other state-of-art-methods of blood vessels segmentation of DRIVE, STARE, and HRF datasets.
[...] Read more.Image reconstruction is the process of generating an image of an object from the signals captured by the scanning machine. Medical imaging is an interdisciplinary field combining physics, biology, mathematics and computational sciences. This paper provides a complete overview of image reconstruction process in MRI (Magnetic Resonance Imaging). It reviews the computational aspect of medical image reconstruction. MRI is one of the commonly used medical imaging techniques. The data collected by MRI scanner for image reconstruction is called the k-space data. For reconstructing an image from k-space data, there are various algorithms such as Homodyne algorithm, Zero Filling method, Dictionary Learning, and Projections onto Convex Set method. All the characteristics of k-space data and MRI data collection technique are reviewed in detail. The algorithms used for image reconstruction discussed in detail along with their pros and cons. Various modern magnetic resonance imaging techniques like functional MRI, diffusion MRI have also been introduced. The concepts of classical techniques like Expectation Maximization, Sensitive Encoding, Level Set Method, and the recent techniques such as Alternating Minimization, Signal Modeling, and Sphere Shaped Support Vector Machine are also reviewed. It is observed that most of these techniques enhance the gradient encoding and reduce the scanning time. Classical algorithms provide undesirable blurring effect when the degree of phase variation is high in partial k-space. Modern reconstructions algorithms such as Dictionary learning works well even with high phase variation as these are iterative procedures.
[...] Read more.The article deals with creating an algorithm for processing information in a digital system for quadrotor flight control. The minimization of L2-gain using simple parametric optimization for the synthesis of the control algorithm based on static output feedback is proposed. The kinematical diagram and mathematical description of the linearized quadrotor model are represented. The transformation of the continuous model into a discrete one has been implemented. The new optimization procedure based on digital static output feedback is developed. Expressions for the optimization criterion and penalty function are given. The features of the creating algorithm and processing information are described. The development of the closed-loop control system with an extended model augmented with some essential nonlinearities inherent to the real control plant is implemented. The simulation of the quadrotor guidance in the turbulent atmosphere has been carried out. The simulation results based on the characteristics of the studied quadrotor are represented. These results prove the efficiency of the proposed algorithm for navigation information processing. The obtained results can be useful for signal processing and designing control systems for unmanned aerial vehicles of the wide class.
[...] Read more.Nowadays, the primary concern of any society is providing safety to an individual. It is very hard to recognize the human behaviour and identify whether it is suspicious or normal. Deep learning approaches paved the way for the development of various machine learning and artificial intelligence. The proposed system detects real-time human activity using a convolutional neural network. The objective of the study is to develop a real-time application for Activity recognition using with and without transfer learning methods. The proposed system considers criminal, suspicious and normal categories of activities. Differentiate suspicious behaviour videos are collected from different peoples(men/women). This proposed system is used to detect suspicious activities of a person. The novel 2D-CNN, pre-trained VGG-16 and ResNet50 is trained on video frames of human activities such as normal and suspicious behaviour. Similarly, the transfer learning in VGG16 and ResNet50 is trained using human suspicious activity datasets. The results show that the novel 2D-CNN, VGG16, and ResNet50 without transfer learning achieve accuracy of 98.96%, 97.84%, and 99.03%, respectively. In Kaggle/real-time video, the proposed system employing 2D-CNN outperforms the pre-trained model VGG16. The trained model is used to classify the activity in the real-time captured video. The performance obtained on ResNet50 with transfer learning accuracy of 99.18% is higher than VGG16 transfer learning accuracy of 98.36%.
[...] Read more.