Sign in to use this feature.

Years

Between: -

Subjects

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Journals

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Article Types

Countries / Regions

remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline
remove_circle_outline

Search Results (27,242)

Search Parameters:
Keywords = image detection

Order results
Result details
Results per page
Select all
Export citation of selected articles as:
25 pages, 7918 KiB  
Article
Enhancing Sustainable Automated Fruit Sorting: Hyperspectral Analysis and Machine Learning Algorithms
by Dmitry O. Khort, Alexey Kutyrev, Igor Smirnov, Nikita Andriyanov, Rostislav Filippov, Andrey Chilikin, Maxim E. Astashev, Elena A. Molkova, Ruslan M. Sarimov, Tatyana A. Matveeva and Sergey V. Gudkov
Sustainability 2024, 16(22), 10084; https://doi.org/10.3390/su162210084 - 19 Nov 2024
Abstract
Recognizing and classifying localized lesions on apple fruit surfaces during automated sorting is critical for improving product quality and increasing the sustainability of fruit production. This study is aimed at developing sustainable methods for fruit sorting by applying hyperspectral analysis and machine learning [...] Read more.
Recognizing and classifying localized lesions on apple fruit surfaces during automated sorting is critical for improving product quality and increasing the sustainability of fruit production. This study is aimed at developing sustainable methods for fruit sorting by applying hyperspectral analysis and machine learning to improve product quality and reduce losses. The employed hyperspectral technologies and machine learning algorithms enable the rapid and accurate detection of defects on the surface of fruits, enhancing product quality and reducing the number of rejects, thereby contributing to the sustainability of agriculture. This study seeks to advance commercial fruit quality control by comparing hyperspectral image classification algorithms to detect apple lesions caused by pathogens, including sunburn, scab, and rot, on three apple varieties: Honeycrisp, Gala, and Jonagold. The lesions were confirmed independently using expert judgment, real-time PCR, and 3D fluorimetry, providing a high accuracy of ground truth data and allowing conclusions to be drawn on ways to improve the sustainability and safety of the agrocenosis in which the fruits are grown. Hyperspectral imaging combined with mathematical analysis revealed that Venturia inaequalis is the main pathogen responsible for scab, while Botrytis cinerea and Penicillium expansum are the main causes of rot. This comparative study is important because it provides a detailed analysis of the performance of both supervised and unsupervised classification methods for hyperspectral imagery, which is essential for the development of reliable automated grading systems. Support Vector Machines (SVM) proved to be the most accurate, with the highest average adjusted Rand Index (ARI) scores for sunscald (0.789), scab (0.818), and rot (0.854), making it the preferred approach for classifying apple lesions during grading. K-Means performed well for scab (0.786) and rot (0.84) classes, but showed limitations with lower metrics for other lesion types. A design and technological scheme of an optical system for identifying micro- and macro-damage to fruit tissues is proposed, and the dependence of the percentage of apple damage on the rotation frequency of the sorting line rollers is obtained. The optimal values for the rotation frequency of the rollers, at which the damage to apples is less than 5%, are up to 6 Hz. The results of this study confirm the high potential of hyperspectral data for the non-invasive recognition and classification of apple diseases in automated sorting systems with an accuracy comparable to that of human experts. These results provide valuable insights into the optimization of machine learning algorithms for agricultural applications, contributing to the development of more efficient and accurate fruit quality control systems, improved production sustainability, and the long-term storage of fruits. Full article
(This article belongs to the Special Issue Agricultural Engineering for Sustainable Development)
20 pages, 687 KiB  
Review
Deep Learning-Based Atmospheric Visibility Detection
by Yawei Qu, Yuxin Fang, Shengxuan Ji, Cheng Yuan, Hao Wu, Shengbo Zhu, Haoran Qin and Fan Que
Atmosphere 2024, 15(11), 1394; https://doi.org/10.3390/atmos15111394 - 19 Nov 2024
Abstract
Atmospheric visibility is a crucial meteorological element impacting urban air pollution monitoring, public transportation, and military security. Traditional visibility detection methods, primarily manual and instrumental, have been costly and imprecise. With advancements in data science and computing, deep learning-based visibility detection technologies have [...] Read more.
Atmospheric visibility is a crucial meteorological element impacting urban air pollution monitoring, public transportation, and military security. Traditional visibility detection methods, primarily manual and instrumental, have been costly and imprecise. With advancements in data science and computing, deep learning-based visibility detection technologies have rapidly emerged as a research hotspot in atmospheric science. This paper systematically reviews the applications of various deep learning models—Convolutional Neural Networks (CNNs), Recurrent Neural Networks (RNNs), Generative Adversarial Networks (GANs), and Transformer networks—in visibility estimation, prediction, and enhancement. Each model’s characteristics and application methods are discussed, highlighting the efficiency of CNNs in spatial feature extraction, RNNs in temporal tracking, GANs in image restoration, and Transformers in capturing long-range dependencies. Furthermore, the paper addresses critical challenges in the field, including dataset quality, algorithm optimization, and practical application barriers, proposing future research directions, such as the development of large-scale, accurately labeled datasets, innovative learning strategies, and enhanced model interpretability. These findings highlight the potential of deep learning in enhancing atmospheric visibility detection techniques, providing valuable insights into the literature and contributing to advances in the field of meteorological observation and public safety. Full article
(This article belongs to the Special Issue Air Pollution Modeling and Observations in Asian Megacities)
29 pages, 30487 KiB  
Article
Joint Classification of Hyperspectral and LiDAR Data via Multiprobability Decision Fusion Method
by Tao Chen, Sizuo Chen, Luying Chen, Huayue Chen, Bochuan Zheng and Wu Deng
Remote Sens. 2024, 16(22), 4317; https://doi.org/10.3390/rs16224317 - 19 Nov 2024
Abstract
With the development of sensor technology, the sources of remotely sensed image data for the same region are becoming increasingly diverse. Unlike single-source remote sensing image data, multisource remote sensing image data can provide complementary information for the same feature, promoting its recognition. [...] Read more.
With the development of sensor technology, the sources of remotely sensed image data for the same region are becoming increasingly diverse. Unlike single-source remote sensing image data, multisource remote sensing image data can provide complementary information for the same feature, promoting its recognition. The effective utilization of remote sensing image data from various sources can enhance the extraction of image features and improve the accuracy of feature recognition. Hyperspectral remote sensing (HSI) data and light detection and ranging (LiDAR) data can provide complementary information from different perspectives and are frequently combined in feature identification tasks. However, the process of joint use suffers from data redundancy, low classification accuracy and high time complexity. To address the aforementioned issues and improve feature recognition in classification tasks, this paper introduces a multiprobability decision fusion (PRDRMF) method for the combined classification of HSI and LiDAR data. First, the original HSI data and LiDAR data are downscaled via the principal component–relative total variation (PRTV) method to remove redundant information. In the multifeature extraction module, the local texture features and spatial features of the image are extracted to consider the local texture and spatial structure of the image data. This is achieved by utilizing the local binary pattern (LBP) and extended multiattribute profile (EMAP) for the two types of data after dimensionality reduction. The four extracted features are subsequently input into the corresponding kernel–extreme learning machine (KELM), which has a simple structure and good classification performance, to obtain four classification probability matrices (CPMs). Finally, the four CPMs are fused via a multiprobability decision fusion method to obtain the optimal classification results. Comparison experiments on four classical HSI and LiDAR datasets demonstrate that the method proposed in this paper achieves high classification performance while reducing the overall time complexity of the method. Full article
Show Figures

Figure 1

20 pages, 458 KiB  
Article
Neural Architecture Search via Trainless Pruning Algorithm: A Bayesian Evaluation of a Network with Multiple Indicators
by Yiqi Lin, Yuki Endo, Jinho Lee and Shunsuke Kamijo
Electronics 2024, 13(22), 4547; https://doi.org/10.3390/electronics13224547 - 19 Nov 2024
Abstract
Neural Architecture Search (NAS) has found applications in various areas of computer vision, including image recognition and object detection. An increasing number of algorithms, such as ENAS (Efficient Neural Architecture Search via Parameter Sharing) and DARTS (Differentiable Architecture Search), have been applied to [...] Read more.
Neural Architecture Search (NAS) has found applications in various areas of computer vision, including image recognition and object detection. An increasing number of algorithms, such as ENAS (Efficient Neural Architecture Search via Parameter Sharing) and DARTS (Differentiable Architecture Search), have been applied to NAS. Nevertheless, the current Training-free NAS methods continue to exhibit unreliability and inefficiency. This paper introduces a training-free prune-based algorithm called TTNAS (True-Skill Training-Free Neural Architecture Search), which utilizes a Bayesian method (true-skill algorithm) to combine multiple indicators for evaluating neural networks across different datasets. The algorithm demonstrates highly competitive accuracy and efficiency compared to state-of-the-art approaches on various datasets. Specifically, it achieves 93.90% accuracy on CIFAR-10, 71.91% accuracy on CIFAR-100, and 44.96% accuracy on ImageNet 16-120, using 1466 GPU seconds in NAS-Bench-201. Additionally, the algorithm exhibits improved adaptation to other datasets and tasks. Full article
(This article belongs to the Special Issue Computational Imaging and Its Application)
Show Figures

Figure 1

47 pages, 3373 KiB  
Review
The Application of Nano Drug Delivery Systems in Female Upper Genital Tract Disorders
by Daniélle van Staden, Minja Gerber and Hendrik J. R. Lemmer
Pharmaceutics 2024, 16(11), 1475; https://doi.org/10.3390/pharmaceutics16111475 - 19 Nov 2024
Abstract
The prevalence of female reproductive system disorders is increasing, especially among women of reproductive age, significantly impacting their quality of life and overall health. Managing these diseases effectively is challenging due to the complex nature of the female reproductive system, characterized by dynamic [...] Read more.
The prevalence of female reproductive system disorders is increasing, especially among women of reproductive age, significantly impacting their quality of life and overall health. Managing these diseases effectively is challenging due to the complex nature of the female reproductive system, characterized by dynamic physiological environments and intricate anatomical structures. Innovative drug delivery approaches are necessary to facilitate the precise regulation and manipulation of biological tissues. Nanotechnology is increasingly considered to manage reproductive system disorders, for example, nanomaterial imaging allows for early detection and enhances diagnostic precision to determine disease severity and progression. Additionally, nano drug delivery systems are gaining attention for their ability to target the reproductive system successfully, thereby increasing therapeutic efficacy and decreasing side effects. This comprehensive review outlines the anatomy of the female upper genital tract by highlighting the complex mucosal barriers and their impact on systemic and local drug delivery. Advances in nano drug delivery are described for their sustainable therapeutic action and increased biocompatibility to highlight the potential of nano drug delivery strategies in managing female upper genital tract disorders. Full article
(This article belongs to the Special Issue Drug Delivery in the Reproductive Systems)
Show Figures

Figure 1

21 pages, 12271 KiB  
Article
Detection of Marine Oil Spill from PlanetScope Images Using CNN and Transformer Models
by Jonggu Kang, Chansu Yang, Jonghyuk Yi and Yangwon Lee
J. Mar. Sci. Eng. 2024, 12(11), 2095; https://doi.org/10.3390/jmse12112095 - 19 Nov 2024
Abstract
The contamination of marine ecosystems by oil spills poses a significant threat to the marine environment, necessitating the prompt and effective implementation of measures to mitigate the associated damage. Satellites offer a spatial and temporal advantage over aircraft and unmanned aerial vehicles (UAVs) [...] Read more.
The contamination of marine ecosystems by oil spills poses a significant threat to the marine environment, necessitating the prompt and effective implementation of measures to mitigate the associated damage. Satellites offer a spatial and temporal advantage over aircraft and unmanned aerial vehicles (UAVs) in oil spill detection due to their wide-area monitoring capabilities. While oil spill detection has traditionally relied on synthetic aperture radar (SAR) images, the combined use of optical satellite sensors alongside SAR can significantly enhance monitoring capabilities, providing improved spatial and temporal coverage. The advent of deep learning methodologies, particularly convolutional neural networks (CNNs) and Transformer models, has generated considerable interest in their potential for oil spill detection. In this study, we conducted a comprehensive and objective comparison to evaluate the suitability of CNN and Transformer models for marine oil spill detection. High-resolution optical satellite images were used to optimize DeepLabV3+, a widely utilized CNN model; Swin-UPerNet, a representative Transformer model; and Mask2Former, which employs a Transformer-based architecture for both encoding and decoding. The results of cross-validation demonstrate a mean Intersection over Union (mIoU) of 0.740, 0.840 and 0.804 for all the models, respectively, indicating their potential for detecting oil spills in the ocean. Additionally, we performed a histogram analysis on the predicted oil spill pixels, which allowed us to classify the types of oil. These findings highlight the considerable promise of the Swin Transformer models for oil spill detection in the context of future marine disaster monitoring. Full article
(This article belongs to the Special Issue Remote Sensing Applications in Marine Environmental Monitoring)
Show Figures

Figure 1

11 pages, 50395 KiB  
Article
Detection of Low-Density Foreign Objects in Infant Snacks Using a Continuous-Wave Sub-Terahertz Imaging System for Industrial Applications
by Byeong-Hyeon Na, Dae-Ho Lee, Jaein Choe, Young-Duk Kim and Mi-Kyung Park
Sensors 2024, 24(22), 7374; https://doi.org/10.3390/s24227374 - 19 Nov 2024
Abstract
Low-density foreign objects (LDFOs) in foods pose significant safety risks to consumers. Existing detection methods, such as metal and X-ray detectors, have limitations in identifying low-density and nonmetallic contaminants. To address these challenges, our research group constructed and optimized a continuous-wave sub-terahertz (THz) [...] Read more.
Low-density foreign objects (LDFOs) in foods pose significant safety risks to consumers. Existing detection methods, such as metal and X-ray detectors, have limitations in identifying low-density and nonmetallic contaminants. To address these challenges, our research group constructed and optimized a continuous-wave sub-terahertz (THz) imaging system for the real-time, on-site detection of LDFOs in infant snacks. The system was optimized by adjusting the attenuation value from 0 to 9 dB and image processing parameters [White (W), Black (B), and Gamma (G)] from 0 to 100. Its detectability was evaluated across eight LDFOs underneath snacks with scanning at 30 cm/s. The optimal settings for puffed snacks and freeze-dried chips were found to be 3 dB attenuation with W, B, and G values of 100, 50, and 80, respectively, while others required 0 dB attenuation with W, B, and G set to 100, 0, and 100, respectively. Additionally, the moisture content of infant snacks was measured using a modified AOAC-based drying method at 105 °C, ensuring the removal of all free moisture. Using these optimized settings, the system successfully detected a housefly and a cockroach underneath puffed snacks and freeze-dried chips. It also detected LDFOs as small as 3 mm in size in a single layer of snacks, including polyurethane, polyvinyl chloride, ethylene–propylene–diene–monomer, and silicone, while in two layers of infant snacks, they were detected up to 7.5 mm. The constructed system can rapidly and effectively detect LDFOs in foods, offering a promising approach to enhance safety in the food industry. Full article
(This article belongs to the Section Sensing and Imaging)
Show Figures

Figure 1

20 pages, 15268 KiB  
Article
Automatic Reading and Reporting Weather Information from Surface Fax Charts for Ships Sailing in Actual Northern Pacific and Atlantic Oceans
by Jun Jian, Yingxiang Zhang, Ke Xu and Peter J. Webster
J. Mar. Sci. Eng. 2024, 12(11), 2096; https://doi.org/10.3390/jmse12112096 - 19 Nov 2024
Abstract
This study is aimed to improve the intelligence level, efficiency, and accuracy of ship safety and security systems by contributing to the development of marine weather forecasting. The accurate and prompt recognition of weather fax charts is very important for navigation safety. This [...] Read more.
This study is aimed to improve the intelligence level, efficiency, and accuracy of ship safety and security systems by contributing to the development of marine weather forecasting. The accurate and prompt recognition of weather fax charts is very important for navigation safety. This study employed many artificial intelligent (AI) methods including a vectorization approach and target recognition algorithm to automatically detect the severe weather information from Japanese and US weather charts. This enabled the expansion of an existing auto-response marine forecasting system’s applications toward north Pacific and Atlantic Oceans, thus enhancing decision-making capabilities and response measures for sailing ships at actual sea. The OpenCV image processing method and YOLOv5s/YOLO8vn algorithm were utilized to make template matches and locate warning symbols and weather reports from surface weather charts. After these improvements, the average accuracy of the model significantly increased from 0.920 to 0.928, and the detection rate of a single image reached a maximum of 1.2 ms. Additionally, OCR technology was applied to retract texts from weather reports and highlighted the marine areas where dense fog and great wind conditions are likely to occur. Finally, the field tests confirmed that this auto and intelligent system could assist the navigator within 2–3 min and thus greatly enhance the navigation safety in specific areas in the sailing routes with minor text-based communication costs. Full article
(This article belongs to the Special Issue Ship Performance in Actual Seas)
Show Figures

Figure 1

17 pages, 8145 KiB  
Article
Integrated Anti-Aliasing and Fully Shared Convolution for Small-Ship Detection in Synthetic Aperture Radar (SAR) Images
by Manman He, Junya Liu, Zhen Yang and Zhijian Yin
Electronics 2024, 13(22), 4540; https://doi.org/10.3390/electronics13224540 - 19 Nov 2024
Abstract
Synthetic Aperture Radar (SAR) imaging plays a vital role in maritime surveillance, yet the detection of small vessels poses a significant challenge when employing conventional Constant False Alarm Rate (CFAR) techniques, primarily due to the limitations in resolution and the presence of clutter. [...] Read more.
Synthetic Aperture Radar (SAR) imaging plays a vital role in maritime surveillance, yet the detection of small vessels poses a significant challenge when employing conventional Constant False Alarm Rate (CFAR) techniques, primarily due to the limitations in resolution and the presence of clutter. Deep learning (DL) offers a promising alternative, yet it still struggles with identifying small targets in complex SAR backgrounds because of feature ambiguity and noise. To address these challenges, our team has developed the AFSC network, which combines anti-aliasing techniques with fully shared convolutional layers to improve the detection of small targets in SAR imagery. The network is composed of three key components: the Backbone Feature Extraction Module (BFEM) for initial feature extraction, the Neck Feature Fusion Module (NFFM) for consolidating features, and the Head Detection Module (HDM) for final object detection. The BFEM serves as the principal feature extraction technique, with a primary emphasis on extracting features of small targets, The NFFM integrates an anti-aliasing element and is designed to accentuate the feature details of diminutive objects throughout the fusion procedure, HDM is the detection head module and adopts a new fully shared convolution strategy to make the model more lightweight. Our approach has shown better performance in terms of speed and accuracy for detecting small targets in SAR imagery, surpassing other leading methods on the SSDD dataset. It attained a mean Average Precision (AP) of 69.3% and a specific AP for small targets (APS) of 66.5%. Furthermore, the network’s robustness was confirmed using the HRSID dataset. Full article
(This article belongs to the Special Issue Advances in AI Technology for Remote Sensing Image Processing)
Show Figures

Figure 1

22 pages, 1952 KiB  
Article
A Fully Autonomous On-Board GNC Methodology for Small-Body Environments Based on CNN Image Processing and MPCs
by Pelayo Peñarroya, Alfredo Escalante, Thomas Frekhaug and Manuel Sanjurjo
Aerospace 2024, 11(11), 952; https://doi.org/10.3390/aerospace11110952 (registering DOI) - 19 Nov 2024
Abstract
The increasing need for autonomy in space exploration missions is becoming more and more relevant in the design of missions to small bodies. The long communication latencies and sensitivity of the system to unplanned environmental perturbations mean autonomous methods could be a key [...] Read more.
The increasing need for autonomy in space exploration missions is becoming more and more relevant in the design of missions to small bodies. The long communication latencies and sensitivity of the system to unplanned environmental perturbations mean autonomous methods could be a key design block for this type of mission. In this work, a fully autonomous Guidance, Navigation, and Control (GNC) methodology is introduced. This methodology relies on published CNN-based techniques for surface recognition and pose estimation and also on existing MPC-based techniques for the design of a trajectory to perform a soft landing on an asteroid. Combining Hazard Detection and Avoidance (HDA) with relative navigation systems, a Global Safety Map (GSM) is built on the fly as images are acquired. These GSMs provide the GNC system with information about feasible landing spots and populate a longitude–latitude map with safe/hazardous labels that are later processed to find an optimal landing spot based on mission requirements and a distance-fromhazard metric. The methodology is exemplified using Bennu as the body of interest, and a GSM is built for an arbitrary reconnaissance orbit. Full article
(This article belongs to the Section Astronautics & Space Science)
Show Figures

Figure 1

17 pages, 5786 KiB  
Article
Corn Plant In-Row Distance Analysis Based on Unmanned Aerial Vehicle Imagery and Row-Unit Dynamics
by Marko M. Kostić, Željana Grbović, Rana Waqar, Bojana Ivošević, Marko Panić, Antonio Scarfone and Aristotelis C. Tagarakis
Appl. Sci. 2024, 14(22), 10693; https://doi.org/10.3390/app142210693 - 19 Nov 2024
Abstract
Uniform spatial distribution of plants is crucial in arable crops. Seeding quality is affected by numerous parameters, including the working speed and vibrations of the seeder. Therefore, investigating effective and rapid methods to evaluate seeding quality and the parameters affecting the seeders’ performance [...] Read more.
Uniform spatial distribution of plants is crucial in arable crops. Seeding quality is affected by numerous parameters, including the working speed and vibrations of the seeder. Therefore, investigating effective and rapid methods to evaluate seeding quality and the parameters affecting the seeders’ performance is of high importance. With the latest advancements in unmanned aerial vehicle (UAV) technology, the potential for acquiring accurate agricultural data has significantly increased, making UAVs an ideal tool for scouting applications in agricultural systems. This study investigates the effectiveness of utilizing different plant recognition algorithms applied to UAV-derived images for evaluating seeder performance based on detected plant spacings. Additionally, it examines the impact of seeding unit vibrations on seeding quality by analyzing accelerometer data installed on the seeder. For the image analysis, three plant recognition approaches were tested: an unsupervised segmentation method based on the Visible Atmospherically Resistant Index (VARI), template matching (TM), and a deep learning model called Mask R-CNN. The Mask R-CNN model demonstrated the highest recognition reliability at 96.7%, excelling in detecting seeding errors such as misses and doubles, as well as in evaluating the quality of feed index and precision when compared to ground-truth data. Although the VARI-based unsupervised method and TM outperformed Mask R-CNN in recognizing double spacings, overall, the Mask R-CNN was the most promising. Vibration analysis indicated that the seeder’s working speed significantly affected seeding quality. These findings suggest areas for potential improvements in machine technology to improve sowing operations. Full article
Show Figures

Figure 1

22 pages, 46477 KiB  
Article
Pixelator v2: A Novel Perceptual Image Comparison Method with LAB Colour Space and Sobel Edge Detection for Enhanced Security Analysis
by Somdip Dey, Jabir Alshehabi Al-Ani, Aikaterini Bourazeri, Suman Saha, Rohit Purkait, Samuel Hill and Julian Thompson
Electronics 2024, 13(22), 4541; https://doi.org/10.3390/electronics13224541 - 19 Nov 2024
Abstract
In this paper, we introduce Pixelator v2, a novel perceptual image comparison method designed to enhance security and analysis through improved image difference detection. Unlike traditional metrics such as MSE, Q, and SSIM, which often fail to capture subtle but critical changes in [...] Read more.
In this paper, we introduce Pixelator v2, a novel perceptual image comparison method designed to enhance security and analysis through improved image difference detection. Unlike traditional metrics such as MSE, Q, and SSIM, which often fail to capture subtle but critical changes in images, Pixelator v2 integrates the LAB (CIE-LAB) colour space for perceptual relevance and Sobel edge detection for structural integrity. By combining these techniques, Pixelator v2 offers a more robust and nuanced approach to identifying variations in images, even in cases of minor modifications. The LAB colour space ensures that the method aligns with human visual perception, making it particularly effective at detecting differences that are less visible in RGB space. Sobel edge detection, on the other hand, emphasises structural changes, allowing Pixelator v2 to focus on the most significant areas of an image. This combination makes Pixelator v2 ideal for applications in security, where image comparison plays a vital role in tasks like tamper detection, authentication, and analysis. We evaluate Pixelator v2 against other popular methods, demonstrating its superior performance in detecting both perceptual and structural differences. Our results indicate that Pixelator v2 not only provides more accurate image comparisons but also enhances security by making it more difficult for subtle alterations to go unnoticed. This paper contributes to the growing field of image-based security systems by offering a perceptually-driven, computationally efficient method for image comparison that can be readily applied in information system security. Full article
Show Figures

Figure 1

13 pages, 3184 KiB  
Review
A Comprehensive Review of Performance Metrics for Computer-Aided Detection Systems
by Doohyun Park
Bioengineering 2024, 11(11), 1165; https://doi.org/10.3390/bioengineering11111165 - 19 Nov 2024
Abstract
This paper aims to provide a structured analysis of the performance metrics used in computer-aided detection (CAD) systems, specifically focusing on lung nodule detection in computed tomography (CT) images. By examining key metrics along with their respective strengths and limitations, this study offers [...] Read more.
This paper aims to provide a structured analysis of the performance metrics used in computer-aided detection (CAD) systems, specifically focusing on lung nodule detection in computed tomography (CT) images. By examining key metrics along with their respective strengths and limitations, this study offers guidelines to assist in selecting appropriate metrics. Evaluation methods for CAD systems for lung nodule detection are primarily categorized into per-scan and per-nodule approaches. For per-scan analysis, a key metric is the area under the receiver operating characteristic (ROC) curve (AUROC), which evaluates the ability of the system to distinguish between scans with and without nodules. For per-nodule analysis, the nodule-level sensitivity at fixed false positives per scan is often used, supplemented by the free-response receiver operating characteristic (FROC) curve and the competition performance metric (CPM). However, the CPM does not provide normalized scores because it theoretically ranges from zero to infinity and largely varies depending on the characteristics of the data. To address the advantages and limitations of ROC and FROC curves, an alternative FROC (AFROC) was introduced to combine the strengths of both per-scan and per-nodule analyses. This paper discusses the principles of each metric and their relative strengths, providing insights into their clinical implications and practical utility. Full article
Show Figures

Figure 1

23 pages, 9369 KiB  
Article
A YOLO-Based Method for Head Detection in Complex Scenes
by Ming Xie, Xiaobing Yang, Boxu Li and Yingjie Fan
Sensors 2024, 24(22), 7367; https://doi.org/10.3390/s24227367 - 19 Nov 2024
Abstract
Detecting objects in intricate scenes has always presented a significant challenge in the field of machine vision. Complex scenes typically refer to situations in images or videos where there are numerous densely distributed and mutually occluded objects, making the object detection task even [...] Read more.
Detecting objects in intricate scenes has always presented a significant challenge in the field of machine vision. Complex scenes typically refer to situations in images or videos where there are numerous densely distributed and mutually occluded objects, making the object detection task even more difficult. This paper introduces a novel head detection algorithm, YOLO-Based Head Detection in Complex Scenes (YOLO-HDCS). Firstly, in complex scenes, head detection typically involves a large number of small objects that are randomly distributed. Traditional object detection algorithms struggle to address the challenge of small object detection. For this purpose, two new modules have been constructed: one is a feature fusion module based on context enhancement with scale adjustment, and the other is an attention-based convolutional module. These modules are characterized by high detection efficiency and high accuracy. They significantly improve the model’s multi-scale detection capabilities, thus enhancing the detection ability of the system. Secondly, it was found in practical operations that the original IoU function has a serious problem with overlapping detection in complex scenes. There is an IoU function that can ensure that the final selection boxes cover the object as accurately as possible without overlapping. This not only improves the detection performance but also greatly aids in enhancing the detection efficiency and accuracy. Our method achieves impressive results for head detection in complex scenarios, with average accuracy of 82.2%, and has the advantage of rapid loss convergence during training. Full article
(This article belongs to the Special Issue Intelligent Sensors and Artificial Intelligence in Building)
Show Figures

Figure 1

15 pages, 10336 KiB  
Technical Note
Multi-Scenario Remote Sensing Image Forgery Detection Based on Transformer and Model Fusion
by Jinmiao Zhao, Zelin Shi, Chuang Yu and Yunpeng Liu
Remote Sens. 2024, 16(22), 4311; https://doi.org/10.3390/rs16224311 - 19 Nov 2024
Abstract
Recently, remote sensing image forgery detection has received widespread attention. To improve the detection accuracy, we build a novel scheme based on Transformer and model fusion. Specifically, we model this task as a binary classification task that focuses on global information. First, we [...] Read more.
Recently, remote sensing image forgery detection has received widespread attention. To improve the detection accuracy, we build a novel scheme based on Transformer and model fusion. Specifically, we model this task as a binary classification task that focuses on global information. First, we explore the performance of various excellent feature extraction networks in this task under the constructed unified classification framework. On this basis, we select three high-performance Transformer-based networks that focus on global information, namely, Swin Transformer V1, Swin Transformer V2, and Twins, as the backbone networks and fuse them. Secondly, considering the small number of samples, we use the public ImageNet-1K dataset to pre-train the network to learn more stable feature expressions. At the same time, a circular data divide strategy is proposed, which can fully utilize all the samples to improve the accuracy in the competition. Finally, to promote network optimization, on the one hand, we explore multiple loss functions and select label smooth loss, which can reduce the model’s excessive dependence on training data. On the other hand, we construct a combined learning rate optimization strategy that first uses step degeneration and then cosine annealing, which reduces the risk of the network falling into local optima. Extensive experiments show that the proposed scheme has excellent performance. This scheme won seventh place in the “Forgery Detection in Multi-scenario Remote Sensing Images of Typical Objects” track of the 2024 ISPRS TC I contest on Intelligent Interpretation for Multi-modal Remote Sensing Application. Full article
(This article belongs to the Special Issue Geospatial Artificial Intelligence (GeoAI) in Remote Sensing)
Show Figures

Figure 1

Back to TopTop