Treffer: Comparing hybrid models for recognising objects in thermal images at nighttime.
Weitere Informationen
This research aims to revolutionize urban object recognition by developing cloud-based Python programs using intelligent algorithms. Unlike current models that focus on colour enhancement in nighttime thermal images, this work addresses the critical challenge of accurate object detection in urban landscapes. The proposed method incorporates a binary generative adversarial network (GAN) generator that can switch bidirectionally between daytime colour (DC) and nighttime infrared (NTIR) images. memory-based visual image memory (MVAM), system extracts important descriptive information from urban landscape images, reducing problems related to small sample sizes. This discussion presents a comprehensive improvement and evaluation of a deep learning image classification pipeline using Google Colab, demonstrating advanced image processing. Using TensorFlow, Keres and scikit image libraries combined with advanced algorithms such as DenseNet121 and MobileNetV2 presents a clear approach. We created a Bidirectional GAN + MVAM for object recognition in this work. Our method performed well, with an accuracy of 81.43%, precision of 51.16, recall of 50.11, and F-score of 46.37. The systematic presentation of the code presents a careful strategy to ensure optimal performance, stability, and efficiency of deep learning and image processing tasks. [ABSTRACT FROM AUTHOR]
Copyright of Indonesian Journal of Electrical Engineering & Computer Science is the property of Institute of Advanced Engineering & Science and its content may not be copied or emailed to multiple sites without the copyright holder's express written permission. Additionally, content may not be used with any artificial intelligence tools or machine learning technologies. However, users may print, download, or email articles for individual use. This abstract may be abridged. No warranty is given about the accuracy of the copy. Users should refer to the original published version of the material for the full abstract. (Copyright applies to all Abstracts.)