Efficient Image Super-Resolution Guided by Object Detection
DOI:
https://doi.org/10.6919/ICJE.202506_11(6).0005Keywords:
Neural Network; Machine Vision; Tower Construction.Abstract
To address the challenges of reduced target recognition accuracy and insufficient real-time performance caused by low image resolution in traditional tower assembly construction, this paper proposes an efficient image super-resolution method guided by object detection. The proposed approach employs YOLOv8 to rapidly locate target regions and introduces a discriminative image segmentation module to divide the image into key and background regions. These regions are then processed separately using a specially designed Super-Resolution Guided Efficient Detail Recovery Network (SGERN) and a Background Lightweight Super-Resolution Network (BLSRN), thereby enhancing image quality while significantly reducing computational cost. The SGERN integrates Generative Adversarial Networks (GAN), a simple gating mechanism, and re-parameterization modules to achieve high-quality reconstruction of fine textures in key regions. Experiments conducted on a self-built dataset of hoisting components and public datasets (UDM10 and Urban100) demonstrate that the proposed method outperforms mainstream approaches in terms of PSNR and SSIM. Moreover, it achieves good real-time performance while ensuring high visual quality in the target areas. This study provides a novel and practical solution for target recognition and image enhancement in complex construction scenarios, with promising engineering application value.
Downloads
References
[1] Wang G, Ding P, Huang C, et al. A novel lifting point location optimization method of transmission line tower based on improved grey wolf optimizer[J]. Scientific Reports, 2023, 13(1): 21914.
[2] Zhou L, Zhang L, Konz N. Computer vision techniques in manufacturing[J]. IEEE Transactions on Systems, Man, and Cybernetics: Systems, 2022, 53(1): 105-117.
[3] Varghese R, Sambath M. Yolov8: A novel object detection algorithm with enhanced performance and robustness[C]//2024 International Conference on Advances in Data Engineering and Intelligent Computing Systems (ADICS). IEEE, 2024: 1-6.
[4] Ledig C, Theis L, Huszár F, et al. Photo-realistic single image super-resolution using a generative adversarial network[C]. Proceedings of the IEEE conference on computer vision and pattern recognition. 2017: 4681-4690.
[5] Chan E R, Lin C Z, Chan M A, et al. Efficient geometry-aware 3d generative adversarial networks[C]//Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 2022: 16123-16133.
[6] Navidan H, Moshiri P F, Nabati M, et al. Generative Adversarial Networks (GANs) in networking: A comprehensive survey & evaluation[J]. Computer Networks, 2021, 194: 108149.
[7] De Souza V L T, Marques B A D, Batagelo H C, et al. A review on generative adversarial networks for image generation[J]. Computers & Graphics, 2023, 114: 13-25.
[8] Wang H, Chen X, Ni B, et al. Omni aggregation networks for lightweight image super-resolution[C]//Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition. 2023: 22378-22387.
[9] Chen L, Chu X, Zhang X, et al. Simple baselines for image restoration[C]. European conference on computer vision. Cham: Springer Nature Switzerland, 2022: 17-33.
[10] Ding X, Zhang X, Ma N, et al. Repvgg: Making vgg-style convnets great again[C]//Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. 2021: 13733-13742.
[11] Sun L, Pan J, Tang J. Shufflemixer: An efficient convnet for image super-resolution[J]. Advances in Neural Information Processing Systems, 2022, 35: 17314-17326.
[12] Feng X, Li X, Li J. Multi-scale fractal residual network for image super-resolution[J]. Applied Intelligence, 2021, 51(4): 1845-1856.
[13] Meng G, Wu Y, Li S, et al. Video Super-Resolution with Long-Term Self-Exemplars[J]. arXiv preprint arXiv:2106.12778, 2021.
Downloads
Published
Issue
Section
License
Copyright (c) 2025 International Core Journal of Engineering

This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License.




