Chris Chen shuang.chen@durham.ac.uk
Post Doctoral Research Associate
Chris Chen shuang.chen@durham.ac.uk
Post Doctoral Research Associate
Yifeng He
Barry Lennox
Professor Farshad Arvin farshad.arvin@durham.ac.uk
Professor
Dr Amir Atapour-Abarghouei amir.atapour-abarghouei@durham.ac.uk
Assistant Professor
Long-term monitoring and exploration of extreme environments, such as underwater storage facilities, is costly, labor-intensive, and hazardous. Automating this process with low-cost, collaborative robots can greatly improve efficiency. These robots capture images from different positions, which must be processed simultaneously to create a spatio-temporal model of the facility. In this paper, we propose a novel approach that integrates data simulation, a multi-modal deep learning network for coordinate prediction, and image reassembly to address the challenges posed by environmental disturbances causing drift and rotation in the robots' positions and ori-entations. Our approach enhances the precision of alignment in noisy environments by integrating visual information from snapshots, global positional context from masks, and noisy coordinates. We validate our method through extensive experiments using synthetic data that simulate real-world robotic operations in underwater settings. The results demonstrate very high coordinate prediction accuracy and plausible image assembly, indicating the real-world applicability of our approach. The assembled images provide clear and coherent views of the underwater environment for effective monitoring and inspection, showcasing the potential for broader use in extreme settings, further contributing to improved safety, efficiency, and cost reduction in hazardous field monitoring.
Chen, S., He, Y., Lennox, B., Arvin, F., & Atapour-Abarghouei, A. (2025, May). Deep Learning-Enhanced Visual Monitoring in Hazardous Underwater Environments with a Swarm of Micro-Robots. Presented at IEEE International Conference on Robotics & Automation, Atlanta, USA
Presentation Conference Type | Conference Paper (published) |
---|---|
Conference Name | IEEE International Conference on Robotics & Automation |
Start Date | May 19, 2025 |
End Date | May 23, 2025 |
Acceptance Date | Jan 27, 2025 |
Deposit Date | Mar 7, 2025 |
Peer Reviewed | Peer Reviewed |
Public URL | https://durham-repository.worktribe.com/output/3682342 |
Publisher URL | https://2025.ieee-icra.org/ |
This file is under embargo due to copyright reasons.
One-Index Vector Quantization Based Adversarial Attack on Image Classification
(2024)
Journal Article
Depth-Aware Endoscopic Video Inpainting
(2024)
Presentation / Conference Contribution
INCLG: Inpainting for Non-Cleft Lip Generation with a Multi-Task Image Processing Network
(2023)
Journal Article
A Feasibility Study on Image Inpainting for Non-cleft Lip Generation from Patients with Cleft Lip
(2022)
Presentation / Conference Contribution
HINT: High-quality INpainting Transformer with Mask-Aware Encoding and Enhanced Attention
(2024)
Journal Article
About Durham Research Online (DRO)
Administrator e-mail: dro.admin@durham.ac.uk
This application uses the following open-source libraries:
Apache License Version 2.0 (http://www.apache.org/licenses/)
Apache License Version 2.0 (http://www.apache.org/licenses/)
SIL OFL 1.1 (http://scripts.sil.org/OFL)
MIT License (http://opensource.org/licenses/mit-license.html)
CC BY 3.0 ( http://creativecommons.org/licenses/by/3.0/)
Powered by Worktribe © 2025
Advanced Search