1st Place Solutions for UG2+ Challenge 2022 ATMOSPHERIC TURBULENCE
MITIGATION
- URL: http://arxiv.org/abs/2210.16847v1
- Date: Sun, 30 Oct 2022 14:11:36 GMT
- Title: 1st Place Solutions for UG2+ Challenge 2022 ATMOSPHERIC TURBULENCE
MITIGATION
- Authors: Zhuang Liu, Zhichao Zhao, Ye Yuan, Zhi Qiao, Jinfeng Bai and Zhilong
Ji
- Abstract summary: We propose a unified end-to-end framework to reconstruct a high quality image from distorted frames.
Our framework is efficient and generic, which is adapted to both hot-air image and text pattern.
We achieve the average accuracy of 98.53% on the reconstruction result of the text patterns, ranking 1st on the final leaderboard.
- Score: 11.380487356442863
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: In this technical report, we briefly introduce the solution of our team
''summer'' for Atomospheric Turbulence Mitigation in UG$^2$+ Challenge in CVPR
2022. In this task, we propose a unified end-to-end framework to reconstruct a
high quality image from distorted frames, which is mainly consists of a
Restormer-based image reconstruction module and a NIMA-based image quality
assessment module. Our framework is efficient and generic, which is adapted to
both hot-air image and text pattern. Moreover, we elaborately synthesize more
than 10 thousands of images to simulate atmospheric turbulence. And these
images improve the robustness of the model. Finally, we achieve the average
accuracy of 98.53\% on the reconstruction result of the text patterns, ranking
1st on the final leaderboard.
Related papers
- Emu: Enhancing Image Generation Models Using Photogenic Needles in a
Haystack [75.00066365801993]
Training text-to-image models with web scale image-text pairs enables the generation of a wide range of visual concepts from text.
These pre-trained models often face challenges when it comes to generating highly aesthetic images.
We propose quality-tuning to guide a pre-trained model to exclusively generate highly visually appealing images.
arXiv Detail & Related papers (2023-09-27T17:30:19Z) - Physics-Driven Turbulence Image Restoration with Stochastic Refinement [80.79900297089176]
Image distortion by atmospheric turbulence is a critical problem in long-range optical imaging systems.
Fast and physics-grounded simulation tools have been introduced to help the deep-learning models adapt to real-world turbulence conditions.
This paper proposes the Physics-integrated Restoration Network (PiRN) to help the network to disentangle theity from the degradation and the underlying image.
arXiv Detail & Related papers (2023-07-20T05:49:21Z) - 1st Solution Places for CVPR 2023 UG$^2$+ Challenge Track 2.2-Coded
Target Restoration through Atmospheric Turbulence [12.484269899245515]
This report introduces the solution of our team VIELab-HUST for coded target restoration through atmospheric turbulence in CVPR 2023 UG$2$+ Track 2.2.
We propose an efficient multi-stage framework to restore a high quality image from distorted frames.
Our framework is capable of handling different kinds of coded target dataset provided in the final testing phase, and ranked 1st on the final leaderboard.
arXiv Detail & Related papers (2023-06-15T09:06:48Z) - 1st Solution Places for CVPR 2023 UG$^{\textbf{2}}$+ Challenge Track
2.1-Text Recognition through Atmospheric Turbulence [13.393698451466689]
We present the solution developed by our team VIELab-HUST for text recognition through atmospheric turbulence in Track 2.1 of the CVPR 2023 UG$2$+ challenge.
Our framework can handle both hot-air text dataset and turbulence text dataset provided in the final testing phase and achieved 1st place in text recognition accuracy.
arXiv Detail & Related papers (2023-06-15T08:56:51Z) - A Two-Stage Real Image Deraining Method for GT-RAIN Challenge CVPR 2023
Workshop UG$^{\textbf{2}}$+ Track 3 [15.370704973282848]
We propose an efficient two-stage framework to reconstruct a clear image from rainy frames.
A transformer-based single image deraining network Uformer is implemented to pre-train on large real rain dataset.
Our overall framework is elaborately designed and able to handle both heavy rainy and foggy sequences.
arXiv Detail & Related papers (2023-05-13T18:30:27Z) - Dual-former: Hybrid Self-attention Transformer for Efficient Image
Restoration [6.611849560359801]
We present Dual-former, which combines the powerful global modeling ability of self-attention modules and the local modeling ability of convolutions in an overall architecture.
Experiments demonstrate that Dual-former achieves a 1.91dB gain over the state-of-the-art MAXIM method on the Indoor dataset for single image dehazing.
For single image deraining, it exceeds the SOTA method by 0.1dB PSNR on the average results of five datasets with only 21.5% GFLOPs.
arXiv Detail & Related papers (2022-10-03T16:39:21Z) - 2nd Place Solutions for UG2+ Challenge 2022 -- D$^{3}$Net for Mitigating
Atmospheric Turbulence from Images [5.092028049119383]
D$3$Net was proposed by our team "TUK-IK" for Atmospheric Turbulence Mitigation in $UG2+$ Challenge at CVPR 2022.
The proposed method ranked 2nd on the final leader-board of the aforementioned challenge in the testing phase, respectively.
arXiv Detail & Related papers (2022-08-25T20:20:09Z) - Single Frame Atmospheric Turbulence Mitigation: A Benchmark Study and A
New Physics-Inspired Transformer Model [82.23276183684001]
We propose a physics-inspired transformer model for imaging through atmospheric turbulence.
The proposed network utilizes the power of transformer blocks to jointly extract a dynamical turbulence distortion map.
We present two new real-world turbulence datasets that allow for evaluation with both classical objective metrics and a new task-driven metric using text recognition accuracy.
arXiv Detail & Related papers (2022-07-20T17:09:16Z) - Improved Transformer for High-Resolution GANs [69.42469272015481]
We introduce two key ingredients to Transformer to address this challenge.
We show in the experiments that the proposed HiT achieves state-of-the-art FID scores of 31.87 and 2.95 on unconditional ImageNet $128 times 128$ and FFHQ $256 times 256$, respectively.
arXiv Detail & Related papers (2021-06-14T17:39:49Z) - Aggregated Contextual Transformations for High-Resolution Image
Inpainting [57.241749273816374]
We propose an enhanced GAN-based model, named Aggregated COntextual-Transformation GAN (AOT-GAN) for high-resolution image inpainting.
To enhance context reasoning, we construct the generator of AOT-GAN by stacking multiple layers of a proposed AOT block.
For improving texture synthesis, we enhance the discriminator of AOT-GAN by training it with a tailored mask-prediction task.
arXiv Detail & Related papers (2021-04-03T15:50:17Z) - Gated Fusion Network for Degraded Image Super Resolution [78.67168802945069]
We propose a dual-branch convolutional neural network to extract base features and recovered features separately.
By decomposing the feature extraction step into two task-independent streams, the dual-branch model can facilitate the training process.
arXiv Detail & Related papers (2020-03-02T13:28:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.