-
Notifications
You must be signed in to change notification settings - Fork 4
Commit
- Loading branch information
There are no files selected for viewing
Some generated files are not rendered by default. Learn more about how customized files appear on GitHub.
Some generated files are not rendered by default. Learn more about how customized files appear on GitHub.
Some generated files are not rendered by default. Learn more about how customized files appear on GitHub.
Some generated files are not rendered by default. Learn more about how customized files appear on GitHub.
Some generated files are not rendered by default. Learn more about how customized files appear on GitHub.
Some generated files are not rendered by default. Learn more about how customized files appear on GitHub.
Some generated files are not rendered by default. Learn more about how customized files appear on GitHub.
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -1,2 +1,72 @@ | ||
# IDT | ||
|
||
# [Image Deraining Transformer](https://ieeexplore.ieee.org/document/9798773) (IDT) | ||
<hr /> | ||
|
||
> **Abstract:** *Existing deep learning based de-raining approaches have resorted to the convolutional architectures. However, the intrinsic limitations of convolution, including local receptive fields and independence of input content, hinder the model's ability to capture long-range and complicated rainy artifacts. To overcome these limitations, we propose an effective and efficient transformer-based architecture for the image de-raining. Firstly, we introduce general priors of vision tasks, i.e., locality and hierarchy, into the network architecture so that our model can achieve excellent de-raining performance without costly pre-training. Secondly, since the geometric appearance of rainy artifacts is complicated and of significant variance in space, it is essential for de-raining models to extract both local and non-local features. Therefore, we design the complementary window-based transformer and spatial transformer to enhance locality while capturing long-range dependencies. Besides, to compensate for the positional blindness of self-attention, we establish a separate representative space for modeling positional relationship, and design a new relative position enhanced multi-head self-attention. In this way, our model enjoys powerful abilities to capture dependencies from both content and position, so as to achieve better image content recovery while removing rainy artifacts. Experiments substantiate that our approach attains more appealing results than state-of-the-art methods quantitatively and qualitatively.* | ||
<hr /> | ||
|
||
## Method | ||
|
||
 | ||
|
||
 | ||
|
||
## Results | ||
For research convinience, we release both derained patches as well as full-size images. | ||
- Derain results [Google Drive](https://drive.google.com/drive/folders/17GkFCALmG50RNrc0p4Cl92rAwlVagtWs?usp=sharing) | ||
|
||
## Evaluation | ||
|
||
- Rain200H/Rain200L/DDN/DID/SPA: <a href="evaluaterain.m">evaluaterain.m</a> | ||
- Raindrop: <a href="evaluate_raindrop.py">evaluate_raindrop.py</a> | ||
|
||
## Full-size Evaluation Results | ||
<table> | ||
<tr> | ||
<th align="left">Dataset</th> | ||
<th align="center">Rain200H</th> | ||
<th align="center">Rain200L</th> | ||
<th align="center">DDN-Data</th> | ||
<th align="center">DID-Data</th> | ||
<th align="center">SPA-Data</th> | ||
<th align="center">AGAN-Data</th> | ||
</tr> | ||
<tr> | ||
<td align="left">PSNR</td> | ||
<td align="center">32.10</td> | ||
<td align="center">40.74</td> | ||
<td align="center">33.80</td> | ||
<td align="center">34.85</td> | ||
<td align="center">47.34</td> | ||
<th align="center">31.63</th> | ||
</tr> | ||
<tr> | ||
<td align="left">SSIM</td> | ||
<td align="center">0.9343</td> | ||
<td align="center">0.9884</td> | ||
<td align="center">0.9407</td> | ||
<td align="center">0.9401</td> | ||
<td align="center">0.9929</td> | ||
<th align="center">0.9360</th> | ||
</tr> | ||
</table> | ||
|
||
## Pretrained Model | ||
- Derain [Google Drive](https://drive.google.com/drive/folders/1PrEwOyFPciWKHC_UFe5ZE7x9JqfUkUMP?usp=sharing) | ||
|
||
## Citation | ||
If you use IDT, please consider citing: | ||
|
||
@article{xiao2022image, | ||
title={Image De-raining Transformer}, | ||
author={Xiao, Jie and Fu, Xueyang and Liu, Aiping and Wu, Feng and Zha, Zheng-Jun}, | ||
journal={IEEE Transactions on Pattern Analysis and Machine Intelligence}, | ||
year={2022}, | ||
publisher={IEEE} | ||
} | ||
|
||
## Acknowledgement | ||
This code is based on the [Uformer](https://github.com/ZhendongWang6/Uformer). The evaluation code borrows from [Restormer](https://github.com/swz30/Restormer) and [AttentGAN](https://github.com/rui1996/DeRaindrop). | ||
|
||
## Contact | ||
Please contact us if there is any question([email protected]). | ||
|