This repo is the official implementation of the DRENet in "A Degraded Reconstruction Enhancement-based Method for Tiny Ship Detection in Remote Sensing Images with A New Large-scale Dataset". The paper can be accessed in [IEEE | Lab Server | ResearchGate]. (Accepted by TGRS 2022)
If you encounter any question, please feel free to contact us. You can create an issue or just send email to me windvchen@gmail.com. Also welcome for any idea exchange and discussion.
05/30/2023
Thanks for the discussions in #10. The current code now supports processing different resolutions without manually modifying the yaml configurations. To achieve this, you can comment out L139-141 in common.py. (Note that this feature has not well been tested, if there is a drop in performance, you can get back to the manual modification mentioned in #4).
11/19/2022
We have improved the previous codes, and the current repository support directly detect the images without generating degraded images, which correspond to the file detect.py.
06/10/2022
The code cleanup is finished and the complete codes are provided, also the weights of our model on LEVIR-Ship dataset.
06/06/2022
We will finish the code cleanup within a week, and make both the code and dataset fully public. Please be patient.
We focus on tiny ship detection task in medium-resolution (MR, about 16m/pixel) remote sensing (RS) images . Compared with high-resolution (HR) RS image, an MR image covers a much wider area, thus facilitating quick ship detection. This direction has great research significance, and can greatly benefit the rapid ship detection under massive RS images.
For the task, we propose an effective Degraded Reconstruction Enhancement Network (DRENet), where a degraded reconstruction enhancer is designed to learn to regress an object-aware blurred version of the input image. Our method achieves both great effectiveness and efficiency, and outperforms many recent methods.
Methods | Params(M) | FLOPs(G) | AP | FPS |
---|---|---|---|---|
YOLOv3 | 61.52 | 99.2 | 69.9 | 61 |
YOLOv5s | 7.05 | 10.4 | 75.6 [Google Drive | Baidu Pan (code:ogdm)] |
95 |
Retinanet | 36.33 | 104.4 | 74.9 | 12 |
SSD | 24.39 | 175.2 | 52.6 | 25 |
FasterRCNN | 136.70 | 299.2 | 70.8 | 10 |
EfficientDet-D0 | 3.84 | 4.6 | 71.3 | 32 |
EfficientDet-D2 | 8.01 | 20.0 | 80.9 | 21 |
FCOS | 5.92 | 51.8 | 75.5 | 37 |
CenterNet | 191.24 | 584.6 | 77.7 | 25 |
HSFNet | 157.59 | 538.1 | 73.6 | 7 |
ImYOLOv3 | 62.86 | 101.9 | 72.6 | 51 |
MaskRCNN+DFR+RFE | 24.99 | 237.8 | 76.2 | 6 |
DRENet | 4.79 | 8.3 | 82.4 [Google Drive | Baidu Pan (code:x710)] |
85 |
Please at first download dataset LEVIR-Ship, then prepare the dataset as the following structure:
├── train
├── images
├── img_1.png
├── img_2.png
├── ...
├── degrade
# images processed by Selective Degradation (refer to our paper for detals)
├── degraded_img_1.png
├── degraded_img_2.png
├── ...
├── labels
├── label_1.txt
├── label_2.txt
├── ...
├── val
├── test
Note that apart from the images and labels in LEVIR-Ship dataset, you should also generate the degraded images, which are the supervision of the enhancer (see details in our paper). Here, we provide DegradeGenerate.py to easily generate the degraded images.
After preparing the dataset as above, change the paths in ship.yaml.
(The partitioned dataset, including the degraded images, can all be accessed here)
-
Windows/Linux Support
-
python 3.8
-
pytorch 1.9.0
-
torchvision
-
wandb (Suggested, a good tool to visualize the training process. If not want to use it, you should comment out the related codes.)
-
...... (See more details in requirements.txt)
(The code is constructed based on YOLOv5s, for more details about YOLOv5, please refer to their repo here.)
To train our DRENet
, run:
python train.py --cfg "./models/DRENet.yaml" --epochs 1000 --workers 8 --batch-size 16 --device 0 --project "./LEVIR-Ship" --data "./data/ship.yaml"
Parameters Description
cfg
: You can change it to use different network structures. More structure configurations can be found in models directory, where we provide the baseline YOLOv5s, and the ablation structures of DRENet. You can try them if you are interestedepochs
: A longer training time is suggested, and 1,000 epochs are enough.project
: The path where you want to save your experiments. Also the name of the project in wandb.
Others
The current codes use fixed weight balance, which can also achieve a good result.
If you want to make use of automatic weight balance, please search the key word weightOptimizer
in train.py and uncomment the code lines, also the code lines with the key word ForAuto
in loss.py be uncommented and the other lines be commented out.
To evaluate our DRENet
, you should first train the network or download our provided weights, then run:
python test.py --weights "./DRENet.pt" --project "runs/test" --device 0 --batch-size 16 --data "./data/ship.yaml"
You can set how many detected results to plot by changing the value of plot_batch_num
in test.py. Also ensure that you have changed the val path in ship.yaml into your test path.
Please ensure that there are corresponding degraded images in the degrade folder. (See #issue 4 for more details.)
To directly output the detect results without the need of the degraded images, please run the following command:
python detect.py --weights "./DRENet.pt" --source "images/" --device 0
where "--source" is the path that the images need detection in.
If you find this paper useful in your research, please consider citing:
@ARTICLE{9791363,
author={Chen, Jianqi and Chen, Keyan and Chen, Hao and Zou, Zhengxia and Shi, Zhenwei},
journal={IEEE Transactions on Geoscience and Remote Sensing},
title={A Degraded Reconstruction Enhancement-based Method for Tiny Ship Detection in Remote Sensing Images with A New Large-scale Dataset},
year={2022},
volume={60},
number={},
pages={1-14},
doi={10.1109/TGRS.2022.3180894}}
This project is licensed under the GPL-3.0 License. See LICENSE for details