This is the official repository of work:MS-DETR: Multispectral Pedestrian Detection Transformer with Loosely Coupled Fusion and Modality-Balanced Optimization.
- 2024/7/12 Release our MS-DETR code. At the same time we are focusing on developing RGBT models that are more efficient than MS-DETR. Stay tuned.
- 2023/7/21 Build the official repository of our MS-DETR and upload the evalution scripts and the detection results of our MS-DETR and other sota multispectral detectors on the KAIST dataset. We will update all codes and models after our work is accepted.
To install the necessary dependencies for this project, please follow the steps below:
-
Install Dependencies: Navigate to the directory containing the
requirements.txtfile and run the following command to install all required packages:conda install --file ./requirements.txt
-
Ensure Proper Installation of MSDA Operator: After installing the dependencies, ensure the MSDA operator is correctly installed by navigating to the
exp/official_repo/MS-DETR/models/dab_deformable_detr/opsdirectory:cd exp/official_repo/MS-DETR/models/dab_deformable_detr/ops bash make.sh # install
- Pretrain model link GoogleDisk. Please download them and place them in the pretrain_models directory.
- Please refer to the exp_config folder for more details. There are training and test script commands in each yaml file. Take KAIST training as an example.
torchrun --nproc_per_node=4 --master_port=49104 fusion_main.py --exp_config exp_config/KAIST/kaist.yaml --output_dir <path of your work dir> --action trainpython fusion_main.py --output_dir <output dir> --action test --resume <path of checkpoint.pth> --exp_config <path of exp config>You can evaluate the result files of the models with code.
We draw all the results of state-of-the-art methods in a single figure to make it easy to compare, and the figure represents the miss-rate against false positives per image.
For annotations file, only json is supported, and for result files, json and txt formats are supported.
(multiple --rstFiles are supported)
Example
$ python evaluation_script.py \
--annFile KAIST_annotation.json \
--rstFile state_of_arts/ACF_result.txt \
state_of_arts/ARCNN_result.txt \
state_of_arts/CIAN_result.txt \
state_of_arts/Fusion-RPN+BF_result.txt \
state_of_arts/Halfway-Fusion_result.txt \
state_of_arts/IAF-RCNN_result.txt \
state_of_arts/IATDNN-IAMSS_result.txt \
state_of_arts/MBNet_result.txt \
state_of_arts/GAFF_result.txt \
state_of_arts/MLPD_result.txt \
state_of_arts/MSDS-RCNN_result.txt If you find this code helpful, please kindly cite:
@ARTICLE{10669167,
author={Xing, Yinghui and Yang, Shuo and Wang, Song and Zhang, Shizhou and Liang, Guoqiang and Zhang, Xiuwei and Zhang, Yanning},
journal={IEEE Transactions on Intelligent Transportation Systems},
title={MS-DETR: Multispectral Pedestrian Detection Transformer With Loosely Coupled Fusion and Modality-Balanced Optimization},
year={2024},
volume={},
number={},
pages={1-15},
keywords={Multispectral pedestrian detection;end-to-end detector;loosely coupled fusion;modality-balanced optimization},
doi={10.1109/TITS.2024.3450584}}
