ControlFusion: A Controllable Image Fusion Framework with Language-Vision Degradation Prompts [NeurIPS 2025]
*Equal Contribution †Corresponding Author
-
Clone this repository:
git clone https://github.com/Linfeng-Tang/ControlFusion.git cd ControlFusion
-
Create a Conda environment (recommended):
conda create -n controlfusion python=3.8 -y conda activate controlfusion
-
Install dependency packages:
pip install -r requirements.txt
please refer to genDateset,To simulate light degradation, use Lightroom Classic
Our dataset will be open sourced soon.
Download the pretrained model Mask-DiFuser from Baidu Drive, and put the weight into pretrained_weights/
.
You can use the test.py
script we provide to fuse pairs of images. Please make sure you have downloaded the pre-trained weights.
You can modify ControlFusion.py to select text/auto control by:
text_features = self.get_text_feature(text.expand(b, -1)).to(inp_img_A.dtype)
text_features = imgfeature
You can use the train.py
script we provide to train. Make sure you have organized your train dataset correctly.
If our work is useful for your research, please consider citing and give us a star ⭐:
@inproceedings{Tang2024Mask-DiFuser,
author={Linfeng Tang, Yeda Wang, Zhanchuan Cai, Junjun Jiang, and Jiayi Ma},
title={ControlFusion: A Controllable Image Fusion Network with Language-Vision Degradation Prompts},
booktitle={Advances in Neural Information Processing Systems},
year={2025},
}
Please feel free to contact: [email protected], [email protected]
.
We are very pleased to communicate with you and will maintain this repository during our free time.