Source code for AAAI 2025 (Main Track) paper Attribution Analysis Meets Model Editing: Advancing Knowledge Correction in Vision Language Models with VisEdit.
- Please download the E-EVQA and E-IC datasets from the URL provided in [1] and place the related folders in the
data
directory. - Please modify the
ROOT_PATH
inutils/GLOBAL.py
to the absolute path of the current directory, and updatemodel_path_map
to the absolute paths of each backbone's weights.
Please run contribution_module.py
, using Jupyter Notebook would be better for display.
Please run contribution_visual_reps.py
, using Jupyter Notebook would be better for display.
Please use the following script to train a VEAD:
python vead_train.py -mn llava -dna EVQA -bs 4 -dvc "cuda:0" -edvc 1
Please use the following script to test VEAD:
python vead_test.py -mn llava -dn EVQA -dvc "cuda:0" -ckpt [vead_checkpoint_path]
Please cite our paper if you use VisEdit in your work (The AAAI citation is not yet available).
@article{DBLP:journals/corr/abs-2408-09916,
author = {Qizhou Chen and
Taolin Zhang and
Chengyu Wang and
Xiaofeng He and
Dakan Wang and
Tingting Liu},
title = {Attribution Analysis Meets Model Editing: Advancing Knowledge Correction
in Vision Language Models with VisEdit},
journal = {CoRR},
volume = {abs/2408.09916},
year = {2024},
url = {https://doi.org/10.48550/arXiv.2408.09916},
doi = {10.48550/ARXIV.2408.09916},
eprinttype = {arXiv},
eprint = {2408.09916},
timestamp = {Fri, 15 Nov 2024 07:55:45 +0100},
biburl = {https://dblp.org/rec/journals/corr/abs-2408-09916.bib},
bibsource = {dblp computer science bibliography, https://dblp.org}
}
[1] Can We Edit Multimodal Large Language Models?