Skip to content

Source code of our MM'24 paper Multimodal LLM Enhanced Cross-lingual Cross-modal Retrieval

Notifications You must be signed in to change notification settings

LiJiaBei-7/leccr

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

11 Commits
 
 
 
 
 
 

Repository files navigation

Multimodal LLM Enhanced Cross-lingual Cross-modal Retrieval

source code of our paper Multimodal LLM Enhanced Cross-lingual Cross-modal Retrieval

image

Data

The training and inference data can be obtained from the NRCCR, and we use the videochat2 to generate the image description.

Training

cd LECCR
sh run_multi30k.sh
sh run_mscoco.sh
sh run_video.sh

The codes are modified from NRCCR and CCLM.

Reference

If you find the package useful, please consider citing our paper:

@inproceedings{wang2024multimodal,
  title={Multimodal llm enhanced cross-lingual cross-modal retrieval},
  author={Wang, Yabing and Wang, Le and Zhou, Qiang and Wang, Zhibin and Li, Hao and Hua, Gang and Tang, Wei},
  booktitle={Proceedings of the 32nd ACM International Conference on Multimedia},
  pages={8296--8305},
  year={2024}
}

About

Source code of our MM'24 paper Multimodal LLM Enhanced Cross-lingual Cross-modal Retrieval

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published