Skip to content

Multimodal Relation Extraction via a Mixture of Hierarchical Visual Context Learners. WWW 2024

License

Notifications You must be signed in to change notification settings

liuxiyang641/HVFormer

Folders and files

NameName
Last commit message
Last commit date

Latest commit

9b6a662 · Feb 21, 2024

History

6 Commits
Oct 12, 2023
Oct 12, 2023
Oct 12, 2023
Oct 12, 2023
Oct 12, 2023
Feb 21, 2024
Oct 13, 2023
Oct 12, 2023
Oct 12, 2023
Oct 12, 2023

Repository files navigation

Hierarchical Visual Context Fusion Transformer

The source code for Multimodal Relation Extraction via a Mixture of Hierarchical Visual Context Learners.

Data preprocessing

MNRE dataset

Due to the large size of MNRE dataset, please download the dataset from the original repository.

Unzip the data and rename the directory as mnre, which should be placed in the directory data:

mkdir data logs ckpt

We also use the detected visual objects provided in previous work, which can be downloaded using the commend:

cd data/
wget 120.27.214.45/Data/re/multimodal/data.tar.gz
tar -xzvf data.tar.gz

Dependencies

Install all necessary dependencies:

pip install -r requirements.txt

Training the model

The best hyperparameters we found have been witten in run_mre.sh file.

You can simply run the bash script for multimodal relation extraction:

bash run_mre.sh

About

Multimodal Relation Extraction via a Mixture of Hierarchical Visual Context Learners. WWW 2024

Resources

License

Stars

Watchers

Forks

Packages

No packages published