- Notifications
You must be signed in to change notification settings - Fork34
[SIGIR 2022] Hybrid Transformer with Multi-level Fusion for Multimodal Knowledge Graph Completion
License
zjunlp/MKGformer
Folders and files
| Name | Name | Last commit message | Last commit date | |
|---|---|---|---|---|
Repository files navigation
Code for the SIGIR 2022 paper "Hybrid Transformer with Multi-level Fusion for Multimodal Knowledge Graph Completion"
- ❗NOTE: We provide some KGE baselines atOpenBG-IMG.
- ❗NOTE: We release a new MKG task "Multimodal Analogical Reasoning over Knowledge Graphs (ICLR'2023)" atMKG_Analogy.
Illustration of MKGformer for (a) Unified Multimodal KGC Framework and (b) Detailed M-Encoder.
To run the codes (Python 3.8), you need to install the requirements:
pip install -r requirements.txtTo extract visual object images int MNER and MRE tasks, we first use the NLTK parser to extract noun phrases from the text and apply thevisual grouding toolkit to detect objects. Detailed steps are as follows:
- Using the NLTK parser (or Spacy, textblob) to extract noun phrases from the text.
- Applying thevisual grouding toolkit to detect objects. Taking the twitter2017 dataset as an example, the extracted objects are stored in
twitter2017_aux_images. The images of the object obey the following naming format:id_pred_yolo_crop_num.png, whereidis the order of the raw image corresponding to the object,numis the number of the object predicted by the toolkit. (idis doesn't matter.) - Establishing the correspondence between the raw images and the objects. We construct a dictionary to record the correspondence between the raw images and the objects. Taking
twitter2017/twitter2017_train_dict.pthas an example, the format of the dictionary can be seen as follows:{imgname:['id_pred_yolo_crop_num0.png', 'id_pred_yolo_crop_num1.png', ...] }, where key is the name of raw images, value is a List of the objects (Note that intrain/val/test.txt, text and raw image have a one-to-one relationship, so theimgnaecan be used as a unique identifier for the raw images).
The detected objects and the dictionary of the correspondence between the raw images and the objects are available in our data links.
The datasets that we used in our experiments are as follows:
Twitter2017
You can download the twitter2017 dataset fromGoogle Drive.
For more information regarding the dataset, please refer to theUMT repository.
MRE
The MRE dataset comes fromMEGA, many thanks.
You can download theMRE dataset with detected visual objects fromGoogle Drive or using following command:
cd MREwget 121.41.117.246/Data/re/multimodal/data.tar.gztar -xzvf data.tar.gzMKG
FB15K-237-IMG
You can download the image data of FB15k-237 frommmkb which provides a list of image URLs, and refer to more information of description of entity fromkg-bert repositories.
- ❗NOTE: we have found a severe bug in the code of data preprocessing for FB15k-237-IMG, which leads to the unfair performance comparison; we have updated the performance inarxiv and released thecheckpoints (The model trained with/without the severe bug).
WN18-IMG
Entity images in WN18 can be obtained from ImageNet, the specific steps can refer to RSME. theRSME repository.
We also provide additional network disk links formultimodal KG data (Images) atGoogleDrive orBaidu Pan with extraction (code:ilbd).
The expected structure of files is:
MKGFormer |-- MKG# Multimodal Knowledge Graph | |-- dataset # task data | |-- data # data process file | |-- lit_models # lightning model | |-- models # mkg model | |-- scripts # running script | |-- main.py |-- MNER# Multimodal Named Entity Recognition | |-- data # task data | | |-- twitter2017 | | | |-- twitter17_detect # rcnn detected objects | | | |-- twitter2017_aux_images # visual grounding objects | | | |-- twitter2017_images # raw images | | | |-- train.txt # text data | | | |-- ... | | | |-- twitter2017_train_dict.pth # {imgname: [object-image]} | | | |-- ... | |-- models # mner model | |-- modules # running script | |-- processor # data process file | |-- utils | |-- run_mner.sh | |-- run.py |-- MRE # Multimodal Relation Extraction | |-- data # task data | | |-- img_detect # rcnn detected objects | | |-- img_org # raw images | | |-- img_vg # visual grounding objects | | |-- txt # text data | | | |-- ours_train.txt | | | |-- ours_val.txt | | | |-- ours_test.txt | | | |-- mre_train_dict.pth # {imgid: [object-image]} | | | |-- ... | | |-- vg_data # [(id, imgname, noun_phrase)], not useful | | |-- ours_rel2id.json # relation data | |-- models # mre model | |-- modules # running script | |-- processor # data process file | |-- run_mre.sh | |-- run.py- First run Image-text Incorporated Entity Modeling to train entity embedding.
cd MKG bash scripts/pretrain_fb15k-237-image.sh- Then do Missing Entity Prediction.
bash scripts/fb15k-237-image.sh
To run mner task, run this script.
cd MNERbash run_mner.shTo run mre task, run this script.
cd MREbash run_mre.sh
The acquisition of image data for the multimodal link prediction task refer to the code fromhttps://github.com/wangmengsd/RSME, many thanks.
If you use or extend our work, please cite the paper as follows:
@inproceedings{DBLP:conf/sigir/ChenZLDTXHSC22,author ={Xiang Chen and Ningyu Zhang and Lei Li and Shumin Deng and Chuanqi Tan and Changliang Xu and Fei Huang and Luo Si and Huajun Chen},editor ={Enrique Amig{\'{o}} and Pablo Castells and Julio Gonzalo and Ben Carterette and J. Shane Culpepper and Gabriella Kazai},title ={Hybrid Transformer with Multi-level Fusion for Multimodal Knowledge Graph Completion},booktitle ={{SIGIR} '22: The 45th International {ACM} {SIGIR} Conference on Research and Development in Information Retrieval, Madrid, Spain, July 11 - 15, 2022},pages ={904--915},publisher ={{ACM}},year ={2022},url ={https://doi.org/10.1145/3477495.3531992},doi ={10.1145/3477495.3531992},timestamp ={Mon, 11 Jul 2022 12:19:20 +0200},biburl ={https://dblp.org/rec/conf/sigir/ChenZLDTXHSC22.bib},bibsource ={dblp computer science bibliography, https://dblp.org}}
About
[SIGIR 2022] Hybrid Transformer with Multi-level Fusion for Multimodal Knowledge Graph Completion
Topics
Resources
License
Uh oh!
There was an error while loading.Please reload this page.
Stars
Watchers
Forks
Releases
Packages0
Uh oh!
There was an error while loading.Please reload this page.
Contributors5
Uh oh!
There was an error while loading.Please reload this page.
