Computer Science > Machine Learning
arXiv:2410.09132 (cs)
[Submitted on 11 Oct 2024 (v1), last revised 27 Feb 2025 (this version, v2)]
Title:When Graph meets Multimodal: Benchmarking and Meditating on Multimodal Attributed Graphs Learning
Authors:Hao Yan,Chaozhuo Li,Jun Yin,Zhigang Yu,Weihao Han,Mingzheng Li,Zhengxin Zeng,Hao Sun,Senzhang Wang
View a PDF of the paper titled When Graph meets Multimodal: Benchmarking and Meditating on Multimodal Attributed Graphs Learning, by Hao Yan and 8 other authors
View PDFHTML (experimental)Abstract:Multimodal Attributed Graphs (MAGs) are ubiquitous in real-world applications, encompassing extensive knowledge through multimodal attributes attached to nodes (e.g., texts and images) and topological structure representing node interactions. Despite its potential to advance diverse research fields like social networks and e-commerce, MAG representation learning (MAGRL) remains underexplored due to the lack of standardized datasets and evaluation frameworks. In this paper, we first propose MAGB, a comprehensive MAG benchmark dataset, featuring curated graphs from various domains with both textual and visual attributes. Based on MAGB dataset, we further systematically evaluate two mainstream MAGRL paradigms: $\textit{GNN-as-Predictor}$, which integrates multimodal attributes via Graph Neural Networks (GNNs), and $\textit{VLM-as-Predictor}$, which harnesses Vision Language Models (VLMs) for zero-shot reasoning. Extensive experiments on MAGB reveal following critical insights: $\textit{(i)}$ Modality significances fluctuate drastically with specific domain characteristics. $\textit{(ii)}$ Multimodal embeddings can elevate the performance ceiling of GNNs. However, intrinsic biases among modalities may impede effective training, particularly in low-data scenarios. $\textit{(iii)}$ VLMs are highly effective at generating multimodal embeddings that alleviate the imbalance between textual and visual attributes. These discoveries, which illuminate the synergy between multimodal attributes and graph topologies, contribute to reliable benchmarks, paving the way for future MAG research. The MAGB dataset and evaluation pipeline are publicly available atthis https URL.
Subjects: | Machine Learning (cs.LG); Artificial Intelligence (cs.AI); Computer Vision and Pattern Recognition (cs.CV) |
Cite as: | arXiv:2410.09132 [cs.LG] |
(orarXiv:2410.09132v2 [cs.LG] for this version) | |
https://doi.org/10.48550/arXiv.2410.09132 arXiv-issued DOI via DataCite |
Submission history
From: Hao Yan [view email][v1] Fri, 11 Oct 2024 13:24:57 UTC (3,439 KB)
[v2] Thu, 27 Feb 2025 14:51:24 UTC (5,705 KB)
Full-text links:
Access Paper:
- View PDF
- HTML (experimental)
- TeX Source
- Other Formats
View a PDF of the paper titled When Graph meets Multimodal: Benchmarking and Meditating on Multimodal Attributed Graphs Learning, by Hao Yan and 8 other authors
References & Citations
Bibliographic and Citation Tools
Bibliographic Explorer(What is the Explorer?)
Connected Papers(What is Connected Papers?)
Litmaps(What is Litmaps?)
scite Smart Citations(What are Smart Citations?)
Code, Data and Media Associated with this Article
alphaXiv(What is alphaXiv?)
CatalyzeX Code Finder for Papers(What is CatalyzeX?)
DagsHub(What is DagsHub?)
Gotit.pub(What is GotitPub?)
Hugging Face(What is Huggingface?)
Papers with Code(What is Papers with Code?)
ScienceCast(What is ScienceCast?)
Demos
Recommenders and Search Tools
Influence Flower(What are Influence Flowers?)
CORE Recommender(What is CORE?)
IArxiv Recommender(What is IArxiv?)
arXivLabs: experimental projects with community collaborators
arXivLabs is a framework that allows collaborators to develop and share new arXiv features directly on our website.
Both individuals and organizations that work with arXivLabs have embraced and accepted our values of openness, community, excellence, and user data privacy. arXiv is committed to these values and only works with partners that adhere to them.
Have an idea for a project that will add value for arXiv's community?Learn more about arXivLabs.