multi-modal-fusion
Here are 28 public repositories matching this topic...
Language:All
Sort:Most stars
Knowledge Graphs Meet Multi-Modal Learning: A Comprehensive Survey
- Updated
Dec 10, 2024
[Paper][AAAI 2025] (MyGO)Tokenization, Fusion, and Augmentation: Towards Fine-grained Multi-modal Entity Representation
- Updated
Dec 19, 2024 - Python
The official repository of Achelous and Achelous++
- Updated
Jul 8, 2024 - Python
Seed, Code, Harvest: Grow Your Own App with Tree of Thoughts!
- Updated
Jul 27, 2023 - Python
[IEEE TCYB 2023] The first large-scale tracking dataset by fusing RGB and Event cameras.
- Updated
Feb 14, 2025 - Python
Implementation of MoE Mamba from the paper: "MoE-Mamba: Efficient Selective State Space Models with Mixture of Experts" in Pytorch and Zeta
- Updated
Jan 27, 2025 - Python
This repository contains the source code for our paper: "Husformer: A Multi-Modal Transformer for Multi-Modal Human State Recognition". For more details, please refer to our paper athttps://arxiv.org/abs/2209.15182.
- Updated
Jul 2, 2023 - Python
Code for J. Wang, J. Li, Y. Shi, J. Lai and X. Tan, "AM3Net: Adaptive Mutual-learning-based Multimodal Data Fusion Network," in IEEE TCSVT, 2022. We conducted the experiments on the hyperspectral and lidar dataset(Houston and Trento) and multispectral and synthetic aperture radar data (grss-dfc-2007 datasets).
- Updated
Mar 27, 2023 - Python
Training for multi-modal image fusion with PyTorch.
- Updated
Nov 30, 2023 - Python
[Paper][SIGIR 2024] NativE: Multi-modal Knowledge Graph Completion in the Wild
- Updated
Aug 12, 2024 - Python
[IV'24] UniBEV: the official implementation of UniBEV
- Updated
Jun 26, 2024 - Python
[Paper][LREC-COLING 2024] Unleashing the Power of Imbalanced Modality Information for Multi-modal Knowledge Graph Completion
- Updated
Apr 16, 2024 - Python
The open source implementation of the model from "Scaling Vision Transformers to 22 Billion Parameters"
- Updated
Jan 27, 2025 - Python
[CVPR-2023 Workshop@NFVLR] Official PyTorch implementation of Learning CLIP Guided Visual-Text Fusion Transformer for Video-based Pedestrian Attribute Recognition
- Updated
Jun 11, 2024 - Python
Adaptive Confidence Multi-View Hashing
- Updated
Dec 13, 2023 - Python
[Paper][ICLR 2025] Multiple Heads are Better than One: Mixture of Modality Knowledge Experts for Entity Representation Learning
- Updated
Mar 14, 2025 - Python
The official implementation of "TFormer: A throughout fusion transformer for multi-modal skin lesion diagnosis"
- Updated
Jan 29, 2024 - Python
IEEE 802.11n CSI and camera synchronization toolkit.
- Updated
Dec 25, 2024 - C
[CHI2021] Hidden emotion detection using multi-modal signals
- Updated
Sep 30, 2021 - Python
Multi-Modal Attention-based Hierarchical Graph Neural Network for Object Interaction Recommendation in Internet of Things (IoT)
- Updated
Dec 15, 2021 - Python
Improve this page
Add a description, image, and links to themulti-modal-fusion topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with themulti-modal-fusion topic, visit your repo's landing page and select "manage topics."