multi-modal
Here are 363 public repositories matching this topic...
Language:All
Sort:Most stars
MiniCPM-o 2.6: A GPT-4o Level MLLM for Vision, Speech and Multimodal Live Streaming on Your Phone
- Updated
Mar 3, 2025 - Python
Database for AI. Store Vectors, Images, Texts, Videos, etc. Use with LLMs/LangChain. Store, query, version, & visualize any AI data. Stream data in real-time to PyTorch/TensorFlow.https://activeloop.ai
- Updated
Apr 23, 2025 - Python
[CVPR 2024 Oral] InternVL Family: A Pioneering Open-Source Alternative to GPT-4o. 接近GPT-4o表现的开源多模态对话模型
- Updated
Apr 27, 2025 - Python
ModelScope: bring the notion of Model-as-a-Service to life.
- Updated
Apr 30, 2025 - Python
Start building LLM-empowered multi-agent applications in an easier way.
- Updated
Apr 30, 2025 - Python
a state-of-the-art-level open visual language model | 多模态预训练模型
- Updated
May 29, 2024 - Python
Implementation / replication of DALL-E, OpenAI's Text to Image Transformer, in Pytorch
- Updated
Feb 17, 2024 - Python
Chinese version of CLIP which achieves Chinese cross-modal retrieval and representation generation.
- Updated
Aug 6, 2024 - Python
Unified embedding generation and search engine. Also available on cloud - cloud.marqo.ai
- Updated
May 1, 2025 - Python
Open Source Routing Engine for OpenStreetMap
- Updated
Apr 30, 2025 - C++
Data processing for and with foundation models! 🍎 🍋 🌽 ➡️ ➡️🍸 🍹 🍷
- Updated
Apr 30, 2025 - Python
Chinese and English multimodal conversational language model | 多模态中英双语对话语言模型
- Updated
Aug 23, 2024 - Python
OmniGen: Unified Image Generation.https://arxiv.org/pdf/2409.11340
- Updated
Feb 20, 2025 - Jupyter Notebook
[EMNLP 2022] An Open Toolkit for Knowledge Graph Extraction and Construction
- Updated
Apr 22, 2025 - Python
【EMNLP 2024🔥】Video-LLaVA: Learning United Visual Representation by Alignment Before Projection
- Updated
Dec 3, 2024 - Python
Represent, send, store and search multimodal data
- Updated
Apr 24, 2025 - Python
GPT4V-level open-source multi-modal model based on Llama3-8B
- Updated
Mar 3, 2025 - Python
Open-source evaluation toolkit of large multi-modality models (LMMs), support 220+ LMMs, 80+ benchmarks
- Updated
Apr 30, 2025 - Python
Project Page for "LISA: Reasoning Segmentation via Large Language Model"
- Updated
Feb 16, 2025 - Python
Improve this page
Add a description, image, and links to themulti-modal topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with themulti-modal topic, visit your repo's landing page and select "manage topics."