Movatterモバイル変換


[0]ホーム

URL:


Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Sign up
Appearance settings
#

llava

Here are 249 public repositories matching this topic...

[NeurIPS'23 Oral] Visual Instruction Tuning (LLaVA) built towards GPT-4V level capabilities and beyond.

  • UpdatedAug 12, 2024
  • Python

SUPIR aims at developing Practical Algorithms for Photo-Realistic Image Restoration In the Wild. Our new online demo is also released at suppixel.ai.

  • UpdatedMay 12, 2025
  • Python

Open-source evaluation toolkit of large multi-modality models (LMMs), support 220+ LMMs, 80+ benchmarks

  • UpdatedFeb 20, 2026
  • Python

中文nlp解决方案(大模型、数据、模型、训练、推理)

  • UpdatedAug 5, 2025
  • Jupyter Notebook

A C#/.NET library to run LLM (🦙LLaMA/LLaVA) on your local device efficiently.

  • UpdatedFeb 17, 2026
  • C#

MLX-VLM is a package for inference and fine-tuning of Vision Language Models (VLMs) on your Mac using MLX.

  • UpdatedFeb 20, 2026
  • Python

ChatGPT爆火,开启了通往AGI的关键一步,本项目旨在汇总那些ChatGPT的开源平替们,包括文本大模型、多模态大模型等,为大家提供一些便利

  • UpdatedAug 14, 2023

[ACL 2024 🔥] Video-ChatGPT is a video conversation model capable of generating meaningful conversation about videos. It combines the capabilities of LLMs with a pretrained visual encoder adapted for spatiotemporal video representation. We also introduce a rigorous 'Quantitative Evaluation Benchmarking' for video-based conversational models.

  • UpdatedAug 5, 2025
  • Python

Tag manager and captioner for image datasets

  • UpdatedOct 11, 2025
  • Python
UForm

Pocket-Sized Multimodal AI for content understanding and generation across multilingual texts, images, and 🔜 video, up to 5x faster than OpenAI CLIP and LLaVA 🖼️ & 🖋️

  • UpdatedOct 30, 2025
  • Python

Famous Vision Language Models and Their Architectures

  • UpdatedJan 11, 2026
  • Markdown

A Framework of Small-scale Large Multimodal Models

  • UpdatedFeb 7, 2026
  • Python

Eagle: Frontier Vision-Language Models with Data-Centric Strategies

  • UpdatedOct 25, 2025
  • Python

OpenCV+YOLO+LLAVA powered video surveillance system

  • UpdatedOct 21, 2025
  • Python

Fully Open Framework for Democratized Multimodal Training

  • UpdatedDec 27, 2025
  • Python

Paddle Multimodal Integration and eXploration, supporting mainstream multi-modal tasks, including end-to-end large-scale multi-modal pretrain models and diffusion model toolbox. Equipped with high performance and flexibility.

  • UpdatedFeb 3, 2026
  • Python

👁️ + 💬 + 🎧 = 🤖 Curated list of top foundation and multimodal models! [Paper + Code + Examples + Tutorials]

  • UpdatedFeb 29, 2024
  • Python

LLaVA-Mini is a unified large multimodal model (LMM) that can support the understanding of images, high-resolution images, and videos in an efficient manner.

  • UpdatedJun 29, 2025
  • Python

Improve this page

Add a description, image, and links to thellava topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with thellava topic, visit your repo's landing page and select "manage topics."

Learn more


[8]ページ先頭

©2009-2026 Movatter.jp