phi-3-vision
Here are 9 public repositories matching this topic...
Language:All
streamline the fine-tuning process for multimodal models: PaliGemma 2, Florence-2, and Qwen2.5-VL
- Updated
Mar 17, 2025 - Python
🔥🔥 LLaVA++: Extending LLaVA with Phi-3 and LLaMA-3 (LLaVA LLaMA-3, LLaVA Phi-3)
- Updated
Jul 10, 2024 - Python
Azure OpenAI (demos, documentation, accelerators).
- Updated
Mar 14, 2025 - Jupyter Notebook
Phi-3.5 for Mac: Locally-run Vision and Language Models for Apple Silicon
- Updated
Sep 7, 2024 - Jupyter Notebook
Chat with Phi 3.5/3 Vision LLMs. Phi-3.5-vision is a lightweight, state-of-the-art open multimodal model built upon datasets which include - synthetic data and filtered publicly available websites - with a focus on very high-quality, reasoning dense data both on text and vision.
- Updated
Jan 2, 2025 - Jupyter Notebook
Phi-3-Vision model test - running locally
- Updated
May 29, 2024 - Jupyter Notebook
Microsoft Phi-3 Vision-the first Multimodal model By Microsoft- Demo With Huggingface
- Updated
May 24, 2024
Microsoft Phi-3 Vision-the first Multimodal model By Microsoft- Demo With Huggingface
- Updated
May 26, 2024 - Jupyter Notebook
Improve this page
Add a description, image, and links to thephi-3-vision topic page so that developers can more easily learn about it.
Add this topic to your repo
To associate your repository with thephi-3-vision topic, visit your repo's landing page and select "manage topics."