Official code for Paper "Mantis: Multi-Image Instruction Tuning"
-
Updated
Jul 12, 2024 - Python
Official code for Paper "Mantis: Multi-Image Instruction Tuning"
This project is the official implementation of 'LLMGA: Multimodal Large Language Model based Generation Assistant', ECCV2024
DenseFusion-1M: Merging Vision Experts for Comprehensive Multimodal Perception
InternLM-XComposer-2.5: A Versatile Large Vision Language Model Supporting Long-Contextual Input and Output
A toolbox for benchmarking trustworthiness of multimodal large language models (MultiTrust)
Large-scale Self-supervised Pre-training Across Tasks, Languages, and Modalities
The code for "TokenPacker: Efficient Visual Projector for Multimodal LLM".
Image Textualization: An Automatic Framework for Generating Rich and Detailed Image Descriptions
Cambrian-1 is a family of multimodal LLMs with a vision-centric design.
This is the official implementation (code, data) of the paper "MOSSBench: Is Your Multimodal Language Model Oversensitive to Safe Queries?""
MOSSBench: A webpage for an oversensitivity benchmark
EVE: Encoder-Free Vision-Language Models from BAAI
Personal Project: MPP-Qwen14B & MPP-Qwen-Next(Multimodal Pipeline Parallel based on Qwen-LM). Support [video/image/multi-image] {sft/conversations}. Don't let the poverty limit your imagination! Train your own 8B/14B LLaVA-training-like MLLM on RTX3090/4090 24GB.
[ECCV2024] Official code implementation of Merlin: Empowering Multimodal LLMs with Foresight Minds
Composition of Multimodal Language Models From Scratch
Mobile-Agent: The Powerful Mobile Device Operation Assistant Family
Add a description, image, and links to the mllm topic page so that developers can more easily learn about it.
To associate your repository with the mllm topic, visit your repo's landing page and select "manage topics."