School of Computer Science and Technology, Harbin Institute of Technology, Shenzhen
*Equal contribution
†Corresponding author
The Thirty-eighth Annual Conference on Neural Information Processing Systems (NeurIPS 2024)
- [12/2024] Code and checkpoints are released.
- [09/2024] Project page released!
- [09/2024] MoME has been accepted by NeurIPS 2024!
- [07/2024] Arxiv paper released.
This is the github repository of MoME: Mixture of Multimodal Experts for Generalist Multimodal Large Language Models. In this work, we propose a mixture of multimodal experts (MoME) to mitigate task interference and obtain a generalist MLLM.
Our MoME is composed of two key components, a mixture of vision experts (MoVE) and a mixture of language experts (MoLE). MoVE can adaptively modulate the features transformed from various vision encoders, and has a strong compatibility in transformation architecture. MoLE incorporates sparsely gated experts into LLMs to achieve painless improvements with roughly unchanged inference costs.
The architecture of the proposed MoME model:
git clone https://github.com/JiuTian-VL/MoME.git
cd MoME
conda create -n mome python=3.12
conda activate mome
pip install -r requirements.txt
Please download all the required checkpoints by running the download_ckpt.py
script.
python download_ckpt.py
The required checkpoints will be downloaded to the ./checkpoints
directory from huggingface.
We provide an inference example in playground.ipynb
, which includes a minimal example of how to use the MoME model for inference.
A gradio demo used for model testing and router visualization is also provided in demo_mome.py
. You can start the demo by running the following command:
python demo_mome.py
We collected 24 datasets and categorized them into four groups for instruction-tuning and evaluation:
Here we list the multitasking performance comparison of MoME and baselines. Please refer to our paper for more details.
If you find this work useful for your research, please kindly cite our paper:
@inproceedings{shen2024mome,
title={MoME: Mixture of Multimodal Experts for Generalist Multimodal Large Language Models},
author={Shen, Leyang and Chen, Gongwei and Shao, Rui and Guan, Weili and Nie, Liqiang},
booktitle={Advances in neural information processing systems},
year={2024}
}