VLMEvalKit / vlmeval /config.py
Racktic's picture
Upload folder using huggingface_hub
b5beb60 verified
from vlmeval.vlm import *
from vlmeval.api import *
from functools import partial
PandaGPT_ROOT = None
MiniGPT4_ROOT = None
TransCore_ROOT = None
Yi_ROOT = None
OmniLMM_ROOT = None
Mini_Gemini_ROOT = None
VXVERSE_ROOT = None
VideoChat2_ROOT = None
VideoChatGPT_ROOT = None
PLLaVA_ROOT = None
RBDash_ROOT = None
VITA_ROOT = None
LLAVA_V1_7B_MODEL_PTH = "Please set your local path to LLaVA-7B-v1.1 here, the model weight is obtained by merging LLaVA delta weight based on vicuna-7b-v1.1 in https://github.com/haotian-liu/LLaVA/blob/main/docs/MODEL_ZOO.md with vicuna-7b-v1.1. "
video_models = {
"Video-LLaVA-7B": partial(VideoLLaVA, model_path="LanguageBind/Video-LLaVA-7B"),
"Video-LLaVA-7B-HF": partial(
VideoLLaVA_HF, model_path="LanguageBind/Video-LLaVA-7B-hf"
),
"VideoChat2-HD": partial(
VideoChat2_HD,
model_path="OpenGVLab/VideoChat2_HD_stage4_Mistral_7B",
root=VideoChat2_ROOT,
config_file="./vlmeval/vlm/video_llm/configs/videochat2_hd.json",
),
"Chat-UniVi-7B": partial(Chatunivi, model_path="Chat-UniVi/Chat-UniVi"),
"Chat-UniVi-7B-v1.5": partial(
Chatunivi, model_path="Chat-UniVi/Chat-UniVi-7B-v1.5"
),
"LLaMA-VID-7B": partial(
LLaMAVID, model_path="YanweiLi/llama-vid-7b-full-224-video-fps-1"
),
"Video-ChatGPT": partial(
VideoChatGPT, model_path="MBZUAI/Video-ChatGPT-7B", dir_root=VideoChatGPT_ROOT
),
"PLLaVA-7B": partial(PLLaVA, model_path="ermu2001/pllava-7b", dir_root=PLLaVA_ROOT),
"PLLaVA-13B": partial(
PLLaVA, model_path="ermu2001/pllava-13b", dir_root=PLLaVA_ROOT
),
"PLLaVA-34B": partial(
PLLaVA, model_path="ermu2001/pllava-34b", dir_root=PLLaVA_ROOT
),
}
ungrouped = {
"AKI": partial(AKI, name="AKI", ckpt_pth="Sony/AKI-4B-phi-3.5-mini"),
"TransCore_M": partial(TransCoreM, root=TransCore_ROOT),
"PandaGPT_13B": partial(PandaGPT, name="PandaGPT_13B", root=PandaGPT_ROOT),
"flamingov2": partial(
OpenFlamingo,
name="v2",
mpt_pth="anas-awadalla/mpt-7b",
ckpt_pth="openflamingo/OpenFlamingo-9B-vitl-mpt7b",
),
"VisualGLM_6b": partial(VisualGLM, model_path="THUDM/visualglm-6b"),
"mPLUG-Owl2": partial(mPLUG_Owl2, model_path="MAGAer13/mplug-owl2-llama2-7b"),
"mPLUG-Owl3": partial(mPLUG_Owl3, model_path="mPLUG/mPLUG-Owl3-7B-240728"),
"OmniLMM_12B": partial(
OmniLMM12B, model_path="openbmb/OmniLMM-12B", root=OmniLMM_ROOT
),
"MGM_7B": partial(
Mini_Gemini, model_path="YanweiLi/MGM-7B-HD", root=Mini_Gemini_ROOT
),
"Bunny-llama3-8B": partial(BunnyLLama3, model_path="BAAI/Bunny-v1_1-Llama-3-8B-V"),
"VXVERSE": partial(VXVERSE, model_name="XVERSE-V-13B", root=VXVERSE_ROOT),
"360VL-70B": partial(QH_360VL, model_path="qihoo360/360VL-70B"),
"Llama-3-MixSenseV1_1": partial(
LLama3Mixsense, model_path="Zero-Vision/Llama-3-MixSenseV1_1"
),
"Parrot": partial(Parrot, model_path="AIDC-AI/Parrot-7B"),
"OmChat": partial(OmChat, model_path="omlab/omchat-v2.0-13B-single-beta_hf"),
"RBDash_72b": partial(
RBDash, model_path="RBDash-Team/RBDash-v1.5", root=RBDash_ROOT
),
"Pixtral-12B": partial(Pixtral, model_path="mistralai/Pixtral-12B-2409"),
"Falcon2-VLM-11B": partial(Falcon2VLM, model_path="tiiuae/falcon-11B-vlm"),
}
o1_key = 'XXX' # noqa: E501
o1_apis = {
'o1': partial(
GPT4V,
model="o1-2024-12-17",
key=o1_key,
api_base='OFFICIAL',
temperature=0,
img_detail='high',
retry=10,
verbose=False,
),
}
api_models = {
# GPT
"GPT4V": partial(
GPT4V,
model="gpt-4-1106-vision-preview",
temperature=0,
img_size=512,
img_detail="low",
retry=10,
verbose=False,
),
"GPT4V_HIGH": partial(
GPT4V,
model="gpt-4-1106-vision-preview",
temperature=0,
img_size=-1,
img_detail="high",
retry=10,
verbose=False,
),
"GPT4V_20240409": partial(
GPT4V,
model="gpt-4-turbo-2024-04-09",
temperature=0,
img_size=512,
img_detail="low",
retry=10,
verbose=False,
),
"GPT4V_20240409_HIGH": partial(
GPT4V,
model="gpt-4-turbo-2024-04-09",
temperature=0,
img_size=-1,
img_detail="high",
retry=10,
verbose=False,
),
"GPT4o": partial(
GPT4V,
model="gpt-4o-2024-05-13",
temperature=0,
img_size=512,
img_detail="low",
retry=10,
verbose=False,
),
"GPT4o_HIGH": partial(
GPT4V,
model="gpt-4o-2024-05-13",
temperature=0,
img_size=-1,
img_detail="high",
retry=10,
verbose=False,
),
"GPT4o_20240806": partial(
GPT4V,
model="gpt-4o-2024-08-06",
temperature=0,
img_size=-1,
img_detail="high",
retry=10,
verbose=False,
),
"GPT4o_20241120": partial(
GPT4V,
model="gpt-4o-2024-11-20",
temperature=0,
img_size=-1,
img_detail="high",
retry=10,
verbose=False,
),
"GPT4o_MINI": partial(
GPT4V,
model="gpt-4o-mini-2024-07-18",
temperature=0,
img_size=-1,
img_detail="high",
retry=10,
verbose=False,
),
"GPT4.5": partial(
GPT4V,
model='gpt-4.5-preview-2025-02-27',
temperature=0,
timeout=600,
img_size=-1,
img_detail='high',
retry=10,
verbose=False,
),
# Gemini
"GeminiPro1-0": partial(
GeminiProVision, model="gemini-1.0-pro", temperature=0, retry=10
), # now GeminiPro1-0 is only supported by vertex backend
"GeminiPro1-5": partial(
GeminiProVision, model="gemini-1.5-pro", temperature=0, retry=10
),
"GeminiFlash1-5": partial(
GeminiProVision, model="gemini-1.5-flash", temperature=0, retry=10
),
"GeminiFlash2-0": partial(
GeminiProVision, model="gemini-2.0-flash", temperature=0, retry=10
),
"GeminiPro2-0": partial(
GeminiProVision, model="gemini-2.0-pro-exp", temperature=0, retry=10
),
"GeminiPro1-5-002": partial(
GPT4V, model="gemini-1.5-pro-002", temperature=0, retry=10
), # Internal Use Only
"GeminiFlash1-5-002": partial(
GPT4V, model="gemini-1.5-flash-002", temperature=0, retry=10
), # Internal Use Only
# Qwen-VL
"QwenVLPlus": partial(QwenVLAPI, model="qwen-vl-plus", temperature=0, retry=10),
"QwenVLMax": partial(QwenVLAPI, model="qwen-vl-max", temperature=0, retry=10),
# Reka
"RekaEdge": partial(Reka, model="reka-edge-20240208"),
"RekaFlash": partial(Reka, model="reka-flash-20240226"),
"RekaCore": partial(Reka, model="reka-core-20240415"),
# Step1V
"Step1V": partial(
GPT4V,
model="step-1v-32k",
api_base="https://api.stepfun.com/v1/chat/completions",
temperature=0,
retry=10,
img_size=-1,
img_detail="high",
),
"Step1.5V-mini": partial(
GPT4V,
model="step-1.5v-mini",
api_base="https://api.stepfun.com/v1/chat/completions",
temperature=0,
retry=10,
img_size=-1,
img_detail="high",
),
"Step1o": partial(
GPT4V,
model="step-1o-vision-32k",
api_base="https://api.stepfun.com/v1/chat/completions",
temperature=0,
retry=10,
img_size=-1,
img_detail="high",
),
# Yi-Vision
"Yi-Vision": partial(
GPT4V,
model="yi-vision",
api_base="https://api.lingyiwanwu.com/v1/chat/completions",
temperature=0,
retry=10,
),
# Claude
"Claude3V_Opus": partial(
Claude3V, model="claude-3-opus-20240229", temperature=0, retry=10, verbose=False
),
"Claude3V_Sonnet": partial(
Claude3V,
model="claude-3-sonnet-20240229",
temperature=0,
retry=10,
verbose=False,
),
"Claude3V_Haiku": partial(
Claude3V,
model="claude-3-haiku-20240307",
temperature=0,
retry=10,
verbose=False,
),
"Claude3-5V_Sonnet": partial(
Claude3V,
model="claude-3-5-sonnet-20240620",
temperature=0,
retry=10,
verbose=False,
),
"Claude3-5V_Sonnet_20241022": partial(
Claude3V,
model="claude-3-5-sonnet-20241022",
temperature=0,
retry=10,
verbose=False,
),
"Claude3-7V_Sonnet": partial(
Claude3V,
model="claude-3-7-sonnet-20250219",
temperature=0,
retry=10,
verbose=False,
),
# GLM4V
"GLM4V": partial(GLMVisionAPI, model="glm4v-biz-eval", temperature=0, retry=10),
"GLM4V_PLUS": partial(GLMVisionAPI, model="glm-4v-plus", temperature=0, retry=10),
"GLM4V_PLUS_20250111": partial(
GLMVisionAPI, model="glm-4v-plus-0111", temperature=0, retry=10
),
# MiniMax abab
"abab6.5s": partial(
GPT4V,
model="abab6.5s-chat",
api_base="https://api.minimax.chat/v1/chat/completions",
temperature=0,
retry=10,
),
"abab7-preview": partial(
GPT4V,
model="abab7-chat-preview",
api_base="https://api.minimax.chat/v1/chat/completions",
temperature=0,
retry=10,
),
# CongRong
"CloudWalk": partial(CWWrapper, model="cw-congrong-v1.5", temperature=0, retry=10),
# SenseChat-V
"SenseChat-Vision": partial(
SenseChatVisionAPI, model="SenseChat-Vision", temperature=0, retry=10
),
"HunYuan-Vision": partial(
HunyuanVision, model="hunyuan-vision", temperature=0, retry=10
),
"HunYuan-Standard-Vision": partial(
HunyuanVision, model="hunyuan-standard-vision", temperature=0, retry=10
),
"BailingMM-Lite-1203": partial(
bailingMMAPI, model="BailingMM-Lite-1203", temperature=0, retry=10
),
"BailingMM-Pro-0120": partial(
bailingMMAPI, model="BailingMM-Pro-0120", temperature=0, retry=10
),
# BlueLM-V
"BlueLM_V": partial(BlueLM_V_API, model="BlueLM-VL-v3.0", temperature=0, retry=10),
# JiuTian-VL
"JTVL": partial(JTVLChatAPI, model="jt-vl-chat", temperature=0, retry=10),
"Taiyi": partial(TaiyiAPI, model="taiyi", temperature=0, retry=10),
# TeleMM
"TeleMM": partial(TeleMMAPI, model="TeleAI/TeleMM", temperature=0, retry=10),
# lmdeploy api
"lmdeploy": partial(
LMDeployAPI,
api_base="http://0.0.0.0:23333/v1/chat/completions",
temperature=0,
retry=10,
),
"lmdeploy_internvl_78B_MPO": partial(
LMDeployAPI,
api_base="http://0.0.0.0:23333/v1/chat/completions",
temperature=0,
retry=10,
timeout=100,
),
"lmdeploy_qvq_72B_preview": partial(
LMDeployAPI,
api_base="http://0.0.0.0:23333/v1/chat/completions",
temperature=0,
retry=10,
timeout=300,
),
# Taichu-VL
# "Taichu-VL-2B": partial(
# TaichuVLAPI,
# model="Taichu-VL-2B",
# url="https://platform.wair.ac.cn/api/v1/infer/10381/v1/chat/completions",
# ),
'Taichu-VLR-3B': partial(
TaichuVLRAPI,
model='taichu_vlr_3b',
url="https://platform.wair.ac.cn/maas/v1/chat/completions"
),
'Taichu-VLR-7B': partial(
TaichuVLRAPI,
model='taichu_vlr_7b',
url="https://platform.wair.ac.cn/maas/v1/chat/completions"
),
# doubao_vl
"DoubaoVL": partial(
DoubaoVL, model="Doubao-1.5-vision-pro", temperature=0, retry=10, verbose=False
),
# Shopee MUG-U
'MUG-U-7B': partial(
MUGUAPI,
model='MUG-U',
temperature=0,
retry=10,
verbose=False,
timeout=300),
# grok
"grok-vision-beta": partial(
GPT4V,
model="grok-vision-beta",
api_base="https://api.x.ai/v1/chat/completions",
temperature=0,
retry=10,
),
"grok-2-vision-1212": partial(
GPT4V,
model="grok-2-vision",
api_base="https://api.x.ai/v1/chat/completions",
temperature=0,
retry=10,
),
# kimi
"moonshot-v1-8k": partial(
GPT4V,
model="moonshot-v1-8k-vision-preview",
api_base="https://api.moonshot.cn/v1/chat/completions",
temperature=0,
retry=10,
),
"moonshot-v1-32k": partial(
GPT4V,
model="moonshot-v1-32k-vision-preview",
api_base="https://api.moonshot.cn/v1/chat/completions",
temperature=0,
retry=10,
),
"moonshot-v1-128k": partial(
GPT4V,
model="moonshot-v1-128k-vision-preview",
api_base="https://api.moonshot.cn/v1/chat/completions",
temperature=0,
retry=10,
),
}
emu_series = {
"emu2_chat": partial(Emu, model_path="BAAI/Emu2-Chat"),
"emu3_chat": partial(Emu3_chat, model_path="BAAI/Emu3-Chat"),
"emu3_gen": partial(Emu3_gen, model_path="BAAI/Emu3-Gen"),
}
mmalaya_series = {
"MMAlaya": partial(MMAlaya, model_path="DataCanvas/MMAlaya"),
"MMAlaya2": partial(MMAlaya2, model_path="DataCanvas/MMAlaya2"),
}
minicpm_series = {
"MiniCPM-V": partial(MiniCPM_V, model_path="openbmb/MiniCPM-V"),
"MiniCPM-V-2": partial(MiniCPM_V, model_path="openbmb/MiniCPM-V-2"),
"MiniCPM-Llama3-V-2_5": partial(
MiniCPM_Llama3_V, model_path="openbmb/MiniCPM-Llama3-V-2_5"
),
"MiniCPM-V-2_6": partial(MiniCPM_V_2_6, model_path="openbmb/MiniCPM-V-2_6"),
"MiniCPM-o-2_6": partial(MiniCPM_o_2_6, model_path="openbmb/MiniCPM-o-2_6"),
}
xtuner_series = {
"llava-internlm2-7b": partial(
LLaVA_XTuner,
llm_path="internlm/internlm2-chat-7b",
llava_path="xtuner/llava-internlm2-7b",
visual_select_layer=-2,
prompt_template="internlm2_chat",
),
"llava-internlm2-20b": partial(
LLaVA_XTuner,
llm_path="internlm/internlm2-chat-20b",
llava_path="xtuner/llava-internlm2-20b",
visual_select_layer=-2,
prompt_template="internlm2_chat",
),
"llava-internlm-7b": partial(
LLaVA_XTuner,
llm_path="internlm/internlm-chat-7b",
llava_path="xtuner/llava-internlm-7b",
visual_select_layer=-2,
prompt_template="internlm_chat",
),
"llava-v1.5-7b-xtuner": partial(
LLaVA_XTuner,
llm_path="lmsys/vicuna-7b-v1.5",
llava_path="xtuner/llava-v1.5-7b-xtuner",
visual_select_layer=-2,
prompt_template="vicuna",
),
"llava-v1.5-13b-xtuner": partial(
LLaVA_XTuner,
llm_path="lmsys/vicuna-13b-v1.5",
llava_path="xtuner/llava-v1.5-13b-xtuner",
visual_select_layer=-2,
prompt_template="vicuna",
),
"llava-llama-3-8b": partial(
LLaVA_XTuner,
llm_path="xtuner/llava-llama-3-8b-v1_1",
llava_path="xtuner/llava-llama-3-8b-v1_1",
visual_select_layer=-2,
prompt_template="llama3_chat",
),
}
qwen_series = {
"qwen_base": partial(QwenVL, model_path="Qwen/Qwen-VL"),
"qwen_chat": partial(QwenVLChat, model_path="Qwen/Qwen-VL-Chat"),
"monkey": partial(Monkey, model_path="echo840/Monkey"),
"monkey-chat": partial(MonkeyChat, model_path="echo840/Monkey-Chat"),
"minimonkey": partial(MiniMonkey, model_path="mx262/MiniMonkey"),
}
llava_series = {
"llava_v1.5_7b": partial(LLaVA, model_path="liuhaotian/llava-v1.5-7b"),
"llava_v1.5_13b": partial(LLaVA, model_path="liuhaotian/llava-v1.5-13b"),
"llava_v1_7b": partial(LLaVA, model_path=LLAVA_V1_7B_MODEL_PTH),
"sharegpt4v_7b": partial(LLaVA, model_path="Lin-Chen/ShareGPT4V-7B"),
"sharegpt4v_13b": partial(LLaVA, model_path="Lin-Chen/ShareGPT4V-13B"),
"llava_next_vicuna_7b": partial(
LLaVA_Next, model_path="llava-hf/llava-v1.6-vicuna-7b-hf"
),
"llava_next_vicuna_13b": partial(
LLaVA_Next, model_path="llava-hf/llava-v1.6-vicuna-13b-hf"
),
"llava_next_mistral_7b": partial(
LLaVA_Next, model_path="llava-hf/llava-v1.6-mistral-7b-hf"
),
"llava_next_yi_34b": partial(LLaVA_Next, model_path="llava-hf/llava-v1.6-34b-hf"),
"llava_next_llama3": partial(
LLaVA_Next, model_path="llava-hf/llama3-llava-next-8b-hf"
),
"llava_next_72b": partial(LLaVA_Next, model_path="llava-hf/llava-next-72b-hf"),
"llava_next_110b": partial(LLaVA_Next, model_path="llava-hf/llava-next-110b-hf"),
"llava_next_qwen_32b": partial(
LLaVA_Next2, model_path="lmms-lab/llava-next-qwen-32b"
),
"llava_next_interleave_7b": partial(
LLaVA_Next, model_path="llava-hf/llava-interleave-qwen-7b-hf"
),
"llava_next_interleave_7b_dpo": partial(
LLaVA_Next, model_path="llava-hf/llava-interleave-qwen-7b-dpo-hf"
),
"llava-onevision-qwen2-0.5b-ov-hf": partial(
LLaVA_OneVision_HF, model_path="llava-hf/llava-onevision-qwen2-0.5b-ov-hf"
),
"llava-onevision-qwen2-0.5b-si-hf": partial(
LLaVA_OneVision_HF, model_path="llava-hf/llava-onevision-qwen2-0.5b-si-hf"
),
"llava-onevision-qwen2-7b-ov-hf": partial(
LLaVA_OneVision_HF, model_path="llava-hf/llava-onevision-qwen2-7b-ov-hf"
),
"llava-onevision-qwen2-7b-si-hf": partial(
LLaVA_OneVision_HF, model_path="llava-hf/llava-onevision-qwen2-7b-si-hf"
),
"llava_onevision_qwen2_0.5b_si": partial(
LLaVA_OneVision, model_path="lmms-lab/llava-onevision-qwen2-0.5b-si"
),
"llava_onevision_qwen2_7b_si": partial(
LLaVA_OneVision, model_path="lmms-lab/llava-onevision-qwen2-7b-si"
),
"llava_onevision_qwen2_72b_si": partial(
LLaVA_OneVision, model_path="lmms-lab/llava-onevision-qwen2-72b-si"
),
"llava_onevision_qwen2_0.5b_ov": partial(
LLaVA_OneVision, model_path="lmms-lab/llava-onevision-qwen2-0.5b-ov"
),
"llava_onevision_qwen2_7b_ov": partial(
LLaVA_OneVision, model_path="lmms-lab/llava-onevision-qwen2-7b-ov"
),
"llava_onevision_qwen2_72b_ov": partial(
LLaVA_OneVision, model_path="lmms-lab/llava-onevision-qwen2-72b-ov-sft"
),
"Aquila-VL-2B": partial(LLaVA_OneVision, model_path="BAAI/Aquila-VL-2B-llava-qwen"),
"llava_video_qwen2_7b": partial(
LLaVA_OneVision, model_path="lmms-lab/LLaVA-Video-7B-Qwen2"
),
"llava_video_qwen2_72b": partial(
LLaVA_OneVision, model_path="lmms-lab/LLaVA-Video-72B-Qwen2"
),
"varco-vision-hf": partial(
LLaVA_OneVision_HF, model_path="NCSOFT/VARCO-VISION-14B-HF"
),
}
vita_series = {
"vita": partial(VITA, model_path="VITA-MLLM/VITA", root=VITA_ROOT),
"vita_qwen2": partial(VITAQwen2, model_path="VITA-MLLM/VITA-1.5", root=VITA_ROOT),
}
long_vita_series = {
"Long-VITA-16K": partial(
LongVITA, model_path="VITA-MLLM/Long-VITA-16K_HF", max_num_frame=128
),
"Long-VITA-128K": partial(
LongVITA, model_path="VITA-MLLM/Long-VITA-128K_HF", max_num_frame=256
),
"Long-VITA-1M": partial(
LongVITA, model_path="VITA-MLLM/Long-VITA-1M_HF", max_num_frame=256
),
}
internvl_series = {
"InternVL-Chat-V1-1": partial(
InternVLChat, model_path="OpenGVLab/InternVL-Chat-V1-1", version="V1.1"
),
"InternVL-Chat-V1-2": partial(
InternVLChat, model_path="OpenGVLab/InternVL-Chat-V1-2", version="V1.2"
),
"InternVL-Chat-V1-2-Plus": partial(
InternVLChat, model_path="OpenGVLab/InternVL-Chat-V1-2-Plus", version="V1.2"
),
# InternVL1.5 series
"InternVL-Chat-V1-5": partial(
InternVLChat,
model_path="OpenGVLab/InternVL-Chat-V1-5",
version="V1.5",
),
"Mini-InternVL-Chat-2B-V1-5": partial(
InternVLChat, model_path="OpenGVLab/Mini-InternVL-Chat-2B-V1-5", version="V1.5"
),
"Mini-InternVL-Chat-4B-V1-5": partial(
InternVLChat, model_path="OpenGVLab/Mini-InternVL-Chat-4B-V1-5", version="V1.5"
),
# InternVL2 series
"InternVL2-1B": partial(
InternVLChat, model_path="OpenGVLab/InternVL2-1B", version="V2.0"
),
"InternVL2-2B": partial(
InternVLChat, model_path="OpenGVLab/InternVL2-2B", version="V2.0"
),
"InternVL2-4B": partial(
InternVLChat, model_path="OpenGVLab/InternVL2-4B", version="V2.0"
),
"InternVL2-8B": partial(
InternVLChat, model_path="OpenGVLab/InternVL2-8B", version="V2.0"
),
"InternVL2-26B": partial(
InternVLChat, model_path="OpenGVLab/InternVL2-26B", version="V2.0"
),
"InternVL2-40B": partial(
InternVLChat, model_path="OpenGVLab/InternVL2-40B", version="V2.0"
),
"InternVL2-76B": partial(
InternVLChat, model_path="OpenGVLab/InternVL2-Llama3-76B", version="V2.0"
),
# InternVL2 MPO series
"InternVL2-8B-MPO": partial(
InternVLChat, model_path="OpenGVLab/InternVL2-8B-MPO", version="V2.0"
),
"InternVL2-8B-MPO-CoT": partial(
InternVLChat,
model_path="OpenGVLab/InternVL2-8B-MPO",
version="V2.0",
use_mpo_prompt=True,
),
# InternVL2.5 series
"InternVL2_5-1B": partial(
InternVLChat, model_path="OpenGVLab/InternVL2_5-1B", version="V2.0"
),
"InternVL2_5-2B": partial(
InternVLChat, model_path="OpenGVLab/InternVL2_5-2B", version="V2.0"
),
"InternVL2_5-4B": partial(
InternVLChat, model_path="OpenGVLab/InternVL2_5-4B", version="V2.0"
),
"InternVL2_5-8B": partial(
InternVLChat, model_path="OpenGVLab/InternVL2_5-8B", version="V2.0"
),
"InternVL2_5-26B": partial(
InternVLChat, model_path="OpenGVLab/InternVL2_5-26B", version="V2.0"
),
"InternVL2_5-38B": partial(
InternVLChat, model_path="OpenGVLab/InternVL2_5-38B", version="V2.0"
),
"InternVL2_5-78B": partial(
InternVLChat, model_path="OpenGVLab/InternVL2_5-78B", version="V2.0"
),
# InternVL2.5 series with Best-of-N evaluation
"InternVL2_5-8B-BoN-8": partial(
InternVLChat, model_path="OpenGVLab/InternVL2_5-8B", version="V2.0",
best_of_n=8, reward_model_path="OpenGVLab/VisualPRM-8B",
),
# InternVL2.5-MPO series
"InternVL2_5-1B-MPO": partial(
InternVLChat,
model_path="OpenGVLab/InternVL2_5-1B-MPO",
version="V2.0",
use_mpo_prompt=True,
),
"InternVL2_5-2B-MPO": partial(
InternVLChat,
model_path="OpenGVLab/InternVL2_5-2B-MPO",
version="V2.0",
use_mpo_prompt=True,
),
"InternVL2_5-4B-MPO": partial(
InternVLChat,
model_path="OpenGVLab/InternVL2_5-4B-MPO",
version="V2.0",
use_mpo_prompt=True,
),
"InternVL2_5-8B-MPO": partial(
InternVLChat,
model_path="OpenGVLab/InternVL2_5-8B-MPO",
version="V2.0",
use_mpo_prompt=True,
),
"InternVL2_5-26B-MPO": partial(
InternVLChat,
model_path="OpenGVLab/InternVL2_5-26B-MPO",
version="V2.0",
use_mpo_prompt=True,
),
"InternVL2_5-38B-MPO": partial(
InternVLChat,
model_path="OpenGVLab/InternVL2_5-38B-MPO",
version="V2.0",
use_mpo_prompt=True,
),
"InternVL2_5-78B-MPO": partial(
InternVLChat,
model_path="OpenGVLab/InternVL2_5-78B-MPO",
version="V2.0",
use_mpo_prompt=True,
),
}
sail_series = {
"SAIL-VL-2B": partial(SailVL, model_path="BytedanceDouyinContent/SAIL-VL-2B")
}
yivl_series = {
"Yi_VL_6B": partial(Yi_VL, model_path="01-ai/Yi-VL-6B", root=Yi_ROOT),
"Yi_VL_34B": partial(Yi_VL, model_path="01-ai/Yi-VL-34B", root=Yi_ROOT),
}
xcomposer_series = {
"XComposer": partial(XComposer, model_path="internlm/internlm-xcomposer-vl-7b"),
"sharecaptioner": partial(ShareCaptioner, model_path="Lin-Chen/ShareCaptioner"),
"XComposer2": partial(XComposer2, model_path="internlm/internlm-xcomposer2-vl-7b"),
"XComposer2_1.8b": partial(
XComposer2, model_path="internlm/internlm-xcomposer2-vl-1_8b"
),
"XComposer2_4KHD": partial(
XComposer2_4KHD, model_path="internlm/internlm-xcomposer2-4khd-7b"
),
"XComposer2d5": partial(
XComposer2d5, model_path="internlm/internlm-xcomposer2d5-7b"
),
}
minigpt4_series = {
"MiniGPT-4-v2": partial(MiniGPT4, mode="v2", root=MiniGPT4_ROOT),
"MiniGPT-4-v1-7B": partial(MiniGPT4, mode="v1_7b", root=MiniGPT4_ROOT),
"MiniGPT-4-v1-13B": partial(MiniGPT4, mode="v1_13b", root=MiniGPT4_ROOT),
}
idefics_series = {
"idefics_9b_instruct": partial(
IDEFICS, model_path="HuggingFaceM4/idefics-9b-instruct"
),
"idefics_80b_instruct": partial(
IDEFICS, model_path="HuggingFaceM4/idefics-80b-instruct"
),
"idefics2_8b": partial(IDEFICS2, model_path="HuggingFaceM4/idefics2-8b"),
# Idefics3 follows Idefics2 Pattern
"Idefics3-8B-Llama3": partial(
IDEFICS2, model_path="HuggingFaceM4/Idefics3-8B-Llama3"
),
}
smolvlm_series = {
"SmolVLM-256M": partial(SmolVLM, model_path="HuggingFaceTB/SmolVLM-256M-Instruct"),
"SmolVLM-500M": partial(SmolVLM, model_path="HuggingFaceTB/SmolVLM-500M-Instruct"),
"SmolVLM": partial(SmolVLM, model_path="HuggingFaceTB/SmolVLM-Instruct"),
"SmolVLM-DPO": partial(SmolVLM, model_path="HuggingFaceTB/SmolVLM-Instruct-DPO"),
"SmolVLM-Synthetic": partial(SmolVLM, model_path="HuggingFaceTB/SmolVLM-Synthetic"),
"SmolVLM2-256M": partial(
SmolVLM2, model_path="HuggingFaceTB/SmolVLM2-256M-Video-Instruct"
),
"SmolVLM2-500M": partial(
SmolVLM2, model_path="HuggingFaceTB/SmolVLM2-500M-Video-Instruct"
),
"SmolVLM2": partial(SmolVLM2, model_path="HuggingFaceTB/SmolVLM2-2.2B-Instruct"),
}
instructblip_series = {
"instructblip_7b": partial(InstructBLIP, name="instructblip_7b"),
"instructblip_13b": partial(InstructBLIP, name="instructblip_13b"),
}
deepseekvl_series = {
"deepseek_vl_7b": partial(DeepSeekVL, model_path="deepseek-ai/deepseek-vl-7b-chat"),
"deepseek_vl_1.3b": partial(
DeepSeekVL, model_path="deepseek-ai/deepseek-vl-1.3b-chat"
),
}
deepseekvl2_series = {
"deepseek_vl2_tiny": partial(
DeepSeekVL2, model_path="deepseek-ai/deepseek-vl2-tiny"
),
"deepseek_vl2_small": partial(
DeepSeekVL2, model_path="deepseek-ai/deepseek-vl2-small"
),
"deepseek_vl2": partial(DeepSeekVL2, model_path="deepseek-ai/deepseek-vl2"),
}
janus_series = {
"Janus-1.3B": partial(Janus, model_path="deepseek-ai/Janus-1.3B"),
"Janus-Pro-7B": partial(Janus, model_path="deepseek-ai/Janus-Pro-7B"),
}
cogvlm_series = {
"cogvlm-grounding-generalist": partial(
CogVlm,
model_path="THUDM/cogvlm-grounding-generalist-hf",
tokenizer_name="lmsys/vicuna-7b-v1.5",
),
"cogvlm-chat": partial(
CogVlm, model_path="THUDM/cogvlm-chat-hf", tokenizer_name="lmsys/vicuna-7b-v1.5"
),
"cogvlm2-llama3-chat-19B": partial(
CogVlm, model_path="THUDM/cogvlm2-llama3-chat-19B"
),
"glm-4v-9b": partial(GLM4v, model_path="THUDM/glm-4v-9b"),
}
wemm_series = {
"WeMM": partial(WeMM, model_path="feipengma/WeMM"),
}
cambrian_series = {
"cambrian_8b": partial(Cambrian, model_path="nyu-visionx/cambrian-8b"),
"cambrian_13b": partial(Cambrian, model_path="nyu-visionx/cambrian-13b"),
"cambrian_34b": partial(Cambrian, model_path="nyu-visionx/cambrian-34b"),
}
chameleon_series = {
"chameleon_7b": partial(Chameleon, model_path="facebook/chameleon-7b"),
"chameleon_30b": partial(Chameleon, model_path="facebook/chameleon-30b"),
}
vila_series = {
"VILA1.5-3b": partial(VILA, model_path="Efficient-Large-Model/VILA1.5-3b"),
"Llama-3-VILA1.5-8b": partial(
VILA, model_path="Efficient-Large-Model/Llama-3-VILA1.5-8b"
),
"VILA1.5-13b": partial(VILA, model_path="Efficient-Large-Model/VILA1.5-13b"),
"VILA1.5-40b": partial(VILA, model_path="Efficient-Large-Model/VILA1.5-40b"),
}
ovis_series = {
"Ovis1.5-Llama3-8B": partial(Ovis, model_path="AIDC-AI/Ovis1.5-Llama3-8B"),
"Ovis1.5-Gemma2-9B": partial(Ovis, model_path="AIDC-AI/Ovis1.5-Gemma2-9B"),
"Ovis1.6-Gemma2-9B": partial(Ovis1_6, model_path="AIDC-AI/Ovis1.6-Gemma2-9B"),
"Ovis1.6-Llama3.2-3B": partial(Ovis1_6, model_path="AIDC-AI/Ovis1.6-Llama3.2-3B"),
"Ovis1.6-Gemma2-27B": partial(
Ovis1_6_Plus, model_path="AIDC-AI/Ovis1.6-Gemma2-27B"
),
"Ovis2-1B": partial(Ovis2, model_path="AIDC-AI/Ovis2-1B"),
"Ovis2-2B": partial(Ovis2, model_path="AIDC-AI/Ovis2-2B"),
"Ovis2-4B": partial(Ovis2, model_path="AIDC-AI/Ovis2-4B"),
"Ovis2-8B": partial(Ovis2, model_path="AIDC-AI/Ovis2-8B"),
"Ovis2-16B": partial(Ovis2, model_path="AIDC-AI/Ovis2-16B"),
"Ovis2-34B": partial(Ovis2, model_path="AIDC-AI/Ovis2-34B"),
}
mantis_series = {
"Mantis-8B-siglip-llama3": partial(
Mantis, model_path="TIGER-Lab/Mantis-8B-siglip-llama3"
),
"Mantis-8B-clip-llama3": partial(
Mantis, model_path="TIGER-Lab/Mantis-8B-clip-llama3"
),
"Mantis-8B-Idefics2": partial(Mantis, model_path="TIGER-Lab/Mantis-8B-Idefics2"),
"Mantis-8B-Fuyu": partial(Mantis, model_path="TIGER-Lab/Mantis-8B-Fuyu"),
}
phi3_series = {
"Phi-3-Vision": partial(
Phi3Vision, model_path="microsoft/Phi-3-vision-128k-instruct"
),
"Phi-3.5-Vision": partial(
Phi3_5Vision, model_path="microsoft/Phi-3.5-vision-instruct"
),
}
phi4_series = {
'Phi-4-Vision': partial(Phi4Multimodal, model_path='microsoft/Phi-4-multimodal-instruct'),
}
xgen_mm_series = {
"xgen-mm-phi3-interleave-r-v1.5": partial(
XGenMM, model_path="Salesforce/xgen-mm-phi3-mini-instruct-interleave-r-v1.5"
),
"xgen-mm-phi3-dpo-r-v1.5": partial(
XGenMM, model_path="Salesforce/xgen-mm-phi3-mini-instruct-dpo-r-v1.5"
),
}
qwen2vl_series = {
"Qwen2-VL-7B-560": partial(
Qwen2VLChat,
model_path="/root/xuqixin/reinforce_perception/model/checkpoint-560",
min_pixels=3136,
max_pixels=12845056,
),
"Qwen2-VL-7B-mix": partial(
Qwen2VLChat,
model_path="/root/kongly/project/Qwen-SFT/saves/qwen2_vl-7b/full/sft",
min_pixels=3136,
max_pixels=12845056,
use_reasoning_prompt=1,
use_vllm=True,
do_sample=True,
max_new_tokens=2048,
top_p=1.0,
top_k=-1,
temperature=1.0,
),
"Qwen2-VL-7B-box-mix": partial(
Qwen2VLChat,
model_path="/root/kongly/project/Qwen-SFT/saves/qwen2_vl-7b/full/sft-box",
min_pixels=3136,
max_pixels=12845056,
use_reasoning_prompt=1,
use_vllm=True,
do_sample=True,
max_new_tokens=2048,
top_p=1.0,
top_k=-1,
temperature=1.0,
),
"Qwen2-VL-7B-GRPO-2000": partial(
Qwen2VLChat,
model_path="/root/xuqixin/reinforce_perception/model/r1_distilled_grpo_step2500",
min_pixels=3136,
max_pixels=12845056,
use_reasoning_prompt=0,
use_vllm=True,
do_sample=True,
max_new_tokens=2048,
top_p=1.0,
top_k=-1,
temperature=1.0,
),
"Qwen2-VL-7B-GRPO-2000-Prompt": partial(
Qwen2VLChat,
model_path="/root/xuqixin/reinforce_perception/model/r1_distilled_grpo_step2500",
min_pixels=3136,
max_pixels=12845056,
use_reasoning_prompt=2,
use_vllm=True,
do_sample=True,
max_new_tokens=2048,
top_p=1.0,
top_k=-1,
temperature=1.0,
system_prompt="You are a helpful assistant.",
),
"Qwen2-VL-7B-GRPO-2000-HF": partial(
Qwen2VLChat,
model_path="/root/xuqixin/reinforce_perception/model/r1_distilled_grpo_step2500",
min_pixels=3136,
max_pixels=12845056,
use_reasoning_prompt=1,
use_vllm=False,
do_sample=True,
max_new_tokens=2048,
top_p=1.0,
top_k=100,
temperature=1.0,
),
"Qwen-VL-Max-0809": partial(
Qwen2VLAPI,
model="qwen-vl-max-0809",
min_pixels=1280 * 28 * 28,
max_pixels=16384 * 28 * 28,
),
"Qwen-VL-Plus-0809": partial(
Qwen2VLAPI,
model="qwen-vl-plus-0809",
min_pixels=1280 * 28 * 28,
max_pixels=16384 * 28 * 28,
),
"QVQ-72B-Preview": partial(
Qwen2VLChat,
model_path="Qwen/QVQ-72B-Preview",
min_pixels=1280 * 28 * 28,
max_pixels=16384 * 28 * 28,
system_prompt="You are a helpful and harmless assistant. You are Qwen developed by Alibaba. You should think step-by-step.",
max_new_tokens=8192,
post_process=False,
),
"Qwen2-VL-72B-Instruct": partial(
Qwen2VLChat,
model_path="Qwen/Qwen2-VL-72B-Instruct",
min_pixels=1280 * 28 * 28,
max_pixels=16384 * 28 * 28,
),
"Qwen2-VL-7B-Instruct": partial(
Qwen2VLChat,
model_path="Qwen/Qwen2-VL-7B-Instruct",
min_pixels=1280 * 28 * 28,
max_pixels=16384 * 28 * 28,
),
"Qwen2-VL-7B-Instruct-AWQ": partial(
Qwen2VLChat,
model_path="Qwen/Qwen2-VL-7B-Instruct-AWQ",
min_pixels=1280 * 28 * 28,
max_pixels=16384 * 28 * 28,
),
"Qwen2-VL-7B-Instruct-GPTQ-Int4": partial(
Qwen2VLChat,
model_path="Qwen/Qwen2-VL-7B-Instruct-GPTQ-Int4",
min_pixels=1280 * 28 * 28,
max_pixels=16384 * 28 * 28,
),
"Qwen2-VL-7B-Instruct-GPTQ-Int8": partial(
Qwen2VLChat,
model_path="Qwen/Qwen2-VL-7B-Instruct-GPTQ-Int8",
min_pixels=1280 * 28 * 28,
max_pixels=16384 * 28 * 28,
),
"Qwen2-VL-2B-Instruct": partial(
Qwen2VLChat,
model_path="Qwen/Qwen2-VL-2B-Instruct",
min_pixels=1280 * 28 * 28,
max_pixels=16384 * 28 * 28,
),
"Qwen2-VL-2B-Instruct-AWQ": partial(
Qwen2VLChat,
model_path="Qwen/Qwen2-VL-2B-Instruct-AWQ",
min_pixels=1280 * 28 * 28,
max_pixels=16384 * 28 * 28,
),
"Qwen2-VL-2B-Instruct-GPTQ-Int4": partial(
Qwen2VLChat,
model_path="Qwen/Qwen2-VL-2B-Instruct-GPTQ-Int4",
min_pixels=1280 * 28 * 28,
max_pixels=16384 * 28 * 28,
),
"Qwen2-VL-2B-Instruct-GPTQ-Int8": partial(
Qwen2VLChat,
model_path="Qwen/Qwen2-VL-2B-Instruct-GPTQ-Int8",
min_pixels=1280 * 28 * 28,
max_pixels=16384 * 28 * 28,
),
"XinYuan-VL-2B-Instruct": partial(
Qwen2VLChat,
model_path="Cylingo/Xinyuan-VL-2B",
min_pixels=1280 * 28 * 28,
max_pixels=16384 * 28 * 28,
),
"Qwen2.5-VL-3B-Instruct": partial(
Qwen2VLChat,
model_path="Qwen/Qwen2.5-VL-3B-Instruct",
min_pixels=1280 * 28 * 28,
max_pixels=16384 * 28 * 28,
use_custom_prompt=False,
),
"Qwen2.5-VL-3B-Instruct-AWQ": partial(
Qwen2VLChat,
model_path="Qwen/Qwen2.5-VL-3B-Instruct-AWQ",
min_pixels=1280 * 28 * 28,
max_pixels=16384 * 28 * 28,
use_custom_prompt=False,
),
"Qwen2.5-VL-7B-Instruct-original": partial(
Qwen2VLChat,
model_path="/user/yaoshu/models/Qwen_25_VL",
min_pixels=1280 * 28 * 28,
max_pixels=16384 * 28 * 28,
use_custom_prompt=False,
use_vllm=True,
),
"Qwen2.5-VL-7B-RL": partial(
Qwen2VLChat,
model_path="/user/xuqixin/checkpoints/qwen25_vl-7b/grpo_v7_exp0_qwen25vl_grpo_opensource_math_onlinefilter_regen/global_step_300/actor/huggingface",
min_pixels=1280 * 28 * 28,
max_pixels=16384 * 28 * 28,
use_custom_prompt=False,
use_vllm=True,
use_reasoning_prompt=2,
do_sample=True,
temperature=1,
top_p=1,
top_k=-1,
),
"Qwen2.5-VL-7B-RL-greedy": partial(
Qwen2VLChat,
model_path="/user/xuqixin/checkpoints/qwen25_vl-7b/grpo_v7_exp0_qwen25vl_grpo_opensource_math_onlinefilter_regen/global_step_300/actor/huggingface",
min_pixels=1280 * 28 * 28,
max_pixels=16384 * 28 * 28,
use_custom_prompt=False,
use_vllm=True,
use_reasoning_prompt=2,
do_sample=True
),
"Qwen2.5-VL-7B-RL-raw-greedy": partial(
Qwen2VLChat,
model_path="/user/xuqixin/checkpoints/qwen25_vl-7b/grpo_v7_exp0_qwen25vl_grpo_opensource_math_onlinefilter_regen/global_step_300/actor/huggingface",
min_pixels=1280 * 28 * 28,
max_pixels=16384 * 28 * 28,
use_custom_prompt=False,
use_vllm=True,
do_sample=True
),
"Qwen2.5-VL-7B-Instruct-AWQ": partial(
Qwen2VLChat,
model_path="Qwen/Qwen2.5-VL-7B-Instruct-AWQ",
min_pixels=1280 * 28 * 28,
max_pixels=16384 * 28 * 28,
use_custom_prompt=False,
),
"Qwen2.5-VL-72B-Instruct": partial(
Qwen2VLChat,
model_path="Qwen/Qwen2.5-VL-72B-Instruct",
min_pixels=1280 * 28 * 28,
max_pixels=16384 * 28 * 28,
use_custom_prompt=False,
),
"Qwen2.5-VL-72B-Instruct-AWQ": partial(
Qwen2VLChat,
model_path="Qwen/Qwen2.5-VL-72B-Instruct-AWQ",
min_pixels=1280 * 28 * 28,
max_pixels=16384 * 28 * 28,
use_custom_prompt=False,
),
"VLM-R1": partial(
VLMR1Chat,
model_path="omlab/VLM-R1-Qwen2.5VL-3B-Math-0305",
min_pixels=1280 * 28 * 28,
max_pixels=16384 * 28 * 28,
use_custom_prompt=False,
),
}
slime_series = {
"Slime-7B": partial(SliME, model_path="yifanzhang114/SliME-vicuna-7B"),
"Slime-8B": partial(SliME, model_path="yifanzhang114/SliME-Llama3-8B"),
"Slime-13B": partial(SliME, model_path="yifanzhang114/SliME-vicuna-13B"),
}
eagle_series = {
"Eagle-X4-8B-Plus": partial(Eagle, model_path="NVEagle/Eagle-X4-8B-Plus"),
"Eagle-X4-13B-Plus": partial(Eagle, model_path="NVEagle/Eagle-X4-13B-Plus"),
"Eagle-X5-7B": partial(Eagle, model_path="NVEagle/Eagle-X5-7B"),
"Eagle-X5-13B": partial(Eagle, model_path="NVEagle/Eagle-X5-13B"),
"Eagle-X5-13B-Chat": partial(Eagle, model_path="NVEagle/Eagle-X5-13B-Chat"),
"Eagle-X5-34B-Chat": partial(Eagle, model_path="NVEagle/Eagle-X5-34B-Chat"),
"Eagle-X5-34B-Plus": partial(Eagle, model_path="NVEagle/Eagle-X5-34B-Plus"),
}
moondream_series = {
"Moondream1": partial(Moondream1, model_path="vikhyatk/moondream1"),
"Moondream2": partial(Moondream2, model_path="vikhyatk/moondream2"),
}
llama_series = {
"Llama-3.2-11B-Vision-Instruct": partial(
llama_vision, model_path="meta-llama/Llama-3.2-11B-Vision-Instruct"
),
"LLaVA-CoT": partial(llama_vision, model_path="Xkev/Llama-3.2V-11B-cot"),
"Llama-3.2-90B-Vision-Instruct": partial(
llama_vision, model_path="meta-llama/Llama-3.2-90B-Vision-Instruct"
),
}
molmo_series = {
"molmoE-1B-0924": partial(molmo, model_path="allenai/MolmoE-1B-0924"),
"molmo-7B-D-0924": partial(molmo, model_path="allenai/Molmo-7B-D-0924"),
"molmo-7B-O-0924": partial(molmo, model_path="allenai/Molmo-7B-O-0924"),
"molmo-72B-0924": partial(molmo, model_path="allenai/Molmo-72B-0924"),
}
kosmos_series = {
"Kosmos2": partial(Kosmos2, model_path="microsoft/kosmos-2-patch14-224")
}
points_series = {
"POINTS-Yi-1.5-9B-Chat": partial(
POINTS, model_path="WePOINTS/POINTS-Yi-1-5-9B-Chat"
),
"POINTS-Qwen-2.5-7B-Chat": partial(
POINTS, model_path="WePOINTS/POINTS-Qwen-2-5-7B-Chat"
),
"POINTSV15-Qwen-2.5-7B-Chat": partial(
POINTSV15, model_path="WePOINTS/POINTS-1-5-Qwen-2-5-7B-Chat"
),
}
nvlm_series = {
"NVLM": partial(NVLM, model_path="nvidia/NVLM-D-72B"),
}
vintern_series = {
"Vintern-3B-beta": partial(VinternChat, model_path="5CD-AI/Vintern-3B-beta"),
"Vintern-1B-v2": partial(VinternChat, model_path="5CD-AI/Vintern-1B-v2"),
}
aria_series = {"Aria": partial(Aria, model_path="rhymes-ai/Aria")}
h2ovl_series = {
"h2ovl-mississippi-2b": partial(H2OVLChat, model_path="h2oai/h2ovl-mississippi-2b"),
"h2ovl-mississippi-1b": partial(
H2OVLChat, model_path="h2oai/h2ovl-mississippi-800m"
),
}
valley_series = {
"valley_eagle": partial(
ValleyEagleChat, model_path="bytedance-research/Valley-Eagle-7B"
),
}
ola_series = {
"ola": partial(Ola, model_path="THUdyh/Ola-7b"),
}
ross_series = {
"ross-qwen2-7b": partial(Ross, model_path="HaochenWang/ross-qwen2-7b"),
}
ursa_series = {"URSA-8B": partial(UrsaChat, model_path="URSA-MATH/URSA-8B")}
gemma_series = {
"paligemma-3b-mix-448": partial(
PaliGemma, model_path="google/paligemma-3b-mix-448"
),
'Gemma3-4B': partial(Gemma3, model_path='google/gemma-3-4b-it'),
'Gemma3-12B': partial(Gemma3, model_path='google/gemma-3-12b-it'),
'Gemma3-27B': partial(Gemma3, model_path='google/gemma-3-27b-it')
}
supported_VLM = {}
model_groups = [
ungrouped,
o1_apis,
api_models,
xtuner_series,
qwen_series,
llava_series,
internvl_series,
yivl_series,
xcomposer_series,
minigpt4_series,
idefics_series,
instructblip_series,
deepseekvl_series,
deepseekvl2_series,
janus_series,
minicpm_series,
cogvlm_series,
wemm_series,
cambrian_series,
chameleon_series,
video_models,
ovis_series,
vila_series,
mantis_series,
mmalaya_series,
phi3_series,
phi4_series,
xgen_mm_series,
qwen2vl_series,
slime_series,
eagle_series,
moondream_series,
llama_series,
molmo_series,
kosmos_series,
points_series,
nvlm_series,
vintern_series,
h2ovl_series,
aria_series,
smolvlm_series,
sail_series,
valley_series,
vita_series,
ross_series,
emu_series,
ola_series,
ursa_series,
gemma_series,
long_vita_series,
]
for grp in model_groups:
supported_VLM.update(grp)