BenchLMM
Paper: BenchLMM: Benchmarking Cross-style Visual Capability of Large Multimodal Models Project Link Publisher: Arxiv Author Affiliation: Nanyang Technological University
Paper: BenchLMM: Benchmarking Cross-style Visual Capability of Large Multimodal Models Project Link Publisher: Arxiv Author Affiliation: Nanyang Technological University
Paper: PixelLM: Pixel Reasoning with Large Multimodal Model GitHub Link Publisher: Arxiv Author Affiliation: Beijing Jiaotong University Functional Division Understanding ...
Paper: RLHF-V: Towards Trustworthy MLLMs via Behavior Alignment from Fine-grained Correctional Human Feedback GitHub Link Publisher: Arxiv Author Affiliation: Tsinghua University Functio...
Paper: RLHF-V: Towards Trustworthy MLLMs via Behavior Alignment from Fine-grained Correctional Human Feedback GitHub Link Publisher: Arxiv Author Affiliation: Tsinghua University Type ...
Paper: Dolphins: Multimodal Language Model for Driving GitHub Link Publisher: Arxiv Author Affiliation: University of Wisconsin-Madison Functional Division Understanding ...
Paper: mPLUG-PaperOwl: Scientific Diagram Analysis with the Multimodal Large Language Model GitHub Link Publisher: Arxiv Author Affiliation: Alibaba Group Functional Division ...
Paper: X-InstructBLIP: A Framework for aligning X-Modal instruction-aware representations to LLMs and Emergent Cross-modal Reasoning GitHub Link Publisher: Arxiv Author Affiliation: Univer...
Paper: X-InstructBLIP: A Framework for aligning X-Modal instruction-aware representations to LLMs and Emergent Cross-modal Reasoning GitHub Link Publisher: Arxiv Author Affiliation: Univer...
Paper: CoDi-2: In-Context, Interleaved, and Interactive Any-to-Any Generation GitHub Link Publisher: Arxiv Author Affiliation: UC Berkeley Functional Division Understanding ...
Paper: VIM: Probing Multimodal Large Language Models for Visual Embedded Instruction Following GitHub Link Publisher: Arxiv Author Affiliation: University of California, Santa Barbara Fu...