JAM
Paper: Jointly Training Large Autoregressive Multimodal Models GitHub Link: None Publisher: Arxiv Author Affiliation: Meta AI Functional Division Understanding Generatio...
Paper: Jointly Training Large Autoregressive Multimodal Models GitHub Link: None Publisher: Arxiv Author Affiliation: Meta AI Functional Division Understanding Generatio...
Paper: AnyMAL: An Efficient and Scalable Any-Modality Augmented Language Model GitHub Link: None Author Affiliation: FAIR, Meta & Meta Reality Labs Functional Division Under...
Paper: Q-Bench: A Benchmark for General-Purpose Foundation Models on Low-level Vision Project Link Publisher: Arxiv Author Affiliation: Nanyang Technological University
Paper: Aligning Large Multimodal Models with Factually Augmented RLHF GitHub Link Publisher: Arxiv Author Affiliation: UC Berkeley Type SFT RLHF Multi-turn ...
Paper: Kosmos-2.5: A Multimodal Literate Model GitHub Link Publisher: Arxiv Author Affiliation: Microsoft Functional Division Understanding Generation Design D...
Paper: DreamLLM: Synergistic Multimodal Comprehension and Creation GitHub Link Publisher: ICLR 2024 Author Affiliation: MEGVII Functional Division Understanding Generati...
Paper: NExT-GPT: Any-to-Any Multimodal LLM GitHub Link Publisher: ICLR 2024 Author Affiliation: National University of Singapore Type SFT RLHF Multi-turn ...
Paper: NExT-GPT: Any-to-Any Multimodal LLM GitHub Link Publisher: ICLR 2024 Author Affiliation: National University of Singapore Functional Division Understanding Genera...
Paper: NExT-GPT: Any-to-Any Multimodal LLM GitHub Link Publisher: ICLR 2024 Author Affiliation: National University of Singapore Type SFT RLHF Multi-turn ...
Paper: Unified Language-Vision Pretraining in LLM with Dynamic Discrete Visual Tokenization GitHub Link Publisher: ICLR 2024 Author Affiliation: Kuaishou Technology Functional Division ...