Authors: Zhiyuan Zhao, Bin Wang, Linke Ouyang, Xiaoyi Dong, Jiaqi Wang, Conghui He
Published on: November 28, 2023
Impact Score: 8.22
Arxiv code: Arxiv:2311.16839
Summary
- What is new: Introduction of Hallucination-Aware Direct Preference Optimization (HA-DPO) to address models generating inaccurate textual descriptions from images.
- Why this is important: Common issue in multimodal large language models known as the hallucination problem.
- What the research proposes: HA-DPO reframes the hallucination problem as a preference selection task, using a pipeline to create high-quality sample pairs for robust preference learning.
- Results: Significant reduction in hallucination issues and improved generalization capabilities in models, with notable improvements in MiniGPT-4 model’s POPE accuracy and MME score.
Technical Details
Technological frameworks used: HA-DPO
Models used: Multimodal models including MiniGPT-4
Data used: Constructed positive and negative sample pairs
Potential Impact
Tech companies developing or utilizing AI-driven content creation and analysis tools, especially those involved in image-text integration services.
Want to implement this idea in a business?
We have generated a startup concept here: ClarityAI.
Leave a Reply