[182] Calibrated Self-Rewarding Vision Language Models

2024๋…„ 10์›” 10์ผ ยท 1 ๋ถ„ ยท long8v ยท 

[178] RLAIF-V: Aligning MLLMs through Open-Source AI Feedback for Super GPT-4V Trustworthiness

2024๋…„ 9์›” 23์ผ ยท 2 ๋ถ„ ยท long8v ยท 

[172] RLHF-V: Towards Trustworthy MLLMs via Behavior Alignment from Fine-grained Correctional Human Feedback

2024๋…„ 8์›” 30์ผ ยท 2 ๋ถ„ ยท long8v ยท 

[170] Unpacking DPO and PPO: Disentangling Best Practices for Learning from Preference Feedback

2024๋…„ 8์›” 27์ผ ยท 2 ๋ถ„ ยท long8v ยท 

[160] ALOHa: A New Measure for Hallucination in Captioning Models

2024๋…„ 6์›” 15์ผ ยท 2 ๋ถ„ ยท long8v ยท 

[157] LeGrad: An Explainability Method for Vision Transformers via Feature Formation Sensitivity

2024๋…„ 5์›” 6์ผ ยท 2 ๋ถ„ ยท long8v ยท 

feat: add LeGrad

2024๋…„ 5์›” 6์ผ ยท 1 ๋ถ„ ยท long8v ยท 

[155] Revisiting Text-to-Image Evaluation with Gecko: On Metrics, Prompts, and Human Ratings

2024๋…„ 5์›” 3์ผ ยท 1 ๋ถ„ ยท long8v ยท 

[154] Mismatch Quest: Visual and Textual Feedback for Image-Text Misalignment

2024๋…„ 4์›” 3์ผ ยท 2 ๋ถ„ ยท long8v ยท