[182] Calibrated Self-Rewarding Vision Language Models

2024λ…„ 10μ›” 10일 Β· 1 λΆ„ Β· long8v Β· 

[178] RLAIF-V: Aligning MLLMs through Open-Source AI Feedback for Super GPT-4V Trustworthiness

2024λ…„ 9μ›” 23일 Β· 2 λΆ„ Β· long8v Β· 

[172] RLHF-V: Towards Trustworthy MLLMs via Behavior Alignment from Fine-grained Correctional Human Feedback

2024λ…„ 8μ›” 30일 Β· 2 λΆ„ Β· long8v Β· 

[170] Unpacking DPO and PPO: Disentangling Best Practices for Learning from Preference Feedback

2024λ…„ 8μ›” 27일 Β· 2 λΆ„ Β· long8v Β· 

[160] ALOHa: A New Measure for Hallucination in Captioning Models

2024λ…„ 6μ›” 15일 Β· 2 λΆ„ Β· long8v Β· 

[157] LeGrad: An Explainability Method for Vision Transformers via Feature Formation Sensitivity

2024λ…„ 5μ›” 6일 Β· 2 λΆ„ Β· long8v Β· 

feat: add LeGrad

2024λ…„ 5μ›” 6일 Β· 1 λΆ„ Β· long8v Β· 

[155] Revisiting Text-to-Image Evaluation with Gecko: On Metrics, Prompts, and Human Ratings

2024λ…„ 5μ›” 3일 Β· 1 λΆ„ Β· long8v Β· 

[154] Mismatch Quest: Visual and Textual Feedback for Image-Text Misalignment

2024λ…„ 4μ›” 3일 Β· 2 λΆ„ Β· long8v Β·