Generate, but Verify: Reducing Hallucination in Vision-Language Models with Retrospective Resampling Paper • 2504.13169 • Published Apr 17 • 39
Puzzled by Puzzles: When Vision-Language Models Can't Take a Hint Paper • 2505.23759 • Published May 29 • 5
Learning Adaptive Parallel Reasoning with Language Models Paper • 2504.15466 • Published Apr 21 • 44
Describe Anything: Detailed Localized Image and Video Captioning Paper • 2504.16072 • Published Apr 22 • 63
Atlas: Multi-Scale Attention Improves Long Context Image Modeling Paper • 2503.12355 • Published Mar 16 • 12
Diffusion Hyperfeatures: Searching Through Time and Space for Semantic Correspondence Paper • 2305.14334 • Published May 23, 2023 • 1
Readout Guidance: Learning Control from Diffusion Features Paper • 2312.02150 • Published Dec 4, 2023 • 3
Multi-Stage Multi-Modal Pre-Training for Automatic Speech Recognition Paper • 2403.19822 • Published Mar 28, 2024
ALOHa: A New Measure for Hallucination in Captioning Models Paper • 2404.02904 • Published Apr 3, 2024
Virtual Personas for Language Models via an Anthology of Backstories Paper • 2407.06576 • Published Jul 9, 2024 • 1
Visual Haystacks: Answering Harder Questions About Sets of Images Paper • 2407.13766 • Published Jul 18, 2024 • 2
view post Post 619 🚨 Launching The Visual Haystacks (VHs) Benchmark: the first "visual-centric" Needle-In-A-Haystack (NIAH) benchmark to assess LMMs' capability in long-context visual retrieval and reasoning. Check it out! tsunghanwu/visual_haystackshttps://visual-haystacks.github.io/https://arxiv.org/abs/2407.13766https://github.com/visual-haystacks/vhs_benchmark 🔥 1 1 + Reply
Rethinking Patch Dependence for Masked Autoencoders Paper • 2401.14391 • Published Jan 25, 2024 • 26
Task Oriented Dialogue as a Catalyst for Self-Supervised Automatic Speech Recognition Paper • 2401.02417 • Published Jan 4, 2024 • 1