PulseAugur
LIVE 07:16:24
ENTITY Video-LLMs

Video-LLMs

PulseAugur coverage of Video-LLMs — every cluster mentioning Video-LLMs across labs, papers, and developer communities, ranked by signal.

Total · 30d
5
5 over 90d
Releases · 30d
0
0 over 90d
Papers · 30d
5
5 over 90d
TIER MIX · 90D
SENTIMENT · 30D

1 day(s) with sentiment data

RECENT · PAGE 1/1 · 5 TOTAL
  1. TOOL · CL_25592 ·

    Video-LLMs struggle with temporal information flow, researchers find

    Researchers have identified a significant bottleneck in how Video Large Language Models (Video-LLMs) process temporal information, hindering their ability to understand the direction of video playback. While video-centr…

  2. RESEARCH · CL_20298 ·

    VTAgent improves Video TextVQA by anchoring keyframes, setting new benchmarks

    Researchers have introduced VTAgent, a novel framework designed to improve video text-based visual question answering (Video TextVQA). The system addresses limitations in current Video-LLMs by focusing on the crucial ta…

  3. RESEARCH · CL_20327 ·

    New research grounds Video-LLMs in physical reality with adversarial curriculum

    A new research paper introduces the Unified Attribution Theory, suggesting that Video-LLMs' struggles with physical reasoning stem from "Semantic Prior Dominance" rather than perceptual issues. To address this, the pape…

  4. RESEARCH · CL_11776 ·

    Researchers benchmark sycophancy in Video-LLMs with new VISE evaluation tool

    Researchers have introduced VISE, the first benchmark designed to evaluate sycophantic behavior in video large language models (Video-LLMs). Sycophancy, where models align with user input despite contradicting visual ev…

  5. RESEARCH · CL_06546 ·

    EMCompress introduces novel compression for Video-LLMs, improving efficiency

    Researchers have introduced EMCompress, a novel method for improving the efficiency of Video-LLMs in long-video reasoning tasks. This approach uses a cognitively-inspired technique called Endomorphic Multimodal Compress…