FlashSVD: Memory-Efficient Inference with Streaming for Low-Rank Models Paper • 2508.01506 • Published Aug 2 • 1
KVCOMM: Online Cross-context KV-cache Communication for Efficient LLM-based Multi-agent Systems Paper • 2510.12872 • Published 29 days ago • 2
Vision-Zero: Scalable VLM Self-Improvement via Strategic Gamified Self-Play Paper • 2509.25541 • Published Sep 29 • 139
MME-Reasoning: A Comprehensive Benchmark for Logical Reasoning in MLLMs Paper • 2505.21327 • Published May 27 • 83
Performance-aware Approximation of Global Channel Pruning for Multitask CNNs Paper • 2303.11923 • Published Mar 21, 2023 • 1
Training-Free Adaptive Diffusion with Bounded Difference Approximation Strategy Paper • 2410.09873 • Published Oct 13, 2024 • 3