QeRL: Beyond Efficiency -- Quantization-enhanced Reinforcement Learning for LLMs Paper • 2510.11696 • Published Oct 13 • 174
QWHA: Quantization-Aware Walsh-Hadamard Adaptation for Parameter-Efficient Fine-Tuning on Large Language Models Paper • 2509.17428 • Published Sep 22 • 9
EpiCache: Episodic KV Cache Management for Long Conversational Question Answering Paper • 2509.17396 • Published Sep 22 • 19
Interleaved Reasoning for Large Language Models via Reinforcement Learning Paper • 2505.19640 • Published May 26 • 14
EpiCache: Episodic KV Cache Management for Long Conversational Question Answering Paper • 2509.17396 • Published Sep 22 • 19
EpiCache: Episodic KV Cache Management for Long Conversational Question Answering Paper • 2509.17396 • Published Sep 22 • 19 • 4
KVzip: Query-Agnostic KV Cache Compression with Context Reconstruction Paper • 2505.23416 • Published May 29 • 11
RILQ: Rank-Insensitive LoRA-based Quantization Error Compensation for Boosting 2-bit Large Language Model Accuracy Paper • 2412.01129 • Published Dec 2, 2024
InfiniPot-V: Memory-Constrained KV Cache Compression for Streaming Video Understanding Paper • 2506.15745 • Published Jun 18 • 13
InfiniPot-V: Memory-Constrained KV Cache Compression for Streaming Video Understanding Paper • 2506.15745 • Published Jun 18 • 13
InfiniPot-V: Memory-Constrained KV Cache Compression for Streaming Video Understanding Paper • 2506.15745 • Published Jun 18 • 13 • 2
Enhancing Computation Efficiency in Large Language Models through Weight and Activation Quantization Paper • 2311.05161 • Published Nov 9, 2023 • 1
Improving Conversational Abilities of Quantized Large Language Models via Direct Preference Alignment Paper • 2407.03051 • Published Jul 3, 2024
InfiniPot: Infinite Context Processing on Memory-Constrained LLMs Paper • 2410.01518 • Published Oct 2, 2024 • 4
InfiniPot: Infinite Context Processing on Memory-Constrained LLMs Paper • 2410.01518 • Published Oct 2, 2024 • 4 • 2
A Controlled Study on Long Context Extension and Generalization in LLMs Paper • 2409.12181 • Published Sep 18, 2024 • 45
Characterizing Prompt Compression Methods for Long Context Inference Paper • 2407.08892 • Published Jul 11, 2024 • 11
LazyLLM: Dynamic Token Pruning for Efficient Long Context LLM Inference Paper • 2407.14057 • Published Jul 19, 2024 • 46