SPINQUANT - LLM QUANTIZATION WITH LEARNED ROTATIONS

November 6, 2025

ViSpec - Accelerating Vision-Language Models with Vision-Aware Speculative Decoding

November 5, 2025

SpecEdge: Scalable Edge-Assisted Serving Framework for Interactive LLMs

November 3, 2025

Ada-KV: Optimizing KV Cache Eviction by Adaptive Budget Allocation for Efficient LLM Inference

November 1, 2025

NOT ALL HEADS MATTER: A HEAD-LEVEL KV CACHE COMPRESSION METHOD WITH INTEGRATED RETRIEVAL AND REASONING

October 31, 2025

QSVD: Efficient Low-rank Approximation for Unified Query-Key-Value Weight Compression in Low-Precision Vision-Language Models

October 30, 2025

SPECVLM: Enhancing Speculative Decoding of Video LLMs via Verifier-Guided Token Pruning

October 29, 2025

AdaSPEC: Selective Knowledge Distillation for Efficient Speculative Decoders

October 27, 2025

R-KV: Redundancy-aware KV Cache Compression for Reasoning Models

October 25, 2025

Beyond Attention or Similarity: Maximizing Conditional Diversity for Token Pruning in MLLMs

October 24, 2025