LLMs

KnapSpec: Self-Speculative Decoding via Adaptive Layer Selection as a Knapsack Problem

A self-speculative decoding method with training-free, knapsack-based adaptive layer selection that accounts for context-dependent attention overhead

PPA-Plan: Proactive Pitfall Avoidance for Reliable Planning in Long-Context LLM Reasoning

A proactive planning strategy for long-context reasoning that focuses on preventing logical pitfalls and false assumptions before plan generation - ___[ACL 2026](https://2026.aclweb.org)___

Detecting Training Data of Large Language Models via Expectation Maximization

Novel membership inference attack framework for large language models via an expectation maximization algorithm - ___[EACL 2026](https://2026.eacl.org)___

AcuRank: Uncertainty-Aware Adaptive Computation for Listwise Reranking

Listwise reranking with uncertainty-aware adaptive computation via Bayesian modeling of documents' relevance with TrueSkill - ___[NeurIPS 2025](https://neurips.cc)___

ScholarBench: A Bilingual Benchmark for Abstraction, Comprehension, and Reasoning Evaluation in Academic Contexts

A bilingual benchmark for abstraction, comprehension, and reasoning evaluation in academic contexts - ___[Findings of EMNLP 2025](https://2025.emnlp.org/)___

MacRAG: Compress, Slice, and Scale-up for Multi-Scale Adaptive Context RAG

Multi-scale adaptive context RAG for long-context large language models