-
Ranking Reasoning LLMs under Test-Time Scaling ACL 2026 Main
-
Quantize What Counts: More for Keys, Less for Values ACL 2026 Findings
-
Donβt Pass@k: A Bayesian Framework for Large Language Model Evaluation ICLR 2026
-
70% Size, 100% Accuracy: Lossless LLM Compression for Efficient GPU Inference via Dynamic-Length Float NeurIPS 2025
- π Pronouns: Vi/Vim



