2025 5 2

3 papers were accepted at ICML 2025. We proposed CSR that builds state-of-the-art shortening embedding models (image/text/multimodal) with sparse coding. We characterized the reasons behind Transformers’ position bias and how LLMs’ length generalization requires output alignment.