반응형

전체 글 1186

Sequential Efficient LLM 논문 -3

https://aclanthology.org/2024.acl-long.536/ Dodo: Dynamic Contextual Compression for Decoder-only LMsGuanghui Qin, Corby Rosset, Ethan Chau, Nikhil Rao, Benjamin Van Durme. Proceedings of the 62nd Annual Meeting of the Association for Computational Linguistics (Volume 1: Long Papers). 2024.aclanthology.orgacl 2024 long에 붙은 논문입니다. 기존 방법들(sparse attention, 커널 등)은 nlp에서 일관적인 효과가 나지 않거나, 대형 llm에 적용이..

Sequential Efficient LLM 논문 -2

https://arxiv.org/abs/2310.01732 Nugget: Neural Agglomerative Embeddings of TextEmbedding text sequences is a widespread requirement in modern language understanding. Existing approaches focus largely on constant-size representations. This is problematic, as the amount of information contained in text often varies with the length of tarxiv.org고정 길이 임베딩은 문장 길이와 정보량이 달라도 동일한 크기로 압축해야 해서 긴 텍스트에서 정보..

Latent Reasoning, Soft Thinking 논문 정리 2

https://aclanthology.org/2025.emnlp-main.36/ CODI: Compressing Chain-of-Thought into Continuous Space via Self-DistillationZhenyi Shen, Hanqi Yan, Linhai Zhang, Zhanghao Hu, Yali Du, Yulan He. Proceedings of the 2025 Conference on Empirical Methods in Natural Language Processing. 2025.aclanthology.orgemnlp 2025 main에 붙었네요 기존 CoT는 토큰 사용량이 너무 많았고, Coconut는 단계적 치환을 통해 latent로 바꾸는데 stage간 망각 가능성을 제..

Latent Reasoning, Soft Thinking 논문 정리 1

https://arxiv.org/abs/2412.06769 Training Large Language Models to Reason in a Continuous Latent SpaceLarge language models (LLMs) are typically constrained to reason in the language space, where they express the reasoning process through a chain-of-thought (CoT) to solve complex problems. However, the language space may not always be optimal for reasoningarxiv.orgCOLM 2025 에 붙었습니다.기존 LLM의 추론은 언..

Privacy AI 관련 조사 12

https://arxiv.org/abs/2505.12540 Harnessing the Universal Geometry of EmbeddingsWe introduce the first method for translating text embeddings from one vector space to another without any paired data, encoders, or predefined sets of matches. Our unsupervised approach translates any embedding to and from a universal latent representatioarxiv.org텍스트 임베딩은 검색, 분류, 클러스터링 등 다양한 곳에 쓰이지만 다른 임베딩 모델은 같은 텍스..

728x90
728x90