반응형

인공지능 633

LLM+P: Empowering Large Language Models with Optimal Planning Proficiency - 논문 리뷰

https://arxiv.org/abs/2304.11477 LLM+P: Empowering Large Language Models with Optimal Planning ProficiencyLarge language models (LLMs) have demonstrated remarkable zero-shot generalization abilities: state-of-the-art chatbots can provide plausible answers to many common questions that arise in daily life. However, so far, LLMs cannot reliably solve long-horizoarxiv.org 로봇을 조종하는 LLM 모델이네요. 장기 계획을..

Language Agent Tree Search Unifies Reasoning Acting and Planning in Language Models

https://arxiv.org/abs/2310.04406 Language Agent Tree Search Unifies Reasoning Acting and Planning in Language ModelsWhile language models (LMs) have shown potential across a range of decision-making tasks, their reliance on simple acting processes limits their broad deployment as autonomous agents. In this paper, we introduce Language Agent Tree Search (LATS) -- the firarxiv.org 이 논문은 강화학습의 MCTS..

ReConcile: Round-Table Conference Improves Reasoning via Consensus among Diverse LLMs - 논문 리뷰

https://arxiv.org/abs/2309.13007 ReConcile: Round-Table Conference Improves Reasoning via Consensus among Diverse LLMsLarge Language Models (LLMs) still struggle with natural language reasoning tasks. Motivated by the society of minds (Minsky, 1988), we propose ReConcile, a multi-model multi-agent framework designed as a round table conference among diverse LLM agents. Rearxiv.org 기존 LLM은 새로운 생각..

Ghost in the Minecraft: Generally Capable Agents for Open-World Environments via Large Language Models with Text-based Knowledge and Memory - 논문 리뷰

https://arxiv.org/abs/2305.17144 Ghost in the Minecraft: Generally Capable Agents for Open-World Environments via Large Language Models with Text-based KnowledgeThe captivating realm of Minecraft has attracted substantial research interest in recent years, serving as a rich platform for developing intelligent agents capable of functioning in open-world environments. However, the current research..

ExpeL: LLM Agents Are Experiential Learners - 논문 리뷰

https://arxiv.org/abs/2308.10144 ExpeL: LLM Agents Are Experiential LearnersThe recent surge in research interest in applying large language models (LLMs) to decision-making tasks has flourished by leveraging the extensive world knowledge embedded in LLMs. While there is a growing demand to tailor LLMs for custom decision-making tarxiv.org  이 논문도 LLM이 새로운 정보에 대해 어떻게 저장하거나 사용할지에 대한 논문입니다.문제: LLM을..

Randomized Positional Encodings Boost Length Generalization of Transformers - 논문 리뷰

https://arxiv.org/abs/2305.16843 Randomized Positional Encodings Boost Length Generalization of TransformersTransformers have impressive generalization capabilities on tasks with a fixed context length. However, they fail to generalize to sequences of arbitrary length, even for seemingly simple tasks such as duplicating a string. Moreover, simply training on lonarxiv.org 학습 혹은 추론 때 토큰 길이에 대한 논문이..

Sparse Autoencoder를 통한 LLM의 Bias 줄이기 - 성에 따른 직업 1

2024.11.05 - [인공지능/논문 리뷰 or 진행] - Bias and Fairness in Large Language Models: A Survey Bias and Fairness in Large Language Models: A Surveyhttps://arxiv.org/abs/2309.00770 Bias and Fairness in Large Language Models: A SurveyRapid advancements of large language models (LLMs) have enabled the processing, understanding, and generation of human-like text, with increasing integration into systemsyoon..

인공지능/XAI 2024.11.26

Editing Large Language Models: Problems, Methods, and Opportunities - 논문 리뷰

https://arxiv.org/abs/2305.13172 Editing Large Language Models: Problems, Methods, and OpportunitiesDespite the ability to train capable LLMs, the methodology for maintaining their relevancy and rectifying errors remains elusive. To this end, the past few years have witnessed a surge in techniques for editing LLMs, the objective of which is to efficientlarxiv.org 이 논문은 새로운 정보를 반영하지 못 하는 LLM의 단점을..

FreshLLMs: Refreshing Large Language Models with Search Engine Augmentation - 논문 리뷰

https://arxiv.org/abs/2310.03214 FreshLLMs: Refreshing Large Language Models with Search Engine AugmentationMost large language models (LLMs) are trained once and never updated; thus, they lack the ability to dynamically adapt to our ever-changing world. In this work, we perform a detailed study of the factuality of LLM-generated text in the context of answeringarxiv.org검색을 통해 LLM의 최신 정보 미달을 해결하..

A Multitask, Multilingual, Multimodal Evaluation of ChatGPT on Reasoning, Hallucination, and Interactivity - 논문 리뷰

https://arxiv.org/abs/2302.04023 A Multitask, Multilingual, Multimodal Evaluation of ChatGPT on Reasoning, Hallucination, and InteractivityThis paper proposes a framework for quantitatively evaluating interactive LLMs such as ChatGPT using publicly available data sets. We carry out an extensive technical evaluation of ChatGPT using 23 data sets covering 8 different common NLP application taskarx..

728x90
728x90