Contextual Moral Value Alignment Through Context-Based Aggregation
arxiv(2024)
摘要
Developing value-aligned AI agents is a complex undertaking and an ongoing
challenge in the field of AI. Specifically within the domain of Large Language
Models (LLMs), the capability to consolidate multiple independently trained
dialogue agents, each aligned with a distinct moral value, into a unified
system that can adapt to and be aligned with multiple moral values is of
paramount importance. In this paper, we propose a system that does contextual
moral value alignment based on contextual aggregation. Here, aggregation is
defined as the process of integrating a subset of LLM responses that are best
suited to respond to a user input, taking into account features extracted from
the user's input. The proposed system shows better results in term of alignment
to human value compared to the state of the art.
更多查看译文
AI 理解论文
溯源树
样例
生成溯源树,研究论文发展脉络
Chat Paper
正在生成论文摘要