Athena: Safe Autonomous Agents with Verbal Contrastive Learning

T Sadhu, A Pesaranghader, Y Chen, DH Yi - arXiv preprint arXiv …, 2024 - arxiv.org
arXiv preprint arXiv:2408.11021, 2024arxiv.org
Due to emergent capabilities, large language models (LLMs) have been utilized as
language-based agents to perform a variety of tasks and make decisions with an increasing
degree of autonomy. These autonomous agents can understand high-level instructions,
interact with their environments, and execute complex tasks using a selection of tools
available to them. As the capabilities of the agents expand, ensuring their safety and
trustworthiness becomes more imperative. In this study, we introduce the Athena framework …
Due to emergent capabilities, large language models (LLMs) have been utilized as language-based agents to perform a variety of tasks and make decisions with an increasing degree of autonomy. These autonomous agents can understand high-level instructions, interact with their environments, and execute complex tasks using a selection of tools available to them. As the capabilities of the agents expand, ensuring their safety and trustworthiness becomes more imperative. In this study, we introduce the Athena framework which leverages the concept of verbal contrastive learning where past safe and unsafe trajectories are used as in-context (contrastive) examples to guide the agent towards safety while fulfilling a given task. The framework also incorporates a critiquing mechanism to guide the agent to prevent risky actions at every step. Furthermore, due to the lack of existing benchmarks on the safety reasoning ability of LLM-based agents, we curate a set of 80 toolkits across 8 categories with 180 scenarios to provide a safety evaluation benchmark. Our experimental evaluation, with both closed- and open-source LLMs, indicates verbal contrastive learning and interaction-level critiquing improve the safety rate significantly.
arxiv.org