The principles of Unified Alignment for Agents (UA2) with human intentions, environmental dynamics, and self-constraints:

Abstract

The rapid progress of foundation models has led to the prosperity of autonomous agents, which leverage the universal capabilities of foundation models to conduct reasoning, decision-making, and environmental interaction. However, the efficacy of agents remains limited when operating in intricate, realistic environments. In this work, we introduce the principles of Unified Alignment for Agents (UA\(^2\)), which advocate for the simultaneous alignment of agents with human intentions, environmental dynamics, and self-constraints such as the limitation of monetary budgets. From the perspective of UA\(^2\), we review the current agent research and highlight the neglected factors in existing agent benchmarks and method candidates. We also conduct proof-of-concept studies by introducing realistic features to WebShop (Yao et al., 2022), including user profiles to demonstrate intentions, personalized reranking for complex environmental dynamics, and runtime cost statistics to reflect self-constraints. We then follow the principles of UA\(^2\) to propose an initial design of our agent and benchmark its performance with several candidate baselines in the retrofitted WebShop. The extensive experimental results further prove the importance of the principles of UA\(^2\). Our research sheds light on the next steps of autonomous agent research with improved general problem-solving abilities.

Principles of Unified Alignment for Agents

Agents, humans, and the environment are the three components that make up a working system of agents. To promote the orchestration of the three roles, the agents should work in the direction of eliminating the gap between agents and humans, agents and the environment, as well as adapting to the constraints imposed on the agents themselves. Based on this, we propose the principles of Unified Alignment for Agents (UA\(^2\)). To enumerate:

Literature Review from the Lens of UA\(^2\)

According to the principles of UA\(^2\), we review the existing benchmarks and methods of agents and check if they suffice the development of agents in holistic real-world scenarios. Detailed analysis can be found in Section 3 of our paper.

Benchmarks

For benchmarks, we review both the digital and the embodied ones, and consider the following three aspects:

The comparative review is summarized in the following table:

From the table, it is witnessed that existing benchmarks still cannot adequately cover all the three sources of alignments for agents. Specifically, in general, the development of digital benchmarks lags behind that of embodied benchmarks. This underscores the need for the construction of more comprehensive and realistic benchmarks, as well as fine-grained evaluation metrics that account for the principles of UA\(^2\).

Methods

We continue to review representative agent methods from the perspective of UA\(^2\). For each method, we analyze whether it actively seeks alignment with human intentions, environmental dynamics, or self-constraints. The analysis is illustrated in the following figure:

In our analysis, we categorize each methods according to the following criteria:

Other basic techniques like ReAct (Yao et al., 2023) are also the fundamental elements in most of advanced agent frameworks. Despite the emergence of diverse agent methodologies, plenty of room still exists for the unified alignment of agents with human intentions, environmental dynamics, and self-constraints simultaneously. A common case is that a method focuses too much on a single source of alignment, but meanwhile violating other sources severely (for example, when aligning with environmental dynamics by sampling a huge amount of trajectories, the agent aligns poorly with self-constraints). Therefore, elaborate agent framework design is required to strike a good balance of alignments with all the three roles. Detailed discussions can be found in our paper.

Proof-of-Concept Studies

To further validate the importance of UA\(^2\) in the design of both benchmarks and methods for agents, we upgrade WebShop with several realistic features according to the principles of UA\(^2\). We then initiate our agent design with UA\(^2\), and benchmark its performance as well as several candidate baselines on top of the retrofitted WebShop.

Environmental Construction

We implement several realistic features on top of the WebShop environment (Yao et al., 2022) to reflect the three sources of alignment that agents should consider.

Try out the retrofitted WebShop at the live site here!

Details can be found in Section 4.1 of our paper. Weโ€™ve also open-sourced our code for detailed setup if you want to deploy the environment locally.

Agent Design and Experiments

Following the principles of UA\(^2\), we initiate our agent by introducing the structured memory module on top of ReAct. The introduced module is formed by two key components: low-level action insights and high-level intra-task experience. Basically, low-level insights are stored as the key actions extracted from the trajectories of successful runs; when new task arrives, the accumulated key action are to be retrieved as high-level experience to facilitate performance generalization.

The benchmarking results on the retrofitted WebShop are shown as follows:

In the table, the averaged reward, success rate (SR) (%), the alignment gap (%) with human intentions (\(\mathbf{G}_\mathrm{HI}\)) and environment dynamics (\(\mathbf{G}_\mathrm{ED}\)), time (s) and money ($) cost of all methods are benchmarked in our retrofitted WebShop environment. The better performance under each metric is indicated by the darker green shades. Still, check out our paper and code for details =)

Actionable Insights

Envisioning the future of autonomous agents powered by foundation models in real-world applications, in this section, we provide insights on the next steps of research from UA\(^{2}\):

Contact

This project is co-led by Zonghan Yang (yangzh20@mails.tsinghua.edu.cn), An Liu (la22@mails.tsinghua.edu.cn), and Zijun Liu (liuzijun20@mails.tsinghua.edu.cn), and is advised by Peng Li (lipeng@air.tsinghua.edu.cn) and Yang Liu (liuyang2011@tsinghua.edu.cn).

Citation

@article{yang2024unified,
  author        = {Yang, Zonghan and Liu, An and Liu, Zijun and Liu, Kaiming and Xiong, Fangzhou and Wang, Yile and Yang, Zeyuan and Hu, Qingyuan and Chen, Xinrui and Zhang, Zhenhe and Luo, Fuwen and Guo, Zhicheng and Li, Peng and Liu, Yang},
  title         = {Towards Unified Alignment Between Agents, Humans, and Environment},
  year          = {2024},
  eprint        = {2402.07744},
  archivePrefix = {arXiv},
  primaryClass  = {cs.AI}
}