Giter VIP home page Giter VIP logo

eto's Introduction

Exploration-Based Trajectory Optimization for LLM Agents

Official repo for Trial and Error: Exploration-Based Trajectory Optimization for LLM Agents

Authors: Yifan Song, Da Yin, Xiang Yue, Jie Huang, Sujian Li, Bill Yuchen Lin.

We introduce ETO (Exploration-based Trajectory Optimization), an agent learning framework inspired by "trial and error" process of human learning. ETO allows an LLM agent to iteratively collect failure trajectories and updates its policy by learning from contrastive failure-success trajectory pairs.

ETO has following features:

  • ๐Ÿ•น๏ธ Learning by Trial and Error
    • ๐ŸŽฒ Learning from Failure Trajectories. Contrary to previous approaches that exclusively train on successful expert trajectories, ETO allows agents to learn from their exploration failures.
    • ๐ŸŽญ Contrastive Trajectory Optimization. ETO applies DPO loss to perform policy learning from failure-success trajectory pairs.
    • ๐ŸŒ Iterative Policy Learning. ETO can be expanded to multiple rounds for further policy enhancement.
  • ๐ŸŽ–๏ธ Superior Performance
    • โš”๏ธ Effectiveness on Three Datasets. ETO significantly outperforms strong baselines, such as RFT, PPO, on WebShop, ScienceWorld, and ALFWorld.
    • ๐Ÿฆพ Generalization on Unseen Scenarios. ETO demonstrates an impressive performance improvement of 22% over SFT on the challenging out-of-distribution test set in ScienceWorld.
    • โŒ› Task-Solving Efficiency. ETO achieves higher rewards within fewer action steps on ScienceWorld.
    • ๐Ÿ’ก Potential in Extreme Scenarios. ETO shows better performance in self-play scenarios where expert trajectories are not available.

๐Ÿ“† Todo

  • Upload the expert trajectories and the agent exploration trajectories on HuggingFace

๐Ÿงฉ Structure of This Project

There are three main folders in this project: envs, eval_agent, fastchat

envs: the interaction environment of WebShop and ScienceWorld. We transform the original WebShop repo into a package.

eval_agent: the evaluation framework of agent tasks, which is inspired by MINT.

fastchat: training scripts for SFT and DPO, which is a modified version of FastChat.

๐Ÿ› ๏ธ Setup

bash setup.sh

The setup script performs the following actions:

  • Install Python dependencies for agent training, deployment, evaluation, and the environments for WebShop, ScienceWorld, ALFWorld
  • Download data and search engine indices for WebShop
  • Download game files for ALFWorld
  • Download expert trajectories for behavioral cloning

๐Ÿš€ Quick Start

The bash script run_eto.sh implements the ETO pipeline. For example, you can run:

# Optional tasks: webshop, sciworld, alfworld
bash run_eto.sh webshop <EXP_NAME> <YOUR_MODEL_PATH> <YOUR_SAVE_PATH>

The script performs the pipeline of ETO:

  1. SFT phase: using the expert trajectories to conduct SFT to get the base agent
  2. Evaluate SFT agent
    1. Launch the FastChat controller
    2. Launch the FastChat model worker
    3. Run the evaluation
    4. Kill the model worker. The controller will be reused in the following steps
  3. Launch multiple FastChat model workers and let the base agent to explore the environment in parallel
  4. Build contrastive failure-success trajectory pairs
  5. Conduct DPO training to learn from the mistakes
  6. Evaluate DPO agent
  7. Repeat 3-6 to iteratively update the policy

๐ŸŽฎ Evaluation

First, launch the controller of FastChat

python -m fastchat.serve.controller

Then, launch the model worker of FastChat

python -m fastchat.serve.model_worker --model-path <YOUR_MODEL_PATH> --port 21002 --worker-address http://localhost:21002

Finally, evaluate the agent

python -m eval_agent.main --agent_config fastchat --model_name <YOUR_MODEL_NAME> --exp_config <TASK_NAME> --split test --verbose

โš™๏ธ How to Add a New Task

  1. Implement your task loader in eval_agent/tasks. You should implement the load_tasks method which returns a task generator.
  2. Implement the corresponding environment in eval_agent/envs. The environment should parse the action generated by the LLM agent, execute the action, and return the observation. The tool/API calling should also be implemented in the environment.
  3. Write the instruction prompt and ICL examples in eval_agent/prompt. The default setting is 1-shot evaluation.
  4. Write a new task config in eval_agent/configs/task. The config defines which task class and environment class to load, and the settings of the environment (e.g., max action steps).

๐Ÿ“‘ The Data Format for Training the Agent

SFT data

[
  {
    "id": "example_0",
    "conversations": [
      {
        "from": "human",
        "value": "Who are you?"
      },
      {
        "from": "gpt",
        "value": "I am Vicuna, a language model trained by researchers from Large Model Systems Organization (LMSYS)."
      },
      {
        "from": "human",
        "value": "Have a nice day!"
      },
      {
        "from": "gpt",
        "value": "You too!"
      }
    ]
  }
]

DPO data

[
    {
        "id": "identity_0",
        "prompt": [
            {
                "from": "human",
                "value": "Hello"
            },
            {
                "from": "gpt",
                "value": "Hi"
            },
            {
                "from": "human",
                "value": "Have a nice day!"
            }
        ],
        "chosen": [
            {
                "from": "gpt",
                "value": "OK!"
            },
            {
                "from": "human",
                "value": "How are you?"
            },
            {
                "from": "gpt",
                "value": "I'm fine"
            }
        ],
        "rejected": [
            {
                "from": "gpt",
                "value": "No, I'm bad"
            }
        ]
    }
]

๐Ÿ“– Citation

If you find this repo helpful, please cite out paper:

@article{song2024trial,
    author={Yifan Song and Da Yin and Xiang Yue and Jie Huang and Sujian Li and Bill Yuchen Lin},
    title={Trial and Error: Exploration-Based Trajectory Optimization for LLM Agents},
    year={2024},
    eprint={2403.02502},
    archivePrefix={arXiv},
    primaryClass={cs.CL}
}

eto's People

Contributors

yifan-song793 avatar

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    ๐Ÿ–– Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. ๐Ÿ“Š๐Ÿ“ˆ๐ŸŽ‰

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google โค๏ธ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.