Skip to content

(ICLR 2025) AgentRefine: Enhancing Agent Generalization through Refinement Tuning

Notifications You must be signed in to change notification settings

Fu-Dayuan/AgentRefine

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

12 Commits
 
 
 
 

Repository files navigation

AgentRefine: Enhancing Agent Generalization through Refinement Tuning

Static Badge arXiv Website

Code for the Paper "AgentRefine: Enhancing Agent Generalization through Refinement Tuning".

🔔 If you have any questions or suggestions, please don't hesitate to let us know. You can post an issue on this repository.

Outlines

💥 News 💥

  • [01/2025] 🔥 Our paper has been accepted by ICLR 2025.

  • [01/2025] 🔥 We will release our model, inference code in one month!

💡 Overview

We introduce AgentRefine, an agent synthesis framework that enables models to learn from observations within trajectories to correct their own errors. AgentRefine significantly outperforms state-of-the-art agent tuning works in terms of generalization capabilities across diverse agent tasks. Our findings establish a relationship between agent generalization and self-improvement, offering a new paradigm for future research.

📝 Training Data

We provided our training data in HuggingFace:

We will also provide inference code and model soon! Thanks for waiting!

📊 Results

The performance comparison of AgentRefine and other methods across different families and sizes.(The underlined text indicates that the training data is sampled in the same environment as the task and is considered as held-in evaluation.)

Method Alfworld BabyAISciWorldPDDLJericho
SuccessProgressSuccessProgressSuccessProgressSuccessProgressSuccessProgress
GPT Series
GPT-4o66.479.948.264.14076.961.769.810.034.0
GPT-4o-mini37.365.036.651.923.349.825.049.110.028.5
LLaMA-3-8B Series
LLaMA-3-8B-Instruct22.446.145.556.57.841.110.038.40.024.3
AgentGen29.147.620.535.0--11.723.0--
AgentGym61.976.947.361.418.947.51.716.60.012.9
Agent-FLAN67.279.725.035.31.110.98.325.50.010.1
AgentRefine44.863.837.550.414.442.616.637.810.032.3
Mistral Series
Mistral-7B-Instruct-v0.312.435.936.645.86.724.713.327.80.017.3
AgentGym76.986.740.256.315.648.31.77.30.013.0
Agent-FLAN77.687.615.221.006.703.20.00.7
AgentRefine51.468.825.942.44.422.411.732.85.028.8
LLaMA-3-70B Series
LLaMA-3-70B-Instruct67.275.248.261.842.275.455.079.825.046.4
Agent-FLAN80.586.832.141.25.516.425.053.70.013.6
AgentRefine67.272.144.659.717.746.438.358.615.037.2

📖 Citation

Please kindly cite our paper if it helps your research:

@inproceedings{fu2025agentrefine,
  title={AgentRefine: Enhancing Agent Generalization through Refinement Tuning},
  author={Dayuan Fu and Keqing He and Yejie Wang and Wentao Hong and Zhuoma GongQue and Weihao Zeng and Wei Wang and Jingang Wang and Xunliang Cai and Weiran Xu},
  booktitle={The Thirteenth International Conference on Learning Representations},
  year={2025},
  url={https://openreview.net/forum?id=FDimWzmcWn}
}

About

(ICLR 2025) AgentRefine: Enhancing Agent Generalization through Refinement Tuning

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published