-
Notifications
You must be signed in to change notification settings - Fork 5.3k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
AgentOptimizer #1053
AgentOptimizer #1053
Conversation
Codecov ReportAll modified and coverable lines are covered by tests ✅
Additional details and impacted files@@ Coverage Diff @@
## main #1053 +/- ##
=======================================
Coverage 30.23% 30.23%
=======================================
Files 30 30
Lines 3989 3989
Branches 899 899
=======================================
Hits 1206 1206
Misses 2704 2704
Partials 79 79
Flags with carried forward coverage won't be shown. Click here to find out more. ☔ View full report in Codecov by Sentry. |
@skzhang1 Thanks for the contribution. To put the PR in perspective, can you a general outline of techniques anticipated for updating the functions? |
@rickyloynd-microsoft Thanks for the suggestion. We just write a blog which provides an overview for this updating. |
* doc and test update * docker update
* first * first * update figure * update figure * update txt+figure * format * blog * blog update * format * update * update * rename notebook * rename --------- Co-authored-by: “skzhang1” <“[email protected]”> Co-authored-by: Qingyun Wu <[email protected]> Co-authored-by: Jieyu Zhang <[email protected]>
Why are these changes needed?
In traditional ML pipeline, we train a model by updating its parameter according to the loss on the training set, while in the era of LLM agents, how should we train an agent? Here, we take an initial step towards the agent training. Inspired by the function calling capabilities provided by OpenAI, we draw an analogy between model parameters and agent functions/skills, and update agent’s functions/skills based on its historical performance on the training set.
In this notebook, we introduce a new class, ‘AgentOptimizer’, which is able to improve the function list of one Assistant-UserProxy pair according to the historical conversation histories. This feature would support agents in improving their ability to solve problems of the same type as previous tasks.
Any suggestions are appreciated.
Related issue number
Checks