Follow with Complete Python Notebook
Iterating on Your Agent
Let’s demonstrate this workflow by creating an improved version of our support agent with enhanced instructions to improve actionability, then running an experiment to compare it against the initial experiment.Create an Improved Agent
We’ll create a new version of the agent with enhanced instructions that emphasize specific, actionable responses. The key change is in theinstructions parameter in the agent’s prompt.
For the complete implementation including the task function, see the reference notebook.
Run Another Experiment
Run an experiment with the improved agent using the same dataset and evaluator to compare performance:
Comparing Experiments
After running both experiments, you can compare the results in the Phoenix UI. To compare experiments:- Navigate to the Experiments page in Phoenix
- Select the experiments you want to compare by checking the boxes next to their names
- Click the Compare button in the toolbar
- The comparison view will open, showing side-by-side output and metrics for each experiment
- See side-by-side metrics, outputs, and evaluation scores for each experiment
- Identify which examples improved or regressed
- Understand the tradeoffs between different quality dimensions

