Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
65 changes: 14 additions & 51 deletions docs/agents.md
Original file line number Diff line number Diff line change
Expand Up @@ -135,30 +135,21 @@ async def main():
UserPromptPart(
content='What is the capital of France?',
timestamp=datetime.datetime(...),
part_kind='user-prompt',
)
],
instructions=None,
kind='request',
]
)
),
CallToolsNode(
model_response=ModelResponse(
parts=[TextPart(content='Paris', part_kind='text')],
parts=[TextPart(content='Paris')],
usage=Usage(
requests=1,
request_tokens=56,
response_tokens=1,
total_tokens=57,
details=None,
requests=1, request_tokens=56, response_tokens=1, total_tokens=57
),
model_name='gpt-4o',
timestamp=datetime.datetime(...),
kind='response',
vendor_id=None,
)
),
End(data=FinalResult(output='Paris', tool_name=None, tool_call_id=None)),
End(data=FinalResult(output='Paris')),
]
"""
print(agent_run.result.output)
Expand Down Expand Up @@ -207,30 +198,24 @@ async def main():
UserPromptPart(
content='What is the capital of France?',
timestamp=datetime.datetime(...),
part_kind='user-prompt',
)
],
instructions=None,
kind='request',
]
)
),
CallToolsNode(
model_response=ModelResponse(
parts=[TextPart(content='Paris', part_kind='text')],
parts=[TextPart(content='Paris')],
usage=Usage(
requests=1,
request_tokens=56,
response_tokens=1,
total_tokens=57,
details=None,
),
model_name='gpt-4o',
timestamp=datetime.datetime(...),
kind='response',
vendor_id=None,
)
),
End(data=FinalResult(output='Paris', tool_name=None, tool_call_id=None)),
End(data=FinalResult(output='Paris')),
]
"""
```
Expand Down Expand Up @@ -370,7 +355,7 @@ if __name__ == '__main__':
[
'=== UserPromptNode: What will the weather be like in Paris on Tuesday? ===',
'=== ModelRequestNode: streaming partial request tokens ===',
"[Request] Starting part 0: ToolCallPart(tool_name='weather_forecast', args=None, tool_call_id='0001', part_kind='tool-call')",
"[Request] Starting part 0: ToolCallPart(tool_name='weather_forecast', tool_call_id='0001')",
'[Request] Part 0 args_delta={"location":"Pa',
'[Request] Part 0 args_delta=ris","forecast_',
'[Request] Part 0 args_delta=date":"2030-01-',
Expand All @@ -379,7 +364,7 @@ if __name__ == '__main__':
'[Tools] The LLM calls tool=\'weather_forecast\' with args={"location":"Paris","forecast_date":"2030-01-01"} (tool_call_id=\'0001\')',
"[Tools] Tool call '0001' returned => The forecast in Paris on 2030-01-01 is 24°C and sunny.",
'=== ModelRequestNode: streaming partial request tokens ===',
"[Request] Starting part 0: TextPart(content='It will be ', part_kind='text')",
"[Request] Starting part 0: TextPart(content='It will be ')",
'[Result] The model produced a final output (tool_name=None)',
"[Request] Part 0 text delta: 'warm and sunny '",
"[Request] Part 0 text delta: 'in Paris on '",
Expand Down Expand Up @@ -417,9 +402,7 @@ result_sync = agent.run_sync(
print(result_sync.output)
#> Rome
print(result_sync.usage())
"""
Usage(requests=1, request_tokens=62, response_tokens=1, total_tokens=63, details=None)
"""
#> Usage(requests=1, request_tokens=62, response_tokens=1, total_tokens=63)

try:
result_sync = agent.run_sync(
Expand Down Expand Up @@ -831,32 +814,22 @@ with capture_run_messages() as messages: # (2)!
UserPromptPart(
content='Please get me the volume of a box with size 6.',
timestamp=datetime.datetime(...),
part_kind='user-prompt',
)
],
instructions=None,
kind='request',
]
),
ModelResponse(
parts=[
ToolCallPart(
tool_name='calc_volume',
args={'size': 6},
tool_call_id='pyd_ai_tool_call_id',
part_kind='tool-call',
)
],
usage=Usage(
requests=1,
request_tokens=62,
response_tokens=4,
total_tokens=66,
details=None,
requests=1, request_tokens=62, response_tokens=4, total_tokens=66
),
model_name='gpt-4o',
timestamp=datetime.datetime(...),
kind='response',
vendor_id=None,
),
ModelRequest(
parts=[
Expand All @@ -865,32 +838,22 @@ with capture_run_messages() as messages: # (2)!
tool_name='calc_volume',
tool_call_id='pyd_ai_tool_call_id',
timestamp=datetime.datetime(...),
part_kind='retry-prompt',
)
],
instructions=None,
kind='request',
]
),
ModelResponse(
parts=[
ToolCallPart(
tool_name='calc_volume',
args={'size': 6},
tool_call_id='pyd_ai_tool_call_id',
part_kind='tool-call',
)
],
usage=Usage(
requests=1,
request_tokens=72,
response_tokens=8,
total_tokens=80,
details=None,
requests=1, request_tokens=72, response_tokens=8, total_tokens=80
),
model_name='gpt-4o',
timestamp=datetime.datetime(...),
kind='response',
vendor_id=None,
),
]
"""
Expand Down
5 changes: 1 addition & 4 deletions docs/api/models/function.md
Original file line number Diff line number Diff line change
Expand Up @@ -28,11 +28,8 @@ async def model_function(
UserPromptPart(
content='Testing my agent...',
timestamp=datetime.datetime(...),
part_kind='user-prompt',
)
],
instructions=None,
kind='request',
]
)
]
"""
Expand Down
15 changes: 2 additions & 13 deletions docs/direct.md
Original file line number Diff line number Diff line change
Expand Up @@ -27,9 +27,7 @@ model_response = model_request_sync(
print(model_response.parts[0].content)
#> Paris
print(model_response.usage)
"""
Usage(requests=1, request_tokens=56, response_tokens=1, total_tokens=57, details=None)
"""
#> Usage(requests=1, request_tokens=56, response_tokens=1, total_tokens=57)
```

_(This example is complete, it can be run "as is")_
Expand Down Expand Up @@ -82,20 +80,11 @@ async def main():
tool_name='divide',
args={'numerator': '123', 'denominator': '456'},
tool_call_id='pyd_ai_2e0e396768a14fe482df90a29a78dc7b',
part_kind='tool-call',
)
],
usage=Usage(
requests=1,
request_tokens=55,
response_tokens=7,
total_tokens=62,
details=None,
),
usage=Usage(requests=1, request_tokens=55, response_tokens=7, total_tokens=62),
model_name='gpt-4.1-nano',
timestamp=datetime.datetime(...),
kind='response',
vendor_id=None,
)
"""
```
Expand Down
Loading