diff --git a/pydantic_ai_slim/pydantic_ai/_cli.py b/pydantic_ai_slim/pydantic_ai/_cli.py index 95fcf8b520..ddf76a4676 100644 --- a/pydantic_ai_slim/pydantic_ai/_cli.py +++ b/pydantic_ai_slim/pydantic_ai/_cli.py @@ -228,7 +228,7 @@ async def run_chat( prog_name: str, config_dir: Path | None = None, deps: AgentDepsT = None, - message_history: list[ModelMessage] | None = None, + message_history: Sequence[ModelMessage] | None = None, ) -> int: prompt_history_path = (config_dir or PYDANTIC_AI_HOME) / PROMPT_HISTORY_FILENAME prompt_history_path.parent.mkdir(parents=True, exist_ok=True) @@ -236,7 +236,7 @@ async def run_chat( session: PromptSession[Any] = PromptSession(history=FileHistory(str(prompt_history_path))) multiline = False - messages: list[ModelMessage] = message_history[:] if message_history else [] + messages: list[ModelMessage] = list(message_history) if message_history else [] while True: try: @@ -272,7 +272,7 @@ async def ask_agent( console: Console, code_theme: str, deps: AgentDepsT = None, - messages: list[ModelMessage] | None = None, + messages: Sequence[ModelMessage] | None = None, ) -> list[ModelMessage]: status = Status('[dim]Working on it…[/dim]', console=console) diff --git a/pydantic_ai_slim/pydantic_ai/agent/__init__.py b/pydantic_ai_slim/pydantic_ai/agent/__init__.py index 72c256e9c4..9ba8420299 100644 --- a/pydantic_ai_slim/pydantic_ai/agent/__init__.py +++ b/pydantic_ai_slim/pydantic_ai/agent/__init__.py @@ -415,7 +415,7 @@ def iter( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: None = None, - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -432,7 +432,7 @@ def iter( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: OutputSpec[RunOutputDataT], - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -449,7 +449,7 @@ async def iter( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: OutputSpec[RunOutputDataT] | None = None, - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -566,7 +566,7 @@ async def main(): # Build the initial state usage = usage or _usage.RunUsage() state = _agent_graph.GraphAgentState( - message_history=message_history[:] if message_history else [], + message_history=list(message_history) if message_history else [], usage=usage, retries=0, run_step=0, @@ -690,7 +690,7 @@ def _run_span_end_attributes( } else: attrs = { - 'pydantic_ai.all_messages': json.dumps(settings.messages_to_otel_messages(state.message_history)), + 'pydantic_ai.all_messages': json.dumps(settings.messages_to_otel_messages(list(state.message_history))), **settings.system_instructions_attributes(literal_instructions), } diff --git a/pydantic_ai_slim/pydantic_ai/agent/abstract.py b/pydantic_ai_slim/pydantic_ai/agent/abstract.py index c5d59f7561..419b5120f4 100644 --- a/pydantic_ai_slim/pydantic_ai/agent/abstract.py +++ b/pydantic_ai_slim/pydantic_ai/agent/abstract.py @@ -128,7 +128,7 @@ async def run( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: None = None, - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -146,7 +146,7 @@ async def run( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: OutputSpec[RunOutputDataT], - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -163,7 +163,7 @@ async def run( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: OutputSpec[RunOutputDataT] | None = None, - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -242,7 +242,7 @@ def run_sync( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: None = None, - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -260,7 +260,7 @@ def run_sync( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: OutputSpec[RunOutputDataT], - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -277,7 +277,7 @@ def run_sync( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: OutputSpec[RunOutputDataT] | None = None, - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -348,7 +348,7 @@ def run_stream( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: None = None, - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -366,7 +366,7 @@ def run_stream( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: OutputSpec[RunOutputDataT], - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -384,7 +384,7 @@ async def run_stream( # noqa C901 user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: OutputSpec[RunOutputDataT] | None = None, - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -560,7 +560,7 @@ def run_stream_events( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: None = None, - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -577,7 +577,7 @@ def run_stream_events( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: OutputSpec[RunOutputDataT], - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -593,7 +593,7 @@ def run_stream_events( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: OutputSpec[RunOutputDataT] | None = None, - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -674,7 +674,7 @@ async def _run_stream_events( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: OutputSpec[RunOutputDataT] | None = None, - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -726,7 +726,7 @@ def iter( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: None = None, - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -743,7 +743,7 @@ def iter( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: OutputSpec[RunOutputDataT], - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -761,7 +761,7 @@ async def iter( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: OutputSpec[RunOutputDataT] | None = None, - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -1112,7 +1112,7 @@ async def to_cli( self: Self, deps: AgentDepsT = None, prog_name: str = 'pydantic-ai', - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, ) -> None: """Run the agent in a CLI chat interface. @@ -1149,7 +1149,7 @@ def to_cli_sync( self: Self, deps: AgentDepsT = None, prog_name: str = 'pydantic-ai', - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, ) -> None: """Run the agent in a CLI chat interface with the non-async interface. diff --git a/pydantic_ai_slim/pydantic_ai/agent/wrapper.py b/pydantic_ai_slim/pydantic_ai/agent/wrapper.py index ba735f0907..4e8646d040 100644 --- a/pydantic_ai_slim/pydantic_ai/agent/wrapper.py +++ b/pydantic_ai_slim/pydantic_ai/agent/wrapper.py @@ -72,7 +72,7 @@ def iter( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: None = None, - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -89,7 +89,7 @@ def iter( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: OutputSpec[RunOutputDataT], - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -106,7 +106,7 @@ async def iter( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: OutputSpec[RunOutputDataT] | None = None, - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, diff --git a/pydantic_ai_slim/pydantic_ai/direct.py b/pydantic_ai_slim/pydantic_ai/direct.py index 26d94717fb..137f39fa3a 100644 --- a/pydantic_ai_slim/pydantic_ai/direct.py +++ b/pydantic_ai_slim/pydantic_ai/direct.py @@ -10,7 +10,7 @@ import queue import threading -from collections.abc import Iterator +from collections.abc import Iterator, Sequence from contextlib import AbstractAsyncContextManager from dataclasses import dataclass, field from datetime import datetime @@ -35,7 +35,7 @@ async def model_request( model: models.Model | models.KnownModelName | str, - messages: list[messages.ModelMessage], + messages: Sequence[messages.ModelMessage], *, model_settings: settings.ModelSettings | None = None, model_request_parameters: models.ModelRequestParameters | None = None, @@ -79,7 +79,7 @@ async def main(): """ model_instance = _prepare_model(model, instrument) return await model_instance.request( - messages, + list(messages), model_settings, model_request_parameters or models.ModelRequestParameters(), ) @@ -87,7 +87,7 @@ async def main(): def model_request_sync( model: models.Model | models.KnownModelName | str, - messages: list[messages.ModelMessage], + messages: Sequence[messages.ModelMessage], *, model_settings: settings.ModelSettings | None = None, model_request_parameters: models.ModelRequestParameters | None = None, @@ -133,7 +133,7 @@ def model_request_sync( return _get_event_loop().run_until_complete( model_request( model, - messages, + list(messages), model_settings=model_settings, model_request_parameters=model_request_parameters, instrument=instrument, @@ -143,7 +143,7 @@ def model_request_sync( def model_request_stream( model: models.Model | models.KnownModelName | str, - messages: list[messages.ModelMessage], + messages: Sequence[messages.ModelMessage], *, model_settings: settings.ModelSettings | None = None, model_request_parameters: models.ModelRequestParameters | None = None, @@ -191,7 +191,7 @@ async def main(): """ model_instance = _prepare_model(model, instrument) return model_instance.request_stream( - messages, + list(messages), model_settings, model_request_parameters or models.ModelRequestParameters(), ) @@ -199,7 +199,7 @@ async def main(): def model_request_stream_sync( model: models.Model | models.KnownModelName | str, - messages: list[messages.ModelMessage], + messages: Sequence[messages.ModelMessage], *, model_settings: settings.ModelSettings | None = None, model_request_parameters: models.ModelRequestParameters | None = None, @@ -246,7 +246,7 @@ def model_request_stream_sync( """ async_stream_cm = model_request_stream( model=model, - messages=messages, + messages=list(messages), model_settings=model_settings, model_request_parameters=model_request_parameters, instrument=instrument, diff --git a/pydantic_ai_slim/pydantic_ai/durable_exec/dbos/_agent.py b/pydantic_ai_slim/pydantic_ai/durable_exec/dbos/_agent.py index eb609cfc4a..14f32bdc97 100644 --- a/pydantic_ai_slim/pydantic_ai/durable_exec/dbos/_agent.py +++ b/pydantic_ai_slim/pydantic_ai/durable_exec/dbos/_agent.py @@ -112,7 +112,7 @@ async def wrapped_run_workflow( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: OutputSpec[RunOutputDataT] | None = None, - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT, @@ -149,7 +149,7 @@ def wrapped_run_sync_workflow( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: OutputSpec[RunOutputDataT] | None = None, - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT, @@ -237,7 +237,7 @@ async def run( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: None = None, - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -255,7 +255,7 @@ async def run( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: OutputSpec[RunOutputDataT], - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -272,7 +272,7 @@ async def run( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: OutputSpec[RunOutputDataT] | None = None, - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -341,7 +341,7 @@ def run_sync( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: None = None, - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -359,7 +359,7 @@ def run_sync( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: OutputSpec[RunOutputDataT], - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -376,7 +376,7 @@ def run_sync( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: OutputSpec[RunOutputDataT] | None = None, - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -444,7 +444,7 @@ def run_stream( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: None = None, - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -462,7 +462,7 @@ def run_stream( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: OutputSpec[RunOutputDataT], - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -480,7 +480,7 @@ async def run_stream( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: OutputSpec[RunOutputDataT] | None = None, - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -553,7 +553,7 @@ def run_stream_events( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: None = None, - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -570,7 +570,7 @@ def run_stream_events( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: OutputSpec[RunOutputDataT], - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -586,7 +586,7 @@ def run_stream_events( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: OutputSpec[RunOutputDataT] | None = None, - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -656,7 +656,7 @@ def iter( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: None = None, - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -674,7 +674,7 @@ def iter( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: OutputSpec[RunOutputDataT], - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -692,7 +692,7 @@ async def iter( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: OutputSpec[RunOutputDataT] | None = None, - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, diff --git a/pydantic_ai_slim/pydantic_ai/durable_exec/temporal/_agent.py b/pydantic_ai_slim/pydantic_ai/durable_exec/temporal/_agent.py index 5338f228a9..84b1b047b0 100644 --- a/pydantic_ai_slim/pydantic_ai/durable_exec/temporal/_agent.py +++ b/pydantic_ai_slim/pydantic_ai/durable_exec/temporal/_agent.py @@ -259,7 +259,7 @@ async def run( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: None = None, - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -277,7 +277,7 @@ async def run( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: OutputSpec[RunOutputDataT], - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -294,7 +294,7 @@ async def run( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: OutputSpec[RunOutputDataT] | None = None, - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -369,7 +369,7 @@ def run_sync( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: None = None, - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -387,7 +387,7 @@ def run_sync( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: OutputSpec[RunOutputDataT], - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -404,7 +404,7 @@ def run_sync( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: OutputSpec[RunOutputDataT] | None = None, - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -477,7 +477,7 @@ def run_stream( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: None = None, - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -495,7 +495,7 @@ def run_stream( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: OutputSpec[RunOutputDataT], - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -513,7 +513,7 @@ async def run_stream( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: OutputSpec[RunOutputDataT] | None = None, - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -586,7 +586,7 @@ def run_stream_events( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: None = None, - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -603,7 +603,7 @@ def run_stream_events( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: OutputSpec[RunOutputDataT], - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -619,7 +619,7 @@ def run_stream_events( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: OutputSpec[RunOutputDataT] | None = None, - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -704,7 +704,7 @@ def iter( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: None = None, - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -722,7 +722,7 @@ def iter( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: OutputSpec[RunOutputDataT], - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None, @@ -740,7 +740,7 @@ async def iter( user_prompt: str | Sequence[_messages.UserContent] | None = None, *, output_type: OutputSpec[RunOutputDataT] | None = None, - message_history: list[_messages.ModelMessage] | None = None, + message_history: Sequence[_messages.ModelMessage] | None = None, deferred_tool_results: DeferredToolResults | None = None, model: models.Model | models.KnownModelName | str | None = None, deps: AgentDepsT = None,