diff --git a/src/app/endpoints/a2a.py b/src/app/endpoints/a2a.py index a00ece91e..e5fe2abd1 100644 --- a/src/app/endpoints/a2a.py +++ b/src/app/endpoints/a2a.py @@ -340,7 +340,7 @@ async def _process_task_streaming( # pylint: disable=too-many-locals stream = await client.responses.create(**responses_params.model_dump()) except APIConnectionError as e: error_message = ( - f"Unable to connect to Llama Stack backend service: {str(e)}. " + f"Unable to connect to Llama Stack backend service: {e!s}. " "The service may be temporarily unavailable. Please try again later." ) logger.error( diff --git a/src/app/endpoints/health.py b/src/app/endpoints/health.py index 7a8058d79..563f2fca1 100644 --- a/src/app/endpoints/health.py +++ b/src/app/endpoints/health.py @@ -91,7 +91,7 @@ async def get_providers_health_statuses() -> list[ProviderHealthStatus]: ProviderHealthStatus( provider_id="unknown", status=HealthStatus.ERROR.value, - message=f"Failed to initialize health check: {str(e)}", + message=f"Failed to initialize health check: {e!s}", ) ]