Skip to content
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
11 changes: 3 additions & 8 deletions splunklib/ai/engines/langchain.py
Original file line number Diff line number Diff line change
Expand Up @@ -1266,7 +1266,7 @@ def _convert_model_result_from_lc(model_response: LC_ModelCallResult) -> ModelRe


def _convert_agent_state_to_lc(state: AgentState) -> LC_AgentState[Any]:
messages = [_map_message_to_langchain(m) for m in state.response.messages]
messages = [_map_message_to_langchain(m) for m in state.messages]
return LC_AgentState(messages=messages)


Expand Down Expand Up @@ -1627,14 +1627,9 @@ def _convert_agent_state_from_langchain(
messages = state["messages"]
total_tokens_counter = _get_approximate_token_counter(model)
total_tokens = total_tokens_counter(messages)

response = AgentResponse[Any | None](
messages=[_map_message_from_langchain(m) for m in state["messages"]],
structured_output=state.get("structured_response"),
)

messages = [_map_message_from_langchain(m) for m in state["messages"]]
return AgentState(
response=response,
messages=messages,
total_steps=len(messages),
token_count=total_tokens,
)
Expand Down
6 changes: 3 additions & 3 deletions splunklib/ai/middleware.py
Original file line number Diff line number Diff line change
Expand Up @@ -12,7 +12,7 @@
# License for the specific language governing permissions and limitations
# under the License.

from collections.abc import Awaitable, Callable
from collections.abc import Sequence, Awaitable, Callable
from dataclasses import dataclass
from typing import Any, override

Expand All @@ -35,7 +35,7 @@ class AgentState:
"""AgentState is available through certain middlewares and contains information about the current state of an agent execution."""

# holds messages exchanged so far in the conversation
response: AgentResponse[Any | None]
messages: Sequence[BaseMessage]
# steps taken so far in the conversation
total_steps: int
# tokens used so far in the conversation
Expand Down Expand Up @@ -96,7 +96,7 @@ def __post_init__(self) -> None:

@dataclass(frozen=True)
class AgentRequest:
messages: list[BaseMessage]
messages: Sequence[BaseMessage]


AgentMiddlewareHandler = Callable[[AgentRequest], Awaitable[AgentResponse[Any | None]]]
Expand Down
2 changes: 1 addition & 1 deletion tests/integration/ai/test_agent.py
Original file line number Diff line number Diff line change
Expand Up @@ -532,7 +532,7 @@ async def _model_call_middleware(
req: ModelRequest, _handler: ModelMiddlewareHandler
) -> ModelResponse:
if after_subagent_call:
msgs = req.state.response.messages
msgs = req.state.messages
assert isinstance(msgs[-1], SubagentMessage)
assert isinstance(msgs[-1].result, SubagentFailureResult)

Expand Down
14 changes: 7 additions & 7 deletions tests/integration/ai/test_conversation_store.py
Original file line number Diff line number Diff line change
Expand Up @@ -66,9 +66,9 @@ async def _model_middleware(

if after_first_call:
# Previous messages included.
assert len(request.state.response.messages) == 3
assert len(request.state.messages) == 3
else:
assert len(request.state.response.messages) == 1
assert len(request.state.messages) == 1
return await handler(request)

@agent_middleware
Expand Down Expand Up @@ -166,7 +166,7 @@ async def _model_middleware(
nonlocal model_middleware_called
model_middleware_called = True

assert len(request.state.response.messages) == 1
assert len(request.state.messages) == 1
return await handler(request)

async with Agent(
Expand Down Expand Up @@ -276,9 +276,9 @@ async def _model_middleware(
nonlocal after_first_call

if after_first_call:
assert len(request.state.response.messages) == 3
assert len(request.state.messages) == 3
else:
assert len(request.state.response.messages) == 1
assert len(request.state.messages) == 1

after_first_call = True
return await handler(request)
Expand Down Expand Up @@ -347,9 +347,9 @@ async def _model_middleware(
nonlocal after_first_call

if after_first_call:
assert len(request.state.response.messages) == 3
assert len(request.state.messages) == 3
else:
assert len(request.state.response.messages) == 1
assert len(request.state.messages) == 1

after_first_call = True
return await handler(request)
Expand Down
4 changes: 2 additions & 2 deletions tests/integration/ai/test_hooks.py
Original file line number Diff line number Diff line change
Expand Up @@ -47,15 +47,15 @@ def test_hook_before(req: ModelRequest) -> None:
hook_calls += 1

assert req.system_message.startswith("Your name is stefan")
assert len(req.state.response.messages) == 1
assert len(req.state.messages) == 1

@before_model
async def test_async_hook_before(req: ModelRequest) -> None:
nonlocal hook_calls
hook_calls += 1

assert req.system_message.startswith("Your name is stefan")
assert len(req.state.response.messages) == 1
assert len(req.state.messages) == 1

@after_model
def test_hook_after(resp: ModelResponse) -> None:
Expand Down
7 changes: 2 additions & 5 deletions tests/integration/ai/test_middleware.py
Original file line number Diff line number Diff line change
Expand Up @@ -78,7 +78,7 @@ async def test_middleware(
assert call.args == {"city": "Krakow"}

state = request.state
assert len(state.response.messages) == 2
assert len(state.messages) == 2

response = await handler(request)
assert isinstance(response.result, ToolResult)
Expand Down Expand Up @@ -699,10 +699,7 @@ async def mutating_middleware(
) -> ModelResponse:
new_state = replace(
request.state,
response=replace(
request.state.response,
messages=[HumanMessage(content="What is the capital of France?")],
),
messages=[HumanMessage(content="What is the capital of France?")],
)
return await handler(replace(request, state=new_state))

Expand Down
5 changes: 2 additions & 3 deletions tests/unit/ai/test_default_limits.py
Original file line number Diff line number Diff line change
Expand Up @@ -48,7 +48,7 @@ def _make_agent_request() -> AgentRequest:

def _make_model_request(token_count: int = 0, total_steps: int = 0) -> ModelRequest:
state = AgentState(
response=AgentResponse(messages=[], structured_output=None),
messages=[],
total_steps=total_steps,
token_count=token_count,
)
Expand Down Expand Up @@ -141,7 +141,7 @@ async def test_timeout_fires_when_deadline_exceeded(self) -> None:
mw = TimeoutLimitMiddleware(60.0)
mw._deadline = monotonic() - 1.0 # pyright: ignore[reportPrivateUsage] # already in the past

state = AgentState(response=AgentResponse(messages=[], structured_output=None), total_steps=0, token_count=0)
state = AgentState(messages=[], total_steps=0, token_count=0)
request = ModelRequest(system_message="", state=state)

with self.assertRaises(TimeoutExceededException):
Expand All @@ -166,4 +166,3 @@ async def test_raises_when_steps_in_request_reach_limit(self) -> None:
await mw.model_middleware(_make_model_request(total_steps=2), _noop_model_handler)
with self.assertRaises(StepsLimitExceededException):
await mw.model_middleware(_make_model_request(total_steps=3), _noop_model_handler)

Loading