Agent
src.data_models.agent.StreamState
Bases: Enum
Defines the possible states of the streaming agent during processing.
The StreamState enum represents different operational states that the streaming agent can be in at any given time during message processing and tool execution.
Attributes:
Name | Type | Description |
---|---|---|
STREAMING |
Currently streaming response content |
|
TOOL_DETECTION |
Analyzing stream for potential tool calls |
|
EXECUTING_TOOLS |
Currently executing detected tools |
|
INTERMEDIATE |
Temporary state between major operations |
Example
state = StreamState.IDLE
if processing_started:
state = StreamState.STREAMING
Source code in src/data_models/agent.py
12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 |
|
src.data_models.agent.StreamResult
Bases: BaseModel
Represents the result of processing an individual stream chunk.
This class encapsulates the various pieces of information that can be produced when processing a single chunk of a stream, including content, errors, and status updates.
Attributes:
Name | Type | Description |
---|---|---|
content |
Optional[str]
|
The actual content from the stream chunk. May be None if chunk contained no content (e.g., only status updates). |
error |
Optional[str]
|
Error message if any issues occurred during processing. None if processing was successful. |
status |
Optional[str]
|
Status message indicating state changes or completion. Used to communicate processing progress. |
should_continue |
bool
|
Flag indicating if streaming should continue. Defaults to True, set to False for terminating streaming. |
Example
result = StreamResult(
content="Generated text response",
status="Processing complete",
should_continue=True
)
Source code in src/data_models/agent.py
39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 |
|
src.data_models.agent.StreamContext
Bases: BaseModel
Context and state for a streaming conversation session.
This class stores all necessary information for managing a streaming conversation, including conversation history, available tool definitions, state buffers, and session metadata. It also tracks the number of times the streaming state has been initiated.
Attributes:
Name | Type | Description |
---|---|---|
conversation_history |
List[TextChatMessage]
|
The full conversation history, including a system message at the start if available. |
tool_definitions |
List[Tool]
|
Definitions of available tools for execution. |
message_buffer |
str
|
Buffer for accumulating generated response text. |
tool_call_buffer |
str
|
Buffer for accumulating potential tool call text until parsing. |
current_tool_call |
Optional[List[ToolCall]]
|
The currently processing tool calls, if any. |
current_state |
StreamState
|
The current state of the stream processing. |
streaming_entry_count |
int
|
Counter tracking the number of times the streaming state has been entered. |
max_streaming_iterations |
int
|
The maximum allowed number of times the streaming state can be initiated. |
context |
Optional[Dict[str, Any]]
|
Additional metadata associated with the streaming session. |
llm_factory |
Optional[LLMFactory]
|
LLM factory associated with the streaming agent. |
Source code in src/data_models/agent.py
82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 |
|
Config
Pydantic model configuration.
Source code in src/data_models/agent.py
145 146 147 |
|