feat: add chat_tui example for FastGPT with Textual interface
- Introduced a new example script `chat_tui.py` that provides a full-screen Textual interface for interacting with FastGPT. - Implemented streaming chat updates, workflow logging, and modal handling for interactive nodes. - Enhanced FastGPT client with new streaming capabilities and structured event types for better interaction handling. - Normalized base URL handling in the client to prevent duplicate `/api` paths. - Added tests for streaming event parsing and interaction handling.
This commit is contained in:
@@ -1,4 +1,4 @@
|
||||
"""FastGPT Python SDK
|
||||
"""FastGPT Python SDK
|
||||
|
||||
A Python client library for interacting with FastGPT's OpenAPI.
|
||||
"""
|
||||
@@ -17,6 +17,8 @@ from fastgpt_client.exceptions import (
|
||||
StreamParseError,
|
||||
ValidationError,
|
||||
)
|
||||
from fastgpt_client.streaming import aiter_stream_events, iter_stream_events
|
||||
from fastgpt_client.stream_types import FastGPTInteractiveEvent, FastGPTStreamEvent
|
||||
|
||||
__all__ = [
|
||||
# Synchronous clients
|
||||
@@ -34,6 +36,11 @@ __all__ = [
|
||||
"RateLimitError",
|
||||
"ValidationError",
|
||||
"StreamParseError",
|
||||
# Streaming helpers
|
||||
"iter_stream_events",
|
||||
"aiter_stream_events",
|
||||
"FastGPTStreamEvent",
|
||||
"FastGPTInteractiveEvent",
|
||||
]
|
||||
|
||||
__version__ = "0.1.0"
|
||||
|
||||
@@ -44,9 +44,10 @@ class AsyncFastGPTClient(BaseClientMixin):
|
||||
# Initialize base client functionality
|
||||
super().__init__(api_key, base_url, timeout, max_retries, retry_delay, enable_logging)
|
||||
|
||||
connect_timeout = min(float(timeout), 15.0) if timeout and timeout > 0 else 15.0
|
||||
self._client = httpx.AsyncClient(
|
||||
base_url=base_url,
|
||||
timeout=httpx.Timeout(timeout, connect=5.0),
|
||||
base_url=self.base_url,
|
||||
timeout=httpx.Timeout(timeout, connect=connect_timeout),
|
||||
)
|
||||
|
||||
async def __aenter__(self):
|
||||
@@ -270,6 +271,9 @@ class AsyncFastGPTClient(BaseClientMixin):
|
||||
try:
|
||||
error_data = response.json()
|
||||
message = error_data.get("message", f"HTTP {response.status_code}")
|
||||
except httpx.ResponseNotRead:
|
||||
message = f"HTTP {response.status_code}"
|
||||
error_data = None
|
||||
except (ValueError, KeyError, AttributeError):
|
||||
# If we can't parse JSON (e.g., streaming response or invalid JSON), use status code
|
||||
message = f"HTTP {response.status_code}"
|
||||
|
||||
@@ -8,6 +8,15 @@ from typing import Any
|
||||
class BaseClientMixin:
|
||||
"""Mixin class providing retry logic, validation, and logging for FastGPT clients."""
|
||||
|
||||
@staticmethod
|
||||
def _normalize_base_url(base_url: str) -> str:
|
||||
"""Normalize FastGPT base URLs so '/api' is not duplicated by endpoint paths."""
|
||||
|
||||
normalized = str(base_url or "").strip().rstrip("/")
|
||||
if normalized.endswith("/api"):
|
||||
normalized = normalized[:-4]
|
||||
return normalized
|
||||
|
||||
def __init__(
|
||||
self,
|
||||
api_key: str,
|
||||
@@ -28,7 +37,7 @@ class BaseClientMixin:
|
||||
enable_logging: Whether to enable request logging
|
||||
"""
|
||||
self.api_key = api_key
|
||||
self.base_url = base_url
|
||||
self.base_url = self._normalize_base_url(base_url)
|
||||
self.timeout = timeout
|
||||
self.max_retries = max_retries
|
||||
self.retry_delay = retry_delay
|
||||
|
||||
@@ -43,9 +43,10 @@ class FastGPTClient(BaseClientMixin):
|
||||
# Initialize base client functionality
|
||||
super().__init__(api_key, base_url, timeout, max_retries, retry_delay, enable_logging)
|
||||
|
||||
connect_timeout = min(float(timeout), 15.0) if timeout and timeout > 0 else 15.0
|
||||
self._client = httpx.Client(
|
||||
base_url=base_url,
|
||||
timeout=httpx.Timeout(timeout, connect=5.0),
|
||||
base_url=self.base_url,
|
||||
timeout=httpx.Timeout(timeout, connect=connect_timeout),
|
||||
)
|
||||
|
||||
def __enter__(self):
|
||||
@@ -206,6 +207,19 @@ class FastGPTClient(BaseClientMixin):
|
||||
try:
|
||||
error_data = response.json()
|
||||
message = error_data.get("message", f"HTTP {response.status_code}")
|
||||
except httpx.ResponseNotRead:
|
||||
try:
|
||||
response.read()
|
||||
error_data = response.json()
|
||||
message = error_data.get("message", f"HTTP {response.status_code}")
|
||||
except Exception:
|
||||
error_text = ""
|
||||
try:
|
||||
error_text = response.text.strip()
|
||||
except Exception:
|
||||
error_text = ""
|
||||
message = error_text or f"HTTP {response.status_code}"
|
||||
error_data = None
|
||||
except (ValueError, KeyError, AttributeError):
|
||||
# If we can't parse JSON (e.g., streaming response or invalid JSON), use status code
|
||||
message = f"HTTP {response.status_code}"
|
||||
|
||||
41
fastgpt_client/stream_types.py
Normal file
41
fastgpt_client/stream_types.py
Normal file
@@ -0,0 +1,41 @@
|
||||
"""Typed FastGPT SSE stream events."""
|
||||
|
||||
from __future__ import annotations
|
||||
|
||||
from dataclasses import dataclass, field
|
||||
from typing import Any, Literal
|
||||
|
||||
|
||||
FastGPTEventKind = Literal[
|
||||
"data",
|
||||
"answer",
|
||||
"fastAnswer",
|
||||
"flowNodeStatus",
|
||||
"flowResponses",
|
||||
"interactive",
|
||||
"toolCall",
|
||||
"toolParams",
|
||||
"toolResponse",
|
||||
"updateVariables",
|
||||
"error",
|
||||
"done",
|
||||
]
|
||||
|
||||
InteractiveType = Literal["userSelect", "userInput"]
|
||||
|
||||
|
||||
@dataclass(frozen=True)
|
||||
class FastGPTStreamEvent:
|
||||
"""Structured event parsed from a FastGPT SSE response."""
|
||||
|
||||
kind: FastGPTEventKind
|
||||
data: Any = field(default_factory=dict)
|
||||
raw_event: str = ""
|
||||
raw_data: str = ""
|
||||
|
||||
|
||||
@dataclass(frozen=True)
|
||||
class FastGPTInteractiveEvent(FastGPTStreamEvent):
|
||||
"""Interactive workflow event requiring client input."""
|
||||
|
||||
interaction_type: InteractiveType = "userSelect"
|
||||
180
fastgpt_client/streaming.py
Normal file
180
fastgpt_client/streaming.py
Normal file
@@ -0,0 +1,180 @@
|
||||
"""Helpers for parsing FastGPT SSE responses."""
|
||||
|
||||
from __future__ import annotations
|
||||
|
||||
import json
|
||||
from typing import Any, AsyncIterator, Dict, Iterator, Optional
|
||||
|
||||
from .exceptions import StreamParseError
|
||||
from .stream_types import FastGPTInteractiveEvent, FastGPTStreamEvent
|
||||
|
||||
|
||||
def _infer_interaction_type(payload: Dict[str, Any]) -> str:
|
||||
interaction_type = str(payload.get("type") or "").strip()
|
||||
if interaction_type in {"userSelect", "userInput"}:
|
||||
return interaction_type
|
||||
|
||||
params = payload.get("params") if isinstance(payload.get("params"), dict) else {}
|
||||
if isinstance(params.get("inputForm"), list):
|
||||
return "userInput"
|
||||
if isinstance(params.get("userSelectOptions"), list):
|
||||
return "userSelect"
|
||||
return "userSelect"
|
||||
|
||||
|
||||
def _normalize_interactive_payload(payload: Dict[str, Any]) -> Dict[str, Any]:
|
||||
normalized = payload
|
||||
wrapped = normalized.get("interactive")
|
||||
if isinstance(wrapped, dict):
|
||||
normalized = wrapped
|
||||
|
||||
interaction_type = str(normalized.get("type") or "").strip()
|
||||
if interaction_type == "toolChildrenInteractive":
|
||||
params = normalized.get("params") if isinstance(normalized.get("params"), dict) else {}
|
||||
children_response = params.get("childrenResponse")
|
||||
if isinstance(children_response, dict):
|
||||
normalized = children_response
|
||||
|
||||
return normalized
|
||||
|
||||
|
||||
def _coerce_line(line: Any) -> str:
|
||||
if isinstance(line, bytes):
|
||||
return line.decode("utf-8")
|
||||
return str(line)
|
||||
|
||||
|
||||
def _build_event(event_name: str, raw_data: str) -> FastGPTStreamEvent:
|
||||
normalized_event = event_name or "data"
|
||||
if raw_data == "[DONE]":
|
||||
return FastGPTStreamEvent(kind="done", raw_event=normalized_event, raw_data=raw_data)
|
||||
|
||||
try:
|
||||
payload = json.loads(raw_data) if raw_data else {}
|
||||
except json.JSONDecodeError as exc:
|
||||
raise StreamParseError(
|
||||
f"Failed to parse SSE payload for event '{normalized_event}': {exc}"
|
||||
) from exc
|
||||
|
||||
if normalized_event == "interactive":
|
||||
if not isinstance(payload, dict):
|
||||
raise StreamParseError(
|
||||
f"Expected JSON object payload for event '{normalized_event}', got {type(payload).__name__}"
|
||||
)
|
||||
normalized_payload = _normalize_interactive_payload(payload)
|
||||
if not isinstance(normalized_payload, dict):
|
||||
raise StreamParseError(
|
||||
f"Expected JSON object payload for event '{normalized_event}', got {type(normalized_payload).__name__}"
|
||||
)
|
||||
interaction_type = _infer_interaction_type(normalized_payload)
|
||||
return FastGPTInteractiveEvent(
|
||||
kind="interactive",
|
||||
data=normalized_payload,
|
||||
raw_event=normalized_event,
|
||||
raw_data=raw_data,
|
||||
interaction_type=interaction_type,
|
||||
)
|
||||
|
||||
allowed_kinds = {
|
||||
"data",
|
||||
"answer",
|
||||
"fastAnswer",
|
||||
"flowNodeStatus",
|
||||
"flowResponses",
|
||||
"toolCall",
|
||||
"toolParams",
|
||||
"toolResponse",
|
||||
"updateVariables",
|
||||
"error",
|
||||
"done",
|
||||
}
|
||||
kind = normalized_event if normalized_event in allowed_kinds else "data"
|
||||
return FastGPTStreamEvent(kind=kind, data=payload, raw_event=normalized_event, raw_data=raw_data)
|
||||
|
||||
|
||||
def _parse_sse_lines(lines: Iterator[Any]) -> Iterator[FastGPTStreamEvent]:
|
||||
current_event: Optional[str] = None
|
||||
data_lines: list[str] = []
|
||||
|
||||
def flush() -> Optional[FastGPTStreamEvent]:
|
||||
nonlocal current_event, data_lines
|
||||
if not current_event and not data_lines:
|
||||
return None
|
||||
if not data_lines:
|
||||
raise StreamParseError(f"Missing data payload for SSE event '{current_event or 'data'}'")
|
||||
raw_data = "\n".join(data_lines)
|
||||
event = _build_event(current_event or "data", raw_data)
|
||||
current_event = None
|
||||
data_lines = []
|
||||
return event
|
||||
|
||||
for raw_line in lines:
|
||||
line = _coerce_line(raw_line).rstrip("\r\n")
|
||||
if not line:
|
||||
event = flush()
|
||||
if event is not None:
|
||||
yield event
|
||||
continue
|
||||
if line.startswith(":"):
|
||||
continue
|
||||
if line.startswith("event:"):
|
||||
if current_event is not None or data_lines:
|
||||
raise StreamParseError("Encountered new SSE event before previous event payload was terminated")
|
||||
current_event = line[6:].strip() or "data"
|
||||
continue
|
||||
if line.startswith("data:"):
|
||||
data_lines.append(line[5:].strip())
|
||||
continue
|
||||
raise StreamParseError(f"Unsupported SSE line: {line}")
|
||||
|
||||
event = flush()
|
||||
if event is not None:
|
||||
yield event
|
||||
|
||||
|
||||
def iter_stream_events(response) -> Iterator[FastGPTStreamEvent]:
|
||||
"""Yield typed SSE events from a synchronous response stream."""
|
||||
|
||||
yield from _parse_sse_lines(iter(response.iter_lines()))
|
||||
|
||||
|
||||
async def aiter_stream_events(response) -> AsyncIterator[FastGPTStreamEvent]:
|
||||
"""Yield typed SSE events from an asynchronous response stream."""
|
||||
|
||||
current_event: Optional[str] = None
|
||||
data_lines: list[str] = []
|
||||
|
||||
async def flush() -> Optional[FastGPTStreamEvent]:
|
||||
nonlocal current_event, data_lines
|
||||
if not current_event and not data_lines:
|
||||
return None
|
||||
if not data_lines:
|
||||
raise StreamParseError(f"Missing data payload for SSE event '{current_event or 'data'}'")
|
||||
raw_data = "\n".join(data_lines)
|
||||
event = _build_event(current_event or "data", raw_data)
|
||||
current_event = None
|
||||
data_lines = []
|
||||
return event
|
||||
|
||||
async for raw_line in response.aiter_lines():
|
||||
line = _coerce_line(raw_line).rstrip("\r\n")
|
||||
if not line:
|
||||
event = await flush()
|
||||
if event is not None:
|
||||
yield event
|
||||
continue
|
||||
if line.startswith(":"):
|
||||
continue
|
||||
if line.startswith("event:"):
|
||||
if current_event is not None or data_lines:
|
||||
raise StreamParseError("Encountered new SSE event before previous event payload was terminated")
|
||||
current_event = line[6:].strip() or "data"
|
||||
continue
|
||||
if line.startswith("data:"):
|
||||
data_lines.append(line[5:].strip())
|
||||
continue
|
||||
raise StreamParseError(f"Unsupported SSE line: {line}")
|
||||
|
||||
event = await flush()
|
||||
if event is not None:
|
||||
yield event
|
||||
Reference in New Issue
Block a user