Response#
Response schema.
- class llama_index.core.response.schema.PydanticResponse(response: ~typing.Optional[~pydantic.main.BaseModel], source_nodes: ~typing.List[~llama_index.schema.NodeWithScore] = <factory>, metadata: ~typing.Optional[~typing.Dict[str, ~typing.Any]] = None)#
PydanticResponse object.
Returned if streaming=False.
- response#
The response text.
- Type
Optional[pydantic.main.BaseModel]
- get_formatted_sources(length: int = 100) str #
Get formatted sources text.
- class llama_index.core.response.schema.Response(response: ~typing.Optional[str], source_nodes: ~typing.List[~llama_index.schema.NodeWithScore] = <factory>, metadata: ~typing.Optional[~typing.Dict[str, ~typing.Any]] = None)#
Response object.
Returned if streaming=False.
- response#
The response text.
- Type
Optional[str]
- get_formatted_sources(length: int = 100) str #
Get formatted sources text.
- class llama_index.core.response.schema.StreamingResponse(response_gen: ~typing.Generator[str, None, None], source_nodes: ~typing.List[~llama_index.schema.NodeWithScore] = <factory>, metadata: ~typing.Optional[~typing.Dict[str, ~typing.Any]] = None, response_txt: ~typing.Optional[str] = None)#
StreamingResponse object.
Returned if streaming=True.
- response_gen#
The response generator.
- Type
Generator[str, None, None]
- get_formatted_sources(length: int = 100, trim_text: int = True) str #
Get formatted sources text.
- print_response_stream() None #
Print the response stream.