Response#

Response schema.

class llama_index.core.response.schema.PydanticResponse(response: ~typing.Optional[~pydantic.main.BaseModel], source_nodes: ~typing.List[~llama_index.schema.NodeWithScore] = <factory>, metadata: ~typing.Optional[~typing.Dict[str, ~typing.Any]] = None)#

PydanticResponse object.

Returned if streaming=False.

response#

The response text.

Type

Optional[pydantic.main.BaseModel]

get_formatted_sources(length: int = 100) str#

Get formatted sources text.

get_response() Response#

Get a standard response object.

class llama_index.core.response.schema.Response(response: ~typing.Optional[str], source_nodes: ~typing.List[~llama_index.schema.NodeWithScore] = <factory>, metadata: ~typing.Optional[~typing.Dict[str, ~typing.Any]] = None)#

Response object.

Returned if streaming=False.

response#

The response text.

Type

Optional[str]

get_formatted_sources(length: int = 100) str#

Get formatted sources text.

class llama_index.core.response.schema.StreamingResponse(response_gen: ~typing.Generator[str, None, None], source_nodes: ~typing.List[~llama_index.schema.NodeWithScore] = <factory>, metadata: ~typing.Optional[~typing.Dict[str, ~typing.Any]] = None, response_txt: ~typing.Optional[str] = None)#

StreamingResponse object.

Returned if streaming=True.

response_gen#

The response generator.

Type

Generator[str, None, None]

get_formatted_sources(length: int = 100, trim_text: int = True) str#

Get formatted sources text.

get_response() Response#

Get a standard response object.

print_response_stream() None#

Print the response stream.