LlamaCPP๏
- pydantic model llama_index.llms.llama_cpp.LlamaCPP๏
Show JSON schema
{ "title": "LlamaCPP", "description": "Simple abstract base class for custom LLMs.\n\nSubclasses must implement the `__init__`, `complete`,\n `stream_complete`, and `metadata` methods.", "type": "object", "properties": { "callback_manager": { "title": "Callback Manager" }, "model_url": { "title": "Model Url", "description": "The URL llama-cpp model to download and use.", "type": "string" }, "model_path": { "title": "Model Path", "description": "The path to the llama-cpp model to use.", "type": "string" }, "temperature": { "title": "Temperature", "description": "The temperature to use for sampling.", "type": "number" }, "max_new_tokens": { "title": "Max New Tokens", "description": "The maximum number of tokens to generate.", "type": "integer" }, "context_window": { "title": "Context Window", "description": "The maximum number of context tokens for the model.", "type": "integer" }, "generate_kwargs": { "title": "Generate Kwargs", "description": "Kwargs used for generation.", "type": "object" }, "model_kwargs": { "title": "Model Kwargs", "description": "Kwargs used for model initialization.", "type": "object" }, "verbose": { "title": "Verbose", "description": "Whether to print verbose output.", "type": "boolean" } }, "required": [ "temperature", "max_new_tokens", "context_window", "verbose" ] }
- Config
arbitrary_types_allowed: bool = True
- Fields
- Validators
_validate_callback_manager
ยปcallback_manager
- field completion_to_prompt: Callable [Required]๏
The function to convert a completion to a prompt.
- field context_window: int [Required]๏
The maximum number of context tokens for the model.
- field generate_kwargs: Dict[str, Any] [Optional]๏
Kwargs used for generation.
- field max_new_tokens: int [Required]๏
The maximum number of tokens to generate.
- field messages_to_prompt: Callable [Required]๏
The function to convert messages to a prompt.
- field model_kwargs: Dict[str, Any] [Optional]๏
Kwargs used for model initialization.
- field model_path: Optional[str] = None๏
The path to the llama-cpp model to use.
- field model_url: Optional[str] = None๏
The URL llama-cpp model to download and use.
- field temperature: float [Required]๏
The temperature to use for sampling.
- field verbose: bool [Required]๏
Whether to print verbose output.
- chat(messages: Sequence[ChatMessage], **kwargs: Any) Any ๏
Chat endpoint for LLM.
- classmethod class_name() str ๏
Get class name.
- complete(*args: Any, **kwargs: Any) Any ๏
Completion endpoint for LLM.
- stream_chat(messages: Sequence[ChatMessage], **kwargs: Any) Any ๏
Streaming chat endpoint for LLM.
- stream_complete(*args: Any, **kwargs: Any) Any ๏
Streaming completion endpoint for LLM.
- property metadata: LLMMetadata๏
LLM metadata.