module pixeltable.functions.llama_cpp
Pixeltable UDFs for llama.cpp models. Provides integration with llama.cpp for running quantized language models locally, supporting chat completions and embeddings with GGUF format models.udf create_chat_completion()
model_path or repo_id must be provided; if model_path
is provided, then an optional repo_filename can also be specified.
For additional details, see the
llama_cpp create_chat_completions documentation.
Parameters:
messages(Json): A list of messages to generate a response for.model_path(String | None): Path to the model (if using a local model).repo_id(String | None): The Hugging Face model repo id (if using a pretrained model).repo_filename(String | None): A filename or glob pattern to match the model file in the repo (optional, if using a pretrained model).model_kwargs(Json | None): Additional keyword args for the llama_cppcreate_chat_completionsAPI, such asmax_tokens,temperature,top_p, andtop_k. For details, see the llama_cpp create_chat_completions documentation.