udf create_chat_completion()
model_path or repo_id must be provided; if model_path is provided, then an optional repo_filename can also be specified.
For additional details, see the llama_cpp create_chat_completions documentation.
Parameters:
messages(Json): A list of messages to generate a response for.model_path(String | None): Path to the model (if using a local model).repo_id(String | None): The Hugging Face model repo id (if using a pretrained model).repo_filename(String | None): A filename or glob pattern to match the model file in the repo (optional, if using a pretrained model).model_kwargs(Json | None): Additional keyword args for the llama_cppcreate_chat_completionsAPI, such asmax_tokens,temperature,top_p, andtop_k. For details, see the llama_cpp create_chat_completions documentation.