To use a model hosted on a filesystem, specify the path to the model file or folder in the from field:
Supported formats include GGUF, GGML, and SafeTensor for large language models (LLMs) and ONNX for traditional machine learning (ML) models.
fromAn absolute or relative path to the model file or folder:
params (optional)| Param | Description |
|---|---|
model_type | The architecture to load the model as. Supported values: mistral, gemma, mixtral, llama, phi2, phi3, qwen2, gemma2, starcoder2, phi3.5moe, deepseekv2, deepseek |
tools | Which tools should be made available to the model. Set to auto to use all available tools. |
system_prompt | An additional system prompt used for all chat completions to this model. |
chat_template | Customizes the transformation of OpenAI chat messages into a character stream for the model. See Overriding the Chat Template. |
See Large Language Models for additional configuration options.
files (optional)The files field specifies additional files required by the model, such as tokenizer, configuration, and other files.
Note: The folder provided should contain all the expected files (see examples above).
Chat templates convert the OpenAI compatible chat messages (see format) and other components of a request into a stream of characters for the language model. It follows Jinja3 templating syntax.
Further details on chat templates can be found here.
messages: List of chat messages, in the OpenAI format.add_generation_prompt: Boolean flag whether to add a generation prompt.tools: List of callable tools, in the OpenAI format.:::warning[Limitations]
:::