AI Assistant
The AI assistant operates within the browser interface and also provides a Command Line Interface (CLI) that can be accessed either from the host or within a space. The assistant requires access to an external Large Language Model (LLM) that supports the OpenAI API.
Enabling the Chat Interface
To enable the chat interface, add the following configuration to the knot.toml
file:
[server.chat]
ui_style = "avatar"
enabled = true
openai_endpoints = true
openai_base_url = "http://localhost:8085/v1/"
openai_api_key = ""
model = "Qwen3-4B-Thinking-2507-GGUF:Q4_0"
reasoning_effort = "low"
Configuration Options
ui_style
: Defines the style of the assistant’s chat icon. Options are:avatar
(default)icon
enabled
: Must be set totrue
to activate the web assistant.openai_endpoints
: If set totrue
, the knot server will expose an OpenAI-compatible endpoint at the/v1/
path. This allows any client to connect to the knot server, chat with the assistant, and use tools.openai_base_url
: Specifies the address of the server hosting the LLM.openai_api_key
: The API key required to authenticate with the LLM server.model
: The model to be used. The assistant’s performance depends on the selected model.reasoning_effort
: Controls the level of effort applied by the model for reasoning, there’s no default. Options are:none
low
medium
high
If you are using a non-thinking model then do not set reasoning_effort
in the configuration as usually this will stop the model from working.
Supported Models
The assistant is confirmed to work with the following models:
gpt-oss 20B
(via LM Studio)Gemini 2.5 Flash