Fetching models from your API and loading chat adapter templates…
Used in chat display and in requests to the model. Character cards can use {{user}} for your name and {{char}} for the character’s name (like SillyTavern).
JSON templates from the adapters folder (KoboldCPP-style). None keeps standard OpenAI /v1/chat/completions with messages. When you pick an adapter, the UI builds one prompt string and calls /v1/completions — your backend must support that endpoint. Edit adapters/manifest.json to change the list.
Caps how long each assistant message can be. It does not limit the whole conversation; use Context in API requests below to trim what is sent as input.
The full chat is sent by default. Trim older turns here to save tokens; empty-send “continue” adds the line below only to the request (not your saved transcript).
Uses context_length / n_ctx from /v1/models when present. Otherwise the budget above is used. When trimming, room for the next reply is reserved using max reply tokens above.
Applies to message times and the conversation list. Automatic follows your browser (—).
Supports SillyTavern / Chub / chara_card_v2 and common exports.
This deletes all characters, conversations, and settings (including API keys) from this browser.
Private. Chats and settings stay in this browser unless you export them.
http://localhost:5001/v1).max_tokens sent to the API). It is not the context-window size for the whole chat.adapters/ so the request uses the right special tokens; it switches the client to the /v1/completions prompt API instead of chat messages.{{user}} and {{char}} in card text and messages are filled from that and the active character.Hosting, privacy, and legal notices are in .
Private. Chats and settings stay in this browser unless you export them. Markdown and syntax highlighting in messages are rendered only in your browser from bundled scripts (no third-party formatter or CDN).
The server that hosts these static files only sends you the WebUI (HTML, CSS, JavaScript). In normal use, your prompts and model replies are not sent to that host; your browser talks directly to whatever API base URL you configure in Settings (for example your own computer or another service you control).
Your chats and settings are stored in your browser unless you export them. Whoever runs the LLM / API server you point at may process and log traffic according to their own setup—you choose that endpoint.
You are responsible for your use of this tool, for the content you create, for obeying applicable laws and any terms of the APIs or models you use, and for any misuse. This interface is provided as-is, without warranty. To the extent allowed by law, the people who publish this hosted WebUI are not liable for user-generated content, third-party backends, or how others use the software.
This host only serves the interface. Your API calls go to the server you set in Settings, not to the operator of this page, unless you deliberately configure otherwise.
Retrograde WebUI runs in your browser. Chats and settings stay on this device unless you export them.
Your requests go to the API URL you set in Settings—not to whoever hosts these files.
More detail is in Help (top of the page).