diff --git a/admin/inertia/components/chat/index.tsx b/admin/inertia/components/chat/index.tsx index 2c1a983..db4ff6a 100644 --- a/admin/inertia/components/chat/index.tsx +++ b/admin/inertia/components/chat/index.tsx @@ -53,6 +53,7 @@ export default function Chat({ const activeSession = sessions.find((s) => s.id === activeSessionId) const { data: lastModelSetting } = useSystemSetting({ key: 'chat.lastModel', enabled }) + const { data: remoteOllamaUrlSetting } = useSystemSetting({ key: 'ai.remoteOllamaUrl', enabled }) const { data: installedModels = [], isLoading: isLoadingModels } = useQuery({ queryKey: ['installedModels'], @@ -363,6 +364,11 @@ export default function Chat({ {activeSession?.title || 'New Chat'}
+
Connect to any OpenAI-compatible API server — Ollama, LM Studio, llama.cpp, and others are all supported.
- For remote Ollama instances, the host must be started with OLLAMA_HOST=0.0.0.0.
+ For remote Ollama instances, the host must be started with OLLAMA_HOST=0.0.0.0.
- Currently configured: {props.models.settings.remoteOllamaUrl} -
- )}- Model downloading is only supported when using an Ollama backend. If you are connected to an OpenAI API host (e.g. LM Studio), you will need to download models directly in that application. -
+