From fb1a55947dd94c2ff9e4d373cb2d50010caf636e Mon Sep 17 00:00:00 2001 From: Henry Estela Date: Fri, 20 Mar 2026 13:27:54 -0700 Subject: [PATCH] fix(ai-chat): dark/light mode for new settings box, improve settings UI --- admin/inertia/components/chat/index.tsx | 6 ++++++ admin/inertia/pages/settings/models.tsx | 21 ++++++++++----------- 2 files changed, 16 insertions(+), 11 deletions(-) diff --git a/admin/inertia/components/chat/index.tsx b/admin/inertia/components/chat/index.tsx index 2c1a983..db4ff6a 100644 --- a/admin/inertia/components/chat/index.tsx +++ b/admin/inertia/components/chat/index.tsx @@ -53,6 +53,7 @@ export default function Chat({ const activeSession = sessions.find((s) => s.id === activeSessionId) const { data: lastModelSetting } = useSystemSetting({ key: 'chat.lastModel', enabled }) + const { data: remoteOllamaUrlSetting } = useSystemSetting({ key: 'ai.remoteOllamaUrl', enabled }) const { data: installedModels = [], isLoading: isLoadingModels } = useQuery({ queryKey: ['installedModels'], @@ -363,6 +364,11 @@ export default function Chat({ {activeSession?.title || 'New Chat'}
+ {remoteOllamaUrlSetting?.value && ( + + Remote Connected + + )}
-
-

+

+

Connect to any OpenAI-compatible API server — Ollama, LM Studio, llama.cpp, and others are all supported. - For remote Ollama instances, the host must be started with OLLAMA_HOST=0.0.0.0. + For remote Ollama instances, the host must be started with OLLAMA_HOST=0.0.0.0.

- {props.models.settings.remoteOllamaUrl && ( -

- Currently configured: {props.models.settings.remoteOllamaUrl} -

- )}
-

- Model downloading is only supported when using an Ollama backend. If you are connected to an OpenAI API host (e.g. LM Studio), you will need to download models directly in that application. -

+