project-nomad/admin
Henry Estela 1e1da483e9
feat(AI): enable flash_attn by default and disable ollama cloud (#616)
New defaults:
OLLAMA_NO_CLOUD=1 - "Ollama can run in local only mode by disabling
Ollama’s cloud features. By turning off Ollama’s cloud features, you
will lose the ability to use Ollama’s cloud models and web search."
https://ollama.com/blog/web-search
https://docs.ollama.com/faq#how-do-i-disable-ollama%E2%80%99s-cloud-features
example output:
```
ollama run minimax-m2.7:cloud
Error: ollama cloud is disabled: remote model details are unavailable
```
This setting can be safely disabled as you have to click on a link to
login to ollama cloud and theres no real way to do that in nomad outside
of looking at the nomad_ollama logs.

This one can be disabled in settings in case theres a model out there
that doesn't play nice. but that doesnt seem necessary so far.
OLLAMA_FLASH_ATTENTION=1 - "Flash Attention is a feature of most modern
models that can significantly reduce memory usage as the context size
grows. "

Tested with llama3.2:
```
docker logs nomad_ollama --tail 1000 2>&1 |grep --color -i flash_attn
llama_context: flash_attn    = enabled
```

And with second_constantine/deepseek-coder-v2 with is based on
https://huggingface.co/lmstudio-community/DeepSeek-Coder-V2-Lite-Instruct-GGUF
which is a model that specifically calls out that you should disable
flash attention, but during testing it seems ollama can do this for you
automatically:
```
docker logs nomad_ollama --tail 1000 2>&1 |grep --color -i flash_attn
llama_context: flash_attn    = disabled
```
2026-04-01 19:52:11 -07:00
..
app feat(AI): enable flash_attn by default and disable ollama cloud (#616) 2026-04-01 19:52:11 -07:00
bin feat: curated content system overhaul 2026-02-11 15:44:46 -08:00
commands fix(Jobs): improved error handling and robustness 2026-04-02 02:49:33 +00:00
config fix(ui): reduce SSE reconnect churn and polling overhead on navigation 2026-03-20 11:46:10 -07:00
constants feat(AI): enable flash_attn by default and disable ollama cloud (#616) 2026-04-01 19:52:11 -07:00
database fix: bump default ollama and cyberchef versions 2026-03-25 16:30:35 -07:00
docs feat: switch all PNG images to WEBP (#575) 2026-04-01 15:06:15 -07:00
inertia feat(AI): enable flash_attn by default and disable ollama cloud (#616) 2026-04-01 19:52:11 -07:00
providers fix(Maps): static path resolution 2026-01-23 14:17:25 -08:00
public feat: switch all PNG images to WEBP (#575) 2026-04-01 15:06:15 -07:00
resources/views feat: switch all PNG images to WEBP (#575) 2026-04-01 15:06:15 -07:00
start feat(downloads): rich progress, friendly names, cancel, and live status (#554) 2026-04-01 15:55:13 -07:00
tests feat: initial commit 2025-06-29 15:51:08 -07:00
types feat(AI): enable flash_attn by default and disable ollama cloud (#616) 2026-04-01 19:52:11 -07:00
util feat: display model download progress 2026-02-06 16:22:23 -08:00
views feat: initial commit 2025-06-29 15:51:08 -07:00
.editorconfig feat: initial commit 2025-06-29 15:51:08 -07:00
.env.example feat: Add Windows Docker Desktop support for local development 2026-01-19 10:29:24 -08:00
ace.js feat: initial commit 2025-06-29 15:51:08 -07:00
adonisrc.ts feat: background job overhaul with bullmq 2025-12-06 23:59:01 -08:00
eslint.config.js feat: openwebui+ollama and zim management 2025-07-09 09:08:21 -07:00
package-lock.json build: regen lockfile 2026-04-01 17:22:30 +00:00
package.json feat(rag): add EPUB file support for Knowledge Base uploads (#257) 2026-04-01 17:22:30 +00:00
tailwind.config.ts feat: initial commit 2025-06-29 15:51:08 -07:00
tsconfig.json feat: initial commit 2025-06-29 15:51:08 -07:00
vite.config.ts feat: initial commit 2025-06-29 15:51:08 -07:00