import { NomadOllamaModel } from '../types/ollama.js' /** * Fallback basic recommended Ollama models in case fetching from the service fails. */ export const FALLBACK_RECOMMENDED_OLLAMA_MODELS: NomadOllamaModel[] = [ { name: 'llama3.1', description: 'Llama 3.1 is a new state-of-the-art model from Meta available in 8B, 70B and 405B parameter sizes.', estimated_pulls: '109.3M', id: '9fe9c575-e77e-4a51-a743-07359458ee71', first_seen: '2026-01-28T23:37:31.000+00:00', model_last_updated: '1 year ago', tags: [ { name: 'llama3.1:8b-text-q4_1', size: '5.1 GB', context: '128k', input: 'Text', }, ], }, { name: 'deepseek-r1', description: 'DeepSeek-R1 is a family of open reasoning models with performance approaching that of leading models, such as O3 and Gemini 2.5 Pro.', estimated_pulls: '77.2M', id: '0b566560-68a6-4964-b0d4-beb3ab1ad694', first_seen: '2026-01-28T23:37:31.000+00:00', model_last_updated: '7 months ago', tags: [ { name: 'deepseek-r1:1.5b', size: '1.1 GB', context: '128k', input: 'Text', }, ], }, { name: 'llama3.2', description: "Meta's Llama 3.2 goes small with 1B and 3B models.", estimated_pulls: '54.7M', id: 'c9a1bc23-b290-4501-a913-f7c9bb39c3ad', first_seen: '2026-01-28T23:37:31.000+00:00', model_last_updated: '1 year ago', tags: [ { name: 'llama3.2:1b-text-q2_K', size: '581 MB', context: '128k', input: 'Text', }, ], }, ] export const SYSTEM_PROMPTS = { default: ` Format all responses using markdown for better readability. Vanilla markdown or GitHub-flavored markdown is preferred. - Use **bold** and *italic* for emphasis. - Use code blocks with language identifiers for code snippets. - Use headers (##, ###) to organize longer responses. - Use bullet points or numbered lists for clarity. - Use tables when presenting structured data. `, rag_context: (context: string) => ` You have access to the following relevant information from the knowledge base. Use this context to provide accurate and informed responses when relevant: [Context] ${context} If the user's question is related to this context, incorporate it into your response. Otherwise, respond normally. `, }