(import "./config.ncl") & { llm = { provider = "ollama", model = "llama2", api_endpoint = "http://localhost:11434/api", generation = { temperature = 0.5, max_tokens = 1024, }, }, rag = { index_path = "~/.config/typedialog/ai/rag-index-dev", cache_size = 500, }, microservice.enable_cors = true, appearance.suggestion_confidence_threshold = 0.3, }