adding the ability to configure ollama chat options and embeddings params via the config.toml file

This commit is contained in:
Owyn Richen 2025-02-05 13:33:17 -08:00
parent 46541e6c0c
commit ad0826111b
8 changed files with 1052 additions and 1388 deletions

View file

@ -11,4 +11,11 @@ GEMINI = "" # Gemini API key - sk-1234567890abcdef1234567890abcdef
[API_ENDPOINTS]
SEARXNG = "http://localhost:32768" # SearxNG API URL
OLLAMA = "" # Ollama API URL - http://host.docker.internal:11434
OLLAMA = "" # Ollama API URL - http://host.docker.internal:11434
[OLLAMA_CHAT_OPTIONS]
# maps to parameters found here: https://v03.api.js.langchain.com/interfaces/_langchain_ollama.ChatOllamaInput.html
numCtx = 2048 # the default, some models demand more
[OLLAMA_EMBEDDINGS_PARAMS]
# maps to parameters found here: https://v03.api.js.langchain.com/interfaces/_langchain_ollama.OllamaEmbeddingsParams.html