wechsel des LLM (--> Mistral")
This commit is contained in:
parent
c3ef65d069
commit
e512f768ff
|
|
@ -2,7 +2,7 @@
|
|||
app/config.py — zentrale Konfiguration (ENV → Settings)
|
||||
|
||||
Version:
|
||||
0.3.0 (WP-05: LLM & Chat Config)
|
||||
0.3.1 (WP-05: Switch default to Mistral for CPU inference)
|
||||
Stand:
|
||||
2025-12-08
|
||||
"""
|
||||
|
|
@ -25,7 +25,8 @@ class Settings:
|
|||
|
||||
# WP-05 LLM / Ollama
|
||||
OLLAMA_URL: str = os.getenv("MINDNET_OLLAMA_URL", "http://127.0.0.1:11434")
|
||||
LLM_MODEL: str = os.getenv("MINDNET_LLM_MODEL", "llama3") # oder 'mistral', 'gemma'
|
||||
# ÄNDERUNG: Standard auf 'mistral' gesetzt, da bereits lokal vorhanden
|
||||
LLM_MODEL: str = os.getenv("MINDNET_LLM_MODEL", "mistral")
|
||||
PROMPTS_PATH: str = os.getenv("MINDNET_PROMPTS_PATH", "config/prompts.yaml")
|
||||
|
||||
# API
|
||||
|
|
|
|||
Loading…
Reference in New Issue
Block a user