Enhance logging in LLMService: Update log messages for MoE dispatch and default provider usage, and add detailed logging before OpenRouter calls for improved traceability and debugging.

This commit is contained in:
Lars 2026-01-02 07:26:24 +01:00
parent d0eae8e43c
commit f62983b08f

View File

@ -140,11 +140,14 @@ class LLMService:
target_provider = profile.get("provider", target_provider)
target_model = profile.get("model", target_model)
target_temp = profile.get("temperature")
logger.debug(f"🎭 MoE Call: Profil '{profile_name}' -> {target_provider}")
logger.info(f"🎭 MoE Dispatch: Profil='{profile_name}' -> Provider='{target_provider}' | Model='{target_model}'")
else:
logger.warning(f"⚠️ Profil '{profile_name}' nicht in llm_profiles.yaml gefunden!")
# Fallback auf Standard-Provider falls nichts übergeben/definiert wurde
if not target_provider:
target_provider = self.settings.MINDNET_LLM_PROVIDER
logger.info(f" Kein Provider/Profil definiert. Nutze Default: {target_provider}")
if priority == "background":
async with LLMService._background_semaphore:
@ -242,6 +245,8 @@ class LLMService:
temperature: Optional[float] = None
) -> str:
model = model_override or self.settings.OPENROUTER_MODEL
# ERWEITERTES LOGGING VOR DEM CALL
logger.info(f"🛰️ OpenRouter Call: Model='{model}' | Temp={temperature}")
messages = []
if system: messages.append({"role": "system", "content": system})
messages.append({"role": "user", "content": prompt})