import streamlit as st import requests import uuid import os import json import re import yaml import unicodedata from datetime import datetime from pathlib import Path from dotenv import load_dotenv # --- WP-19 GRAPH IMPORTS --- try: from streamlit_agraph import agraph, Node, Edge, Config from qdrant_client import QdrantClient, models except ImportError: st.error("Fehlende Bibliotheken! Bitte installiere: pip install streamlit-agraph qdrant-client") st.stop() # --- CONFIGURATION --- load_dotenv() API_BASE_URL = os.getenv("MINDNET_API_URL", "http://localhost:8002") CHAT_ENDPOINT = f"{API_BASE_URL}/chat" FEEDBACK_ENDPOINT = f"{API_BASE_URL}/feedback" INGEST_ANALYZE_ENDPOINT = f"{API_BASE_URL}/ingest/analyze" INGEST_SAVE_ENDPOINT = f"{API_BASE_URL}/ingest/save" HISTORY_FILE = Path("data/logs/search_history.jsonl") # Qdrant Config (Direct Access for Graph) QDRANT_URL = os.getenv("QDRANT_URL", "http://localhost:6333") QDRANT_KEY = os.getenv("QDRANT_API_KEY", None) if QDRANT_KEY == "": QDRANT_KEY = None COLLECTION_PREFIX = os.getenv("COLLECTION_PREFIX", "mindnet") # Timeout Strategy timeout_setting = os.getenv("MINDNET_API_TIMEOUT") or os.getenv("MINDNET_LLM_TIMEOUT") API_TIMEOUT = float(timeout_setting) if timeout_setting else 300.0 # --- PAGE SETUP --- st.set_page_config(page_title="mindnet v2.6", page_icon="🧠", layout="wide") # --- CSS STYLING --- st.markdown(""" """, unsafe_allow_html=True) # --- SESSION STATE --- if "messages" not in st.session_state: st.session_state.messages = [] if "user_id" not in st.session_state: st.session_state.user_id = str(uuid.uuid4()) # --- GRAPH STYLING CONFIG (WP-19) --- GRAPH_COLORS = { "project": "#ff9f43", # Orange "concept": "#54a0ff", # Blau "decision": "#5f27cd", # Lila "risk": "#ff6b6b", # Rot "person": "#1dd1a1", # Grün "experience": "#feca57",# Gelb "default": "#8395a7" # Grau } EDGE_COLORS = { "depends_on": "#ff6b6b", # Rot (Blocker) "blocks": "#ee5253", # Dunkelrot "related_to": "#c8d6e5", # Hellgrau "next": "#54a0ff", # Blau "derived_from": "#ff9ff3"# Pink } # --- HELPER FUNCTIONS --- def slugify(value): if not value: return "" value = str(value).lower() replacements = {'ä': 'ae', 'ö': 'oe', 'ü': 'ue', 'ß': 'ss', '&': 'und', '+': 'und'} for k, v in replacements.items(): value = value.replace(k, v) value = unicodedata.normalize('NFKD', value).encode('ascii', 'ignore').decode('ascii') value = re.sub(r'[^\w\s-]', '', value).strip() return re.sub(r'[-\s]+', '-', value) def normalize_meta_and_body(meta, body): ALLOWED_KEYS = {"title", "type", "status", "tags", "id", "created", "updated", "aliases", "lang"} clean_meta = {} extra_content = [] if "titel" in meta and "title" not in meta: meta["title"] = meta.pop("titel") tag_candidates = ["tags", "emotionale_keywords", "keywords", "schluesselwoerter"] all_tags = [] for key in tag_candidates: if key in meta: val = meta[key] if isinstance(val, list): all_tags.extend(val) elif isinstance(val, str): all_tags.extend([t.strip() for t in val.split(",")]) for key, val in meta.items(): if key in ALLOWED_KEYS: clean_meta[key] = val elif key in tag_candidates: pass else: if val and isinstance(val, str): header = key.replace("_", " ").title() extra_content.append(f"## {header}\n{val}\n") if all_tags: clean_tags = [] for t in all_tags: t_clean = str(t).replace("#", "").strip() if t_clean: clean_tags.append(t_clean) clean_meta["tags"] = list(set(clean_tags)) if extra_content: new_section = "\n".join(extra_content) final_body = f"{new_section}\n{body}" else: final_body = body return clean_meta, final_body def parse_markdown_draft(full_text): clean_text = full_text.strip() pattern_block = r"```(?:markdown|md|yaml)?\s*(.*?)\s*```" match_block = re.search(pattern_block, clean_text, re.DOTALL | re.IGNORECASE) if match_block: clean_text = match_block.group(1).strip() meta = {} body = clean_text yaml_str = "" parts = re.split(r"^---+\s*$", clean_text, maxsplit=2, flags=re.MULTILINE) if len(parts) >= 3: yaml_str = parts[1] body = parts[2] elif clean_text.startswith("---"): fallback_match = re.search(r"^---\s*(.*?)(?=\n#)", clean_text, re.DOTALL | re.MULTILINE) if fallback_match: yaml_str = fallback_match.group(1) body = clean_text.replace(f"---{yaml_str}", "", 1).strip() if yaml_str: yaml_str_clean = yaml_str.replace("#", "") try: parsed = yaml.safe_load(yaml_str_clean) if isinstance(parsed, dict): meta = parsed except Exception as e: print(f"YAML Parsing Warning: {e}") if not meta.get("title"): h1_match = re.search(r"^#\s+(.*)$", body, re.MULTILINE) if h1_match: meta["title"] = h1_match.group(1).strip() if meta.get("type") == "draft": meta["status"] = "draft" meta["type"] = "experience" return normalize_meta_and_body(meta, body) def build_markdown_doc(meta, body): if "id" not in meta or meta["id"] == "generated_on_save": raw_title = meta.get('title', 'note') clean_slug = slugify(raw_title)[:50] or "note" meta["id"] = f"{datetime.now().strftime('%Y%m%d')}-{clean_slug}" meta["updated"] = datetime.now().strftime("%Y-%m-%d") ordered_meta = {} prio_keys = ["id", "type", "title", "status", "tags"] for k in prio_keys: if k in meta: ordered_meta[k] = meta.pop(k) ordered_meta.update(meta) try: yaml_str = yaml.dump(ordered_meta, default_flow_style=None, sort_keys=False, allow_unicode=True).strip() except: yaml_str = "error: generating_yaml" return f"---\n{yaml_str}\n---\n\n{body}" def load_history_from_logs(limit=10): queries = [] if HISTORY_FILE.exists(): try: with open(HISTORY_FILE, "r", encoding="utf-8") as f: lines = f.readlines() for line in reversed(lines): try: entry = json.loads(line) q = entry.get("query_text") if q and q not in queries: queries.append(q) if len(queries) >= limit: break except: continue except: pass return queries # --- WP-19 GRAPH SERVICE --- class GraphExplorerService: def __init__(self, url, api_key=None, prefix="mindnet"): self.client = QdrantClient(url=url, api_key=api_key) self.prefix = prefix self.notes_col = f"{prefix}_notes" self.chunks_col = f"{prefix}_chunks" self.edges_col = f"{prefix}_edges" def get_ego_graph(self, center_note_id: str): """Erzeugt einen Ego-Graphen (Node + Nachbarn) für die Visualisierung.""" nodes = {} # id -> Node Object edges_list = [] # List of Edge Objects # 1. Zentrale Note laden center_note = self._fetch_note(center_note_id) if not center_note: return [], [] self._add_node(nodes, center_note, is_center=True) # 2. Chunks der Note finden (Source Chunks) scroll_filter = models.Filter( must=[models.FieldCondition(key="note_id", match=models.MatchValue(value=center_note_id))] ) chunks, _ = self.client.scroll( collection_name=self.chunks_col, scroll_filter=scroll_filter, limit=100, with_payload=True ) chunk_ids = [c.id for c in chunks] # 3. Kanten finden if chunk_ids: edge_filter = models.Filter( must=[models.FieldCondition(key="source_id", match=models.MatchAny(any=chunk_ids))] ) raw_edges, _ = self.client.scroll( collection_name=self.edges_col, scroll_filter=edge_filter, limit=100, with_payload=True ) # 4. Targets auflösen for re in raw_edges: payload = re.payload target_chunk_id = payload.get("target_id") kind = payload.get("kind") provenance = payload.get("provenance", "explicit") target_note = self._resolve_note_from_chunk(target_chunk_id) if target_note and target_note['note_id'] != center_note_id: self._add_node(nodes, target_note) # Styling color = EDGE_COLORS.get(kind, "#bdc3c7") is_smart = provenance != "explicit" and provenance != "rule" label = f"{kind}" if is_smart: label += " 🤖" edges_list.append(Edge( source=center_note_id, target=target_note['note_id'], label=label, color=color, dashes=is_smart, title=f"Provenance: {provenance}" )) return list(nodes.values()), edges_list def _fetch_note(self, note_id): res, _ = self.client.scroll( collection_name=self.notes_col, scroll_filter=models.Filter(must=[models.FieldCondition(key="note_id", match=models.MatchValue(value=note_id))]), limit=1, with_payload=True ) return res[0].payload if res else None def _resolve_note_from_chunk(self, chunk_id_or_title): if "#" in chunk_id_or_title: res = self.client.retrieve(collection_name=self.chunks_col, ids=[chunk_id_or_title], with_payload=True) if res: parent_id = res[0].payload.get("note_id") return self._fetch_note(parent_id) else: # Versuch: Direkter Match auf Note Titel (für WikiLinks) # In Production sollte das optimiert werden (Cache) res, _ = self.client.scroll( collection_name=self.notes_col, scroll_filter=models.Filter(must=[models.FieldCondition(key="title", match=models.MatchValue(value=chunk_id_or_title))]), limit=1, with_payload=True ) return res[0].payload if res else None return None def _add_node(self, node_dict, note_payload, is_center=False): nid = note_payload.get("note_id") if nid in node_dict: return ntype = note_payload.get("type", "default") color = GRAPH_COLORS.get(ntype, GRAPH_COLORS["default"]) size = 35 if is_center else 20 node_dict[nid] = Node( id=nid, label=note_payload.get("title", nid), size=size, color=color, shape="dot", title=f"Type: {ntype}\nTags: {note_payload.get('tags')}", font={'color': 'black'} ) # Init Graph Service graph_service = GraphExplorerService(QDRANT_URL, QDRANT_KEY, COLLECTION_PREFIX) # --- API CLIENT --- def send_chat_message(message: str, top_k: int, explain: bool): try: response = requests.post( CHAT_ENDPOINT, json={"message": message, "top_k": top_k, "explain": explain}, timeout=API_TIMEOUT ) response.raise_for_status() return response.json() except Exception as e: return {"error": str(e)} def analyze_draft_text(text: str, n_type: str): try: response = requests.post(INGEST_ANALYZE_ENDPOINT, json={"text": text, "type": n_type}, timeout=15) response.raise_for_status() return response.json() except Exception as e: return {"error": str(e)} def save_draft_to_vault(markdown_content: str, filename: str = None): try: response = requests.post(INGEST_SAVE_ENDPOINT, json={"markdown_content": markdown_content, "filename": filename}, timeout=API_TIMEOUT) response.raise_for_status() return response.json() except Exception as e: return {"error": str(e)} def submit_feedback(query_id, node_id, score, comment=None): try: requests.post(FEEDBACK_ENDPOINT, json={"query_id": query_id, "node_id": node_id, "score": score, "comment": comment}, timeout=2) st.toast(f"Feedback ({score}) gesendet!") except: pass # --- UI COMPONENTS --- def render_sidebar(): with st.sidebar: st.title("🧠 mindnet") st.caption("v2.6 | WP-19 Graph View") mode = st.radio("Modus", ["💬 Chat", "📝 Manueller Editor", "🕸️ Graph Explorer"], index=0) st.divider() st.subheader("⚙️ Settings") top_k = st.slider("Quellen (Top-K)", 1, 10, 5) explain = st.toggle("Explanation Layer", True) st.divider() st.subheader("🕒 Verlauf") for q in load_history_from_logs(8): if st.button(f"🔎 {q[:25]}...", key=f"hist_{q}", use_container_width=True): st.session_state.messages.append({"role": "user", "content": q}) st.rerun() return mode, top_k, explain def render_draft_editor(msg): if "query_id" not in msg or not msg["query_id"]: msg["query_id"] = str(uuid.uuid4()) qid = msg["query_id"] key_base = f"draft_{qid}" # State Keys data_meta_key = f"{key_base}_data_meta" data_sugg_key = f"{key_base}_data_suggestions" widget_body_key = f"{key_base}_widget_body" data_body_key = f"{key_base}_data_body" # INIT STATE if f"{key_base}_init" not in st.session_state: meta, body = parse_markdown_draft(msg["content"]) if "type" not in meta: meta["type"] = "default" if "title" not in meta: meta["title"] = "" tags = meta.get("tags", []) meta["tags_str"] = ", ".join(tags) if isinstance(tags, list) else str(tags) st.session_state[data_meta_key] = meta st.session_state[data_sugg_key] = [] st.session_state[data_body_key] = body.strip() st.session_state[f"{key_base}_wdg_title"] = meta["title"] st.session_state[f"{key_base}_wdg_type"] = meta["type"] st.session_state[f"{key_base}_wdg_tags"] = meta["tags_str"] st.session_state[f"{key_base}_init"] = True # RESURRECTION if widget_body_key not in st.session_state and data_body_key in st.session_state: st.session_state[widget_body_key] = st.session_state[data_body_key] # CALLBACKS def _sync_meta(): meta = st.session_state[data_meta_key] meta["title"] = st.session_state.get(f"{key_base}_wdg_title", "") meta["type"] = st.session_state.get(f"{key_base}_wdg_type", "default") meta["tags_str"] = st.session_state.get(f"{key_base}_wdg_tags", "") st.session_state[data_meta_key] = meta def _sync_body(): st.session_state[data_body_key] = st.session_state[widget_body_key] def _insert_text(text_to_insert): current = st.session_state.get(widget_body_key, "") new_text = f"{current}\n\n{text_to_insert}" st.session_state[widget_body_key] = new_text st.session_state[data_body_key] = new_text def _remove_text(text_to_remove): current = st.session_state.get(widget_body_key, "") new_text = current.replace(text_to_remove, "").strip() st.session_state[widget_body_key] = new_text st.session_state[data_body_key] = new_text # UI LAYOUT st.markdown(f'
{link_text}