feat: Phase 0c - activity_metrics.py module complete
All checks were successful
Deploy Development / deploy (push) Successful in 51s
Build Test / lint-backend (push) Successful in 1s
Build Test / build-frontend (push) Successful in 16s

Data Layer:
- get_activity_summary_data() - count, duration, calories, frequency
- get_activity_detail_data() - detailed activity log with all fields
- get_training_type_distribution_data() - category distribution with percentages

Placeholder Layer:
- get_activity_summary() - refactored to use data layer
- get_activity_detail() - refactored to use data layer
- get_trainingstyp_verteilung() - refactored to use data layer

All 3 activity data functions + 3 placeholder refactors complete.

Co-Authored-By: Claude Opus 4.6 <noreply@anthropic.com>
This commit is contained in:
Lars 2026-03-28 19:11:45 +01:00
parent e1d7670971
commit 6b2ad9fa1c
3 changed files with 332 additions and 60 deletions

View File

@ -31,9 +31,9 @@ from .utils import *
# Metric modules
from .body_metrics import *
from .nutrition_metrics import *
from .activity_metrics import *
# Future imports (will be added as modules are created):
# from .activity_metrics import *
# from .recovery_metrics import *
# from .health_metrics import *
# from .goals import *
@ -56,4 +56,9 @@ __all__ = [
'get_energy_balance_data',
'get_protein_adequacy_data',
'get_macro_consistency_data',
# Activity Metrics
'get_activity_summary_data',
'get_activity_detail_data',
'get_training_type_distribution_data',
]

View File

@ -0,0 +1,277 @@
"""
Activity Metrics Data Layer
Provides structured data for training tracking and analysis.
Functions:
- get_activity_summary_data(): Count, total duration, calories, averages
- get_activity_detail_data(): Detailed activity log entries
- get_training_type_distribution_data(): Training category percentages
All functions return structured data (dict) without formatting.
Use placeholder_resolver.py for formatted strings for AI.
Phase 0c: Multi-Layer Architecture
Version: 1.0
"""
from typing import Dict, List, Optional
from datetime import datetime, timedelta, date
from db import get_db, get_cursor, r2d
from data_layer.utils import calculate_confidence, safe_float, safe_int
def get_activity_summary_data(
profile_id: str,
days: int = 14
) -> Dict:
"""
Get activity summary statistics.
Args:
profile_id: User profile ID
days: Analysis window (default 14)
Returns:
{
"activity_count": int,
"total_duration_min": int,
"total_kcal": int,
"avg_duration_min": int,
"avg_kcal_per_session": int,
"sessions_per_week": float,
"confidence": str,
"days_analyzed": int
}
Migration from Phase 0b:
OLD: get_activity_summary(pid, days) formatted string
NEW: Structured data with all metrics
"""
with get_db() as conn:
cur = get_cursor(conn)
cutoff = (datetime.now() - timedelta(days=days)).strftime('%Y-%m-%d')
cur.execute(
"""SELECT
COUNT(*) as count,
SUM(duration_min) as total_min,
SUM(kcal_active) as total_kcal
FROM activity_log
WHERE profile_id=%s AND date >= %s""",
(profile_id, cutoff)
)
row = cur.fetchone()
if not row or row['count'] == 0:
return {
"activity_count": 0,
"total_duration_min": 0,
"total_kcal": 0,
"avg_duration_min": 0,
"avg_kcal_per_session": 0,
"sessions_per_week": 0.0,
"confidence": "insufficient",
"days_analyzed": days
}
activity_count = row['count']
total_min = safe_int(row['total_min'])
total_kcal = safe_int(row['total_kcal'])
avg_duration = int(total_min / activity_count) if activity_count > 0 else 0
avg_kcal = int(total_kcal / activity_count) if activity_count > 0 else 0
sessions_per_week = (activity_count / days * 7) if days > 0 else 0.0
confidence = calculate_confidence(activity_count, days, "general")
return {
"activity_count": activity_count,
"total_duration_min": total_min,
"total_kcal": total_kcal,
"avg_duration_min": avg_duration,
"avg_kcal_per_session": avg_kcal,
"sessions_per_week": round(sessions_per_week, 1),
"confidence": confidence,
"days_analyzed": days
}
def get_activity_detail_data(
profile_id: str,
days: int = 14,
limit: int = 50
) -> Dict:
"""
Get detailed activity log entries.
Args:
profile_id: User profile ID
days: Analysis window (default 14)
limit: Maximum entries to return (default 50)
Returns:
{
"activities": [
{
"date": date,
"activity_type": str,
"duration_min": int,
"kcal_active": int,
"hr_avg": int | None,
"training_category": str | None
},
...
],
"total_count": int,
"confidence": str,
"days_analyzed": int
}
Migration from Phase 0b:
OLD: get_activity_detail(pid, days) formatted string list
NEW: Structured array with all fields
"""
with get_db() as conn:
cur = get_cursor(conn)
cutoff = (datetime.now() - timedelta(days=days)).strftime('%Y-%m-%d')
cur.execute(
"""SELECT
date,
activity_type,
duration_min,
kcal_active,
hr_avg,
training_category
FROM activity_log
WHERE profile_id=%s AND date >= %s
ORDER BY date DESC
LIMIT %s""",
(profile_id, cutoff, limit)
)
rows = cur.fetchall()
if not rows:
return {
"activities": [],
"total_count": 0,
"confidence": "insufficient",
"days_analyzed": days
}
activities = []
for row in rows:
activities.append({
"date": row['date'],
"activity_type": row['activity_type'],
"duration_min": safe_int(row['duration_min']),
"kcal_active": safe_int(row['kcal_active']),
"hr_avg": safe_int(row['hr_avg']) if row.get('hr_avg') else None,
"training_category": row.get('training_category')
})
confidence = calculate_confidence(len(activities), days, "general")
return {
"activities": activities,
"total_count": len(activities),
"confidence": confidence,
"days_analyzed": days
}
def get_training_type_distribution_data(
profile_id: str,
days: int = 14
) -> Dict:
"""
Calculate training category distribution.
Args:
profile_id: User profile ID
days: Analysis window (default 14)
Returns:
{
"distribution": [
{
"category": str,
"count": int,
"percentage": float
},
...
],
"total_sessions": int,
"categorized_sessions": int,
"uncategorized_sessions": int,
"confidence": str,
"days_analyzed": int
}
Migration from Phase 0b:
OLD: get_trainingstyp_verteilung(pid, days) top 3 formatted
NEW: Complete distribution with percentages
"""
with get_db() as conn:
cur = get_cursor(conn)
cutoff = (datetime.now() - timedelta(days=days)).strftime('%Y-%m-%d')
# Get categorized activities
cur.execute(
"""SELECT
training_category,
COUNT(*) as count
FROM activity_log
WHERE profile_id=%s
AND date >= %s
AND training_category IS NOT NULL
GROUP BY training_category
ORDER BY count DESC""",
(profile_id, cutoff)
)
rows = cur.fetchall()
# Get total activity count (including uncategorized)
cur.execute(
"""SELECT COUNT(*) as total
FROM activity_log
WHERE profile_id=%s AND date >= %s""",
(profile_id, cutoff)
)
total_row = cur.fetchone()
total_sessions = total_row['total'] if total_row else 0
if not rows or total_sessions == 0:
return {
"distribution": [],
"total_sessions": total_sessions,
"categorized_sessions": 0,
"uncategorized_sessions": total_sessions,
"confidence": "insufficient",
"days_analyzed": days
}
categorized_count = sum(row['count'] for row in rows)
uncategorized_count = total_sessions - categorized_count
distribution = []
for row in rows:
count = row['count']
percentage = (count / total_sessions * 100) if total_sessions > 0 else 0
distribution.append({
"category": row['training_category'],
"count": count,
"percentage": round(percentage, 1)
})
confidence = calculate_confidence(categorized_count, days, "general")
return {
"distribution": distribution,
"total_sessions": total_sessions,
"categorized_sessions": categorized_count,
"uncategorized_sessions": uncategorized_count,
"confidence": confidence,
"days_analyzed": days
}

View File

@ -23,6 +23,11 @@ from data_layer.nutrition_metrics import (
get_nutrition_days_data,
get_protein_targets_data
)
from data_layer.activity_metrics import (
get_activity_summary_data,
get_activity_detail_data,
get_training_type_distribution_data
)
# ── Helper Functions ──────────────────────────────────────────────────────────
@ -223,25 +228,18 @@ def get_protein_ziel_high(profile_id: str) -> str:
def get_activity_summary(profile_id: str, days: int = 14) -> str:
"""Get activity summary for recent period."""
with get_db() as conn:
cur = get_cursor(conn)
cutoff = (datetime.now() - timedelta(days=days)).strftime('%Y-%m-%d')
cur.execute(
"""SELECT COUNT(*) as count,
SUM(duration_min) as total_min,
SUM(kcal_active) as total_kcal
FROM activity_log
WHERE profile_id=%s AND date >= %s""",
(profile_id, cutoff)
)
row = r2d(cur.fetchone())
"""
Get activity summary for recent period.
if row['count'] == 0:
Phase 0c: Refactored to use data_layer.activity_metrics.get_activity_summary_data()
This function now only FORMATS the data for AI consumption.
"""
data = get_activity_summary_data(profile_id, days)
if data['confidence'] == 'insufficient':
return f"Keine Aktivitäten in den letzten {days} Tagen"
avg_min = int(row['total_min'] / row['count']) if row['total_min'] else 0
return f"{row['count']} Einheiten in {days} Tagen (Ø {avg_min} min/Einheit, {int(row['total_kcal'] or 0)} kcal gesamt)"
return f"{data['activity_count']} Einheiten in {days} Tagen (Ø {data['avg_duration_min']} min/Einheit, {data['total_kcal']} kcal gesamt)"
def calculate_age(dob) -> str:
@ -263,54 +261,46 @@ def calculate_age(dob) -> str:
def get_activity_detail(profile_id: str, days: int = 14) -> str:
"""Get detailed activity log for analysis."""
with get_db() as conn:
cur = get_cursor(conn)
cutoff = (datetime.now() - timedelta(days=days)).strftime('%Y-%m-%d')
cur.execute(
"""SELECT date, activity_type, duration_min, kcal_active, hr_avg
FROM activity_log
WHERE profile_id=%s AND date >= %s
ORDER BY date DESC
LIMIT 50""",
(profile_id, cutoff)
)
rows = [r2d(r) for r in cur.fetchall()]
"""
Get detailed activity log for analysis.
if not rows:
Phase 0c: Refactored to use data_layer.activity_metrics.get_activity_detail_data()
This function now only FORMATS the data for AI consumption.
"""
data = get_activity_detail_data(profile_id, days)
if data['confidence'] == 'insufficient':
return f"Keine Aktivitäten in den letzten {days} Tagen"
# Format as readable list
# Format as readable list (max 20 entries to avoid token bloat)
lines = []
for r in rows:
hr_str = f" HF={r['hr_avg']}" if r.get('hr_avg') else ""
for activity in data['activities'][:20]:
hr_str = f" HF={activity['hr_avg']}" if activity['hr_avg'] else ""
lines.append(
f"{r['date']}: {r['activity_type']} ({r['duration_min']}min, {r.get('kcal_active', 0)}kcal{hr_str})"
f"{activity['date']}: {activity['activity_type']} "
f"({activity['duration_min']}min, {activity['kcal_active']}kcal{hr_str})"
)
return '\n'.join(lines[:20]) # Max 20 entries to avoid token bloat
return '\n'.join(lines)
def get_trainingstyp_verteilung(profile_id: str, days: int = 14) -> str:
"""Get training type distribution."""
with get_db() as conn:
cur = get_cursor(conn)
cutoff = (datetime.now() - timedelta(days=days)).strftime('%Y-%m-%d')
cur.execute(
"""SELECT training_category, COUNT(*) as count
FROM activity_log
WHERE profile_id=%s AND date >= %s AND training_category IS NOT NULL
GROUP BY training_category
ORDER BY count DESC""",
(profile_id, cutoff)
)
rows = [r2d(r) for r in cur.fetchall()]
"""
Get training type distribution.
if not rows:
Phase 0c: Refactored to use data_layer.activity_metrics.get_training_type_distribution_data()
This function now only FORMATS the data for AI consumption.
"""
data = get_training_type_distribution_data(profile_id, days)
if data['confidence'] == 'insufficient' or not data['distribution']:
return "Keine kategorisierten Trainings"
total = sum(r['count'] for r in rows)
parts = [f"{r['training_category']}: {int(r['count']/total*100)}%" for r in rows[:3]]
# Format top 3 categories with percentages
parts = [
f"{dist['category']}: {int(dist['percentage'])}%"
for dist in data['distribution'][:3]
]
return ", ".join(parts)