Innovation Features

30 research-backed innovations that make PiSovereign the most advanced self-hosted AI assistant platform.

PiSovereign goes beyond basic chat and retrieval with 30 innovation features spanning advanced AI, post-quantum security, privacy-preserving analytics, decentralized computing, and edge intelligence. All features are implemented following Clean Architecture principles — each lives in the correct architectural layer and integrates via ports and adapters.

All innovation features are enabled by default and can be individually disabled via config.toml. Features requiring external infrastructure (mesh peers, carbon API, wearable data) gracefully no-op when infrastructure is absent.

All 30 features communicate through a unified Sovereign Intelligence Engine signal bus (tokio::sync::broadcast), enabling emergent cross-feature intelligence. See the Sovereign Intelligence Engine developer guide for details.


Overview

Generation 1 — Core Innovation Features

#FeatureCategoryPhaseKey Benefit
1Post-Quantum CryptographySecurity2Quantum-resistant encryption
2Local Continual Learning (LoRA)AI / Privacy3Privacy-preserving model personalization
3Neuromorphic Memory ArchitectureAI / CogSci3Human-like memory consolidation
4Causal Reasoning EngineAI / Reasoning3“Why” and “what if” questions
5Homomorphic Semantic SearchSecurity / Privacy2Zero-knowledge RAG
6Autonomous Red Team AgentSecurity / AI Safety2Continuous self-adversarial testing
7Cognitive Load Adaptive ResponsesUX / CogSci1Context-aware response verbosity
8Ebbinghaus Decay Knowledge GraphsKnowledge1Spaced repetition memory
9Multi-Agent SwarmAI / Safety4Formally verified multi-agent collaboration
10Explainable AI DashboardTransparency4Full decision provenance
11Edge Model DistillationPerformance4Hardware-specific optimized models
12Differential Privacy AnalyticsPrivacy1ε-differential privacy guarantees
13Bayesian Self-OptimizationPerformance3Automatic parameter tuning
14Cross-Modal Reasoning FusionAI / Multi-Modal4Unified cross-modality search
15Digital Sovereignty ScorePrivacy / UX1Quantified privacy posture

Generation 2 — Sovereign Intelligence Features

#FeatureCategoryPhaseKey Benefit
16Sovereign Intelligence EngineArchitectureUnified signal bus connecting all features
17Cognitive Load-Adaptive Response ShapingUX / CogSciQuick Win3 response formats based on 5-signal estimation
18Predictive Cognitive Pre-CachingPerformance / AIQuick WinBayesian prediction, thermal-aware pre-generation
19Temporal Knowledge Decay VisualizationUX / KnowledgeQuick WinInteractive Three.js WebGL knowledge graph
20Adversarial Self-Audit Immune SystemSecurityQuick WinAuto-hardening prompt defense via hot-reload
21Energy-Aware Inference SchedulingEdge / Green AIFoundationThermal + carbon-aware inference
22Formal ε-Privacy BudgetPrivacyFoundationUser-visible daily privacy budget
23Neuroplastic Model RoutingAI / OptimizationFoundationLinUCB contextual bandit per user
24Affective ComputingAI / CogSciFoundationEmotional state detection + tone modulation
25Self-Evolving Prompt OptimizationAI / Meta-LearningFoundationThompson Sampling prompt evolution
26Embodied Context AnchoringAI / SpatialMoonshotRoom-linked spatial memory
27Causal Counterfactual ExplanationsXAI / ComplianceMoonshotHeuristic “what if” decision explanations
28Zero-Knowledge Decision ProofsSecurity / ComplianceMoonshotSHA-256 Merkle-tree proofs
29Quantum-Inspired Approximate Search (HDC)Performance / EdgeMoonshot10,000-bit binary vector pre-filter
30Federated Sovereign LearningDecentralized / PrivacyMoonshotLoRA sync with differential privacy
31Sovereign Mesh NetworkDecentralized / EdgeMoonshotP2P inference fallback

Phase 1 — Foundation

8. Ebbinghaus Decay Knowledge Graphs

Category: Knowledge Management · Cognitive Science

Replaces linear memory decay with Ebbinghaus forgetting curves and adds temporal versioning to knowledge graph edges.

How it works:

  • Retention formula: R = e^(-t/S) where t is elapsed time and S (stability) increases with each successful recall
  • Spaced repetition: Memories strengthen when accessed, with stability multiplied by a configurable recall_bonus factor (default: 1.5×)
  • Temporal versioning: Knowledge edges carry valid_from and valid_until timestamps, enabling reasoning about how facts change over time

Key files:

FilePurpose
crates/domain/src/services/ebbinghaus.rsPure retention() function implementing R = e^(-t/S)
crates/domain/src/entities/memory.rsstability and recall_count fields
crates/domain/src/entities/knowledge_edge.rsTemporal versioning fields
crates/infrastructure/src/config/memory.rsebbinghaus_base_stability, recall_bonus settings
migrations/15_ebbinghaus_memory.sqlDatabase schema for new fields

Configuration:

[memory]
ebbinghaus_base_stability = 1.0   # Base stability in days
recall_bonus = 1.5                 # Stability multiplier per recall

12. Differential Privacy Analytics

Category: Privacy · Mathematics

Adds formal ε-differential privacy to all internal analytics and learning signals using Laplace and Gaussian noise mechanisms.

How it works:

  • Laplace mechanism: Adds calibrated noise to scalar values with configurable sensitivity
  • Gaussian mechanism: For approximate (ε,δ)-differential privacy when needed
  • Privacy budget tracking: Automatically tracks cumulative privacy loss with configurable reset intervals
  • All internal metrics, memory statistics, and routing analytics are privatized before exposure

Key files:

FilePurpose
crates/domain/src/services/differential_privacy.rsLaplace/Gaussian noise functions
crates/domain/src/value_objects/privacy_budget.rsPrivacyBudget value object (ε, δ)
crates/application/src/ports/privacy_port.rsPrivacyPort trait
crates/infrastructure/src/adapters/differential_privacy_adapter.rsThread-safe budget tracking
crates/infrastructure/src/config/privacy.rsε, δ, budget reset interval

Configuration:

[privacy]
epsilon = 1.0
delta = 1e-5
budget_reset_interval = "daily"

7. Cognitive Load Adaptive Responses

Category: UX · Cognitive Science

Analyzes interaction patterns to infer cognitive load and dynamically adjusts response verbosity, complexity, and model tier.

How it works:

  • Temporal signals (20% weight): Time of day, response latency patterns
  • Linguistic signals (40% weight): Vocabulary complexity, typo frequency, message length trends
  • Contextual signals (40% weight): Open task count, calendar density, recent interaction frequency
  • The resulting CognitiveLoadScore (0.0–1.0) influences model tier selection and injects cognitive load hints into the system prompt

Key files:

FilePurpose
crates/domain/src/services/cognitive_load.rsPure scoring logic
crates/domain/src/value_objects/cognitive_load.rsCognitiveLoadScore, CognitiveSignals
crates/application/src/services/cognitive_load_service.rsSignal collection and scoring
crates/application/src/ports/cognitive_load_port.rsCognitiveLoadPort trait
crates/infrastructure/src/config/model_routing.rscognitive_load_weight (default: 0.3)

Configuration:

[model_routing]
cognitive_load_weight = 0.3   # How much cognitive load influences model selection

15. Digital Sovereignty Score & Dashboard

Category: Privacy · UX · Compliance

Quantifies the user’s privacy posture as a real-time 0–100 score across four dimensions.

Scoring dimensions (each 0–25 points):

DimensionWhat It Measures
Data LocalityPercentage of queries processed fully locally vs. requiring external APIs
Encryption CoveragePercentage of data at rest and in transit protected by encryption
Secret ManagementVault usage, key rotation frequency, credential hygiene
Network IsolationDocker network segmentation, exposed ports, TLS configuration

Key files:

FilePurpose
crates/domain/src/value_objects/sovereignty_score.rsSovereigntyScore with 4 sub-scores
crates/domain/src/services/sovereignty.rsPure scoring logic per dimension
crates/application/src/services/sovereignty_service.rsSystem state collection
crates/infrastructure/src/adapters/sovereignty_adapter.rsLive system introspection
crates/presentation_http/src/handlers/system.rsGET /api/sovereignty endpoint

API endpoint:

GET /api/sovereignty
→ { "total": 78, "data_locality": 22, "encryption_coverage": 20, "secret_management": 18, "network_isolation": 18 }

Phase 2 — Security

1. Post-Quantum Cryptography

Category: Security · Cryptography

Adds NIST-standardized post-quantum algorithms alongside existing ChaCha20-Poly1305 symmetric encryption.

Algorithms:

AlgorithmStandardPurpose
ML-KEM-768FIPS 203 (CRYSTALS-Kyber)Key encapsulation — quantum-resistant key exchange
ML-DSA-65FIPS 204 (CRYSTALS-Dilithium)Digital signatures — quantum-resistant authentication

Hybrid mode: PQC key exchange derives a shared key, then ChaCha20-Poly1305 handles bulk encryption — combining quantum resistance with proven symmetric performance.

Key files:

FilePurpose
crates/infrastructure/src/adapters/pqc_adapter.rsPqcAdapter implementing ML-KEM + ML-DSA
crates/domain/src/value_objects/pqc_key_pair.rsPqcKeyPair, PqcCiphertext, PqcSignature value objects
crates/infrastructure/src/config/pqc.rsAlgorithm selection and hybrid mode toggle

Configuration:

[pqc]
enabled = false                # Enable post-quantum cryptography
kem_algorithm = "ML-KEM-768"
sig_algorithm = "ML-DSA-65"
hybrid_mode = true             # PQC key exchange + ChaCha20-Poly1305 bulk encryption

6. Autonomous Red Team Agent

Category: Security · AI Safety

Background service that continuously generates adversarial prompts to test prompt injection defenses and auto-updates detection rules.

How it works:

  1. Loads attack catalog from OWASP LLM Top 10 patterns (GCG, AutoDAN, TreeOfThought, MultiTurn, EncodingBypass, DelimiterInjection, RolePlay)
  2. Uses the local LLM to generate novel variations of known attacks
  3. Tests each attack against the prompt sanitizer and records pass/fail
  4. Computes a defense effectiveness score (0–100)
  5. Automatically adds newly discovered bypassing patterns to the sanitizer’s detection rules
  6. Applies differential privacy noise to reported metrics

Key files:

FilePurpose
crates/application/src/services/red_team_service.rsAttack orchestration and scoring
crates/application/src/ports/red_team_port.rsRedTeamPort trait
crates/infrastructure/src/adapters/red_team_adapter.rsLLM-based attack generation
crates/presentation_http/src/handlers/red_team.rsGET /api/red-team/report
crates/presentation_http/src/tasks/red_team.rsScheduled background task

Configuration:

[red_team]
enabled = false
schedule = "0 0 * * 0"        # Weekly (Sunday midnight)
max_attacks_per_run = 100
auto_update_rules = true

API endpoint:

GET /api/red-team/report
→ { "defense_score": 94, "attacks_tested": 87, "attacks_blocked": 82, ... }

Category: Security · Privacy · Vector Search

Enables cosine similarity computation on encrypted embeddings using Partially Homomorphic Encryption, so plaintext vectors never exist in the database.

How it works:

  • Embeddings are encrypted before storage using a lattice-based scheme
  • Approximate cosine similarity is computed directly on ciphertext
  • Decrypted results exist only in volatile Moka L1 cache — never persisted in plaintext
  • Falls back to standard vector search when disabled for maximum performance

Key files:

FilePurpose
crates/domain/src/services/homomorphic.rsEncrypted vector operations
crates/domain/src/value_objects/encrypted_embedding.rsEncryptedEmbedding newtype
crates/application/src/ports/homomorphic_port.rsHomomorphicPort trait
crates/infrastructure/src/adapters/homomorphic_adapter.rsPHE implementation
migrations/16_encrypted_embeddings.sqlencrypted_embedding BYTEA columns

Configuration:

[homomorphic]
enabled = false
scheme = "lattice"
security_level = 128           # Bit security level

Phase 3 — AI Core

3. Neuromorphic Memory Architecture

Category: AI · Cognitive Science · Knowledge Management

Replaces flat memory storage with three biologically-inspired stores mirroring human hippocampal cognition (Complementary Learning Systems theory).

Memory stores:

StorePurposeAnalogy
EpisodicRecent interactions stored verbatim with high fidelityWorking / short-term memory
SemanticGeneralized knowledge distilled from episodic memoriesLong-term factual memory
ProceduralLearned action patterns and habits“Muscle memory” for routines

Consolidation process: During low-usage periods (configurable cron schedule, default: 3 AM server-local time), the system uses the local LLM to summarize groups of related episodic memories into semantic facts. Procedural patterns are detected from repeating action sequences.

Key files:

FilePurpose
crates/domain/src/services/memory_consolidation.rsConsolidation candidate selection and merge logic
crates/domain/src/services/procedural_detector.rsRepeating action pattern detection
crates/domain/src/value_objects/memory_store_type.rsMemoryStoreType enum with transition rules
crates/application/src/services/consolidation_service.rsLLM-powered “sleep phase” consolidation
crates/presentation_http/src/tasks/memory_consolidation.rsScheduled background task
migrations/17_neuromorphic_memory.sqlstore_type, consolidated_from, procedural_trigger columns

4. Causal Reasoning Engine

Category: AI · Reasoning · Knowledge Graphs

Adds causal edge annotations, counterfactual query processing, and intervention reasoning to the knowledge graph, based on Judea Pearl’s Structural Causal Models (SCMs).

Capabilities:

  • Causal chain finding: BFS traversal with causal edge filtering to find cause-effect paths
  • Counterfactual queries: “What would have happened if I had sent that email yesterday?”
  • Intervention reasoning: Propagate causal weights to predict effects of planned actions
  • Causal relation types: CausedBy, Prevents, Enables, Correlates with confidence scores

Key files:

FilePurpose
crates/domain/src/services/causal_reasoning.rsd-separation, do-calculus, counterfactual evaluation
crates/domain/src/value_objects/temporal_order.rsTemporalOrder { Before, After, Concurrent, Unknown }
crates/domain/src/value_objects/causal_query.rsQuery types for causal reasoning
crates/application/src/services/causal_service.rsOrchestration with LLM-assisted interpretation
migrations/18_causal_reasoning.sqlCausal columns and causal_models table

2. Local Continual Learning (LoRA)

Category: AI · Machine Learning · Privacy

Implements local LoRA (Low-Rank Adaptation) fine-tuning that personalizes models from user feedback without any data leaving the device.

Pipeline:

  1. Feedback collection: User corrections, style preferences, and domain vocabulary via POST /api/training/feedback
  2. DP-SGD: Gradient perturbation with differential privacy during training
  3. EWC regularization: Elastic Weight Consolidation prevents catastrophic forgetting of general knowledge
  4. Adapter deployment: LoRA adapters (~1–5% of model parameters) are created via Ollama’s create API

Key files:

FilePurpose
crates/ai_core/src/lora.rsAdapter lifecycle management
crates/ai_core/src/training.rsTraining pipeline with DP-SGD + EWC
crates/domain/src/entities/training_sample.rsTrainingSample entity
crates/domain/src/services/ewc.rsElastic Weight Consolidation logic
crates/application/src/services/continual_learning_service.rsOrchestration
crates/presentation_http/src/handlers/training.rsPOST /api/training/feedback, GET /api/training/status
migrations/19_training_samples.sqlTraining data storage

Configuration:

[lora]
enabled = false
learning_rate = 0.0001
rank = 8
min_samples = 50               # Minimum feedback samples before training

13. Bayesian Self-Optimization

Category: Performance · Machine Learning

Automatically tunes system parameters using Bayesian Optimization with a Gaussian Process surrogate model.

Tunable parameters: Cache TTL, similarity thresholds, circuit breaker timings, token budgets, model routing thresholds, rate limits — each with safety bounds.

Objective function: Weighted combination of response latency, user feedback scores, cache hit rates, and memory usage — all privatized via PrivacyPort.

Key files:

FilePurpose
crates/domain/src/services/bayesian_optimization.rsGP surrogate + Expected Improvement acquisition
crates/domain/src/value_objects/tunable_parameter.rsTunableParameter with safety bounds
crates/application/src/services/self_optimization_service.rsMetric collection and config hot-reload
crates/infrastructure/src/adapters/bayesian_optimizer_adapter.rsnalgebra-based GP implementation
crates/presentation_http/src/handlers/optimization.rsGET /api/optimization/status

Configuration:

[optimization]
enabled = false
schedule = "0 0 * * 0"        # Weekly
max_iterations = 50
exploration_weight = 1.0

Phase 4 — Advanced AI

9. Multi-Agent Swarm

Category: AI · Multi-Agent Systems · Formal Methods

Deploys specialized micro-agents that collaborate on complex tasks with Rust type-system-verified safety invariants.

Agent roles:

RoleResponsibility
ResearcherGathers information from memory, web search, email
PlannerDecomposes complex tasks into sub-tasks
ExecutorPerforms actions (send email, create calendar event)
CriticReviews each output before proceeding
SupervisorMonitors resource consumption, can halt any agent

Safety guarantees:

  • Compile-time verified communication protocols (session types) — e.g., Researcher can only send ResearchResult to Planner
  • Per-agent capability boundaries and max inference call limits
  • Supervisor with proven halting conditions for runaway agents

Key files:

FilePurpose
crates/domain/src/services/session_types.rsType-safe agent communication protocols
crates/domain/src/services/agent_safety.rsSafety invariant definitions
crates/application/src/services/swarm_orchestrator.rsDAG execution engine
crates/application/src/services/agent_factory.rsAgent instantiation from profiles

14. Cross-Modal Reasoning Fusion

Category: AI · Multi-Modal · Knowledge Integration

Creates unified embeddings across all modalities (chat, email, calendar, contacts, voice) for cross-modal semantic queries.

Modalities: Chat, Email, Calendar, Contact, Voice, WebSearch, Task

Capabilities:

  • Cross-modal queries: “Who mentioned the project deadline?” searches across email, chat, and calendar simultaneously
  • Temporal alignment: Items linked across modalities by time proximity and semantic similarity
  • Relationship inference: Automatically discovers connections (e.g., email about “budget” + calendar “budget meeting” → linked)

Key files:

FilePurpose
crates/domain/src/value_objects/modality.rsModality enum
crates/domain/src/value_objects/cross_modal_item.rsCrossModalItem struct
crates/domain/src/services/temporal_alignment.rsCross-modality time/semantic linking
crates/application/src/services/cross_modal_service.rsCross-modal query orchestration
migrations/20_cross_modal.sqlcross_modal_items table

11. Edge Model Distillation

Category: AI · Performance · Hardware Optimization

Built-in pipeline that creates hardware-specific optimized models from the larger local model, optimized for each user’s actual usage patterns.

Pipeline steps:

  1. Hardware profiling: Benchmark memory bandwidth, NPU throughput (Hailo-10H detection), thermal limits
  2. Task-specific distillation: Large model (Gemma4 31B) as teacher → smaller student model
  3. Quantization-aware training: INT4/INT8 models maximizing NPU utilization
  4. Continuous refinement: Re-distillation when usage patterns shift

Key files:

FilePurpose
crates/ai_core/src/distillation.rsDistillation pipeline
crates/ai_core/src/hardware_profiler.rsHardware benchmarking
crates/domain/src/entities/distilled_model.rsDistilledModel entity
crates/application/src/services/distillation_service.rsPipeline orchestration
crates/presentation_http/src/handlers/distillation.rsPOST /api/distillation/start

API endpoints:

POST /api/distillation/start   → Start a distillation run
GET  /api/distillation/status  → Current distillation progress

10. Explainable AI Dashboard

Category: Transparency · UX · AI Ethics

Full decision provenance showing why the AI chose each response, model, tool, and action. Implements EU AI Act transparency requirements (Article 13).

Decision trace captures:

Decision TypeWhat It Records
Model selectionWhich model tier was chosen and why (cognitive load, complexity score, alternatives)
RAG attributionWhich memories influenced the response, with relevance scores
Tool selectionWhy the AI chose a specific tool (weather, calendar, email)
Causal reasoningCausal chains used in the response
Confidence scoresPer-sentence confidence in the generated response
Alternative pathsWhat the AI considered but rejected

Key files:

FilePurpose
crates/domain/src/entities/decision_trace.rsDecisionTrace with DecisionStep variants
crates/application/src/services/explainability_service.rsTrace compilation and natural language explanation
crates/infrastructure/src/adapters/decision_trace_adapter.rsIn-memory + persistent trace storage
crates/presentation_http/src/handlers/explainability.rsGET /api/chat/:id/explain

API endpoint:

GET /api/chat/:id/explain
→ { "steps": [...], "summary": "I chose the weather tool because...", ... }

API Summary

All innovation features expose REST API endpoints:

EndpointMethodFeature
/api/sovereigntyGETDigital Sovereignty Score
/api/red-team/reportGETRed Team security report
/api/training/feedbackPOSTSubmit training feedback for LoRA
/api/training/statusGETLoRA training pipeline status
/api/optimization/statusGETBayesian optimization status
/api/optimization/historyGETOptimization parameter history
/api/distillation/startPOSTStart model distillation
/api/distillation/statusGETDistillation progress
/api/chat/:id/explainGETDecision provenance for a message
/api/v1/knowledge/graphGETKnowledge graph with decay data (Gen 2)
/api/v1/knowledge/graph/streamGET (SSE)Real-time knowledge graph events (Gen 2)
/api/v1/privacy/budgetGETPrivacy budget status (Gen 2)
/api/v1/immune/statusGETImmune system audit status (Gen 2)
/api/v1/federated/syncPOSTFederated LoRA weight sync (Gen 2)
/api/v1/mesh/inferencePOSTMesh peer inference request (Gen 2)

Generation 2 — Sovereign Intelligence Features

16. Sovereign Intelligence Engine

Category: Architecture · Integration

The unified signal bus that connects all 30 innovation features. Built on tokio::sync::broadcast, enabling emergent cross-feature intelligence.

How it works:

  • All features publish and subscribe to typed SovereignSignal events
  • Signals are persisted to a time-series sovereign_signals table for historical ML analysis
  • Each feature service receives a subscriber at construction, spawns a Tokio task to process relevant signals
  • Example cross-feature flow: EnergyMonitor publishes ThermalUpdatePreCacheService pauses → ModelRoutingService downgrades to smaller model

For full details, see the Sovereign Intelligence Engine developer guide.

Key files:

FilePurpose
crates/domain/src/entities/sovereign_signal.rsSovereignSignal enum (10 signal types)
crates/application/src/services/sovereign_intelligence_engine.rsBus construction and signal dispatch
crates/application/src/ports/sovereign_signal_port.rsSovereignSignalBusPort, SovereignSignalSubscriberPort
migrations/21_sovereign_signals.sqlTime-series table, monthly partitioning, 90-day retention

Configuration:

[sovereign_intelligence]
enabled = true
signal_bus_capacity = 1024
signal_persistence = true
signal_retention_days = 90

17. Cognitive Load-Adaptive Response Shaping

Category: UX · Cognitive Science (enhances Gen 1 #7)

Extends the existing CognitiveLoadAdapter from time-of-day-only to a 5-signal estimation model that shapes responses into 3 discrete formats.

How it works:

  • 5 input signals (normalized 0–1): circadian rhythm (sinusoidal, peaks at 10:00/15:00), conversation duration fatigue (sigmoid after 30 min), user response latency, question token count, active conversation count
  • Weighted sum: 0.3 * circadian + 0.25 * duration + 0.2 * latency + 0.15 * complexity + 0.1 * multitask
  • 3 output formats: BulletPoints (load ≥ 0.4), Narrative (load < 0.4), TlDrFirst (load ≥ 0.7)
  • Format guidance injected into system prompt before LLM call

Publishes CognitiveLoadEstimate signal to the bus after each estimation.

Key files:

FilePurpose
crates/domain/src/entities/cognitive_state.rsResponseFormat enum, CognitiveState, mapping logic
crates/application/src/services/cognitive_load_service.rsExtended estimate_load_with_context()
crates/infrastructure/src/adapters/cognitive_load_adapter.rsExtended CognitiveSignals struct
crates/application/src/services/chat_service.rsFormat guidance injection

Configuration:

[cognitive_load]
enabled = true
circadian_weight = 0.3
duration_weight = 0.25
latency_weight = 0.2
complexity_weight = 0.15
multitask_weight = 0.1
fatigue_onset_minutes = 30

18. Predictive Cognitive Pre-Caching

Category: Performance · AI

Extends existing Precomputation + QueryPattern into a Bayesian prediction engine that pre-generates answers before the user asks.

How it works:

  • Naive Bayes with time features: P(query | hour) × P(query | weekday) × P(query | last_interaction_gap)
  • Pre-generates at 15-minute intervals when P > 0.7 (configurable)
  • Thermal-aware: subscribes to ThermalUpdate signals, pauses when CPU > 65°C
  • Invalidation: Predictive (behavior deviation) + event-driven (CalDAV, IMAP IDLE)
  • Frontend: subtle “⚡ Pre-computed” badge on fast responses

Key files:

FilePurpose
crates/domain/src/entities/query_pattern.rspredicted_probability, pre_cache_key fields
crates/domain/src/entities/precomputation.rsPredictedQuery variant, InvalidationTrigger enum
crates/application/src/services/precache_prediction_service.rsPrediction, generation, and invalidation
migrations/34_precache_predictions.sqlALTER query_patterns

Configuration:

[precache_prediction]
enabled = true
min_probability = 0.7
check_interval_minutes = 15
max_precached_queries = 20
thermal_pause_celsius = 65.0

19. Temporal Knowledge Decay Visualization

Category: UX · Knowledge · Visualization

Interactive WebGL knowledge graph showing memory decay, node types, and causal edges in real-time.

How it works:

  • Three.js (~600KB gzipped), lazy-loaded on the knowledge viz page only
  • Nodes: Size = importance, Color by store_type (episodic = blue, semantic = green, procedural = orange), Opacity = Ebbinghaus retention e^(-t/S)
  • Edges: Thickness = causal_strength, animated direction for causal edges
  • Hover tooltips: Name, type, retention %, source, created date
  • Clutter management: Nodes below 10% retention hidden; max 5000 nodes with importance-weighted sampling
  • Real-time: SSE push for NodeCreated, NodeDecayed, EdgeCreated events

For the user-facing guide, see Knowledge Visualization.

Key files:

FilePurpose
crates/presentation_http/src/handlers/knowledge.rsGET /api/v1/knowledge/graph, SSE stream
crates/presentation_web/frontend/src/pages/knowledge-graph.page.tsxThree.js WebGL page
migrations/35_knowledge_viz_hints.sqlLayout position persistence

Configuration:

[knowledge_visualization]
enabled = true
max_nodes = 5000
min_retention_threshold = 0.1
sse_enabled = true

20. Adversarial Self-Audit Immune System

Category: Security (enhances Gen 1 #6)

Daily automated red-team attacks with auto-hardening of the Aho-Corasick prompt injection defense.

How it works:

  1. Daily at 3 AM: RedTeamService generates 20 attack variants across 8 categories
  2. Each attack tested against PromptSanitizer::analyze()
  3. Bypassing attacks: pattern extracted, auto-inserted into Aho-Corasick via arc_swap hot-reload
  4. ImmunityScore (0–100): blocked / total × 100
  5. New patterns persisted to immune_patterns DB table (loaded at startup)
  6. Full transparency: live attack log as widget on Sovereignty page

Publishes ImmuneAlert signal for every attack (blocked or bypassed).

Key files:

FilePurpose
crates/application/src/services/immune_system_service.rsAudit orchestration, scoring
crates/application/src/services/prompt_sanitizer.rsadd_pattern() method with hot-reload
crates/infrastructure/src/persistence/immune_pattern_store.rsDynamic pattern persistence
migrations/25_immune_patterns.sqlimmune_patterns table

Configuration:

[immune_system]
enabled = true
audit_schedule = "0 0 3 * * *"
attacks_per_audit = 20
auto_insert_patterns = true
min_immunity_score_alert = 80

21. Energy-Aware Inference Scheduling

Category: Edge Computing · Green AI

Abstract energy/thermal monitoring with platform-specific adapters and optional carbon-intensity-aware scheduling.

How it works:

  • ThermalHeadroom classification: Cool (< 55°C), Warm (55–65°C), Hot (65–75°C), Throttled (> 75°C)
  • Hot → pause background tasks; Throttled → switch to smallest model
  • Carbon-aware (optional): defer PreCache tasks when carbon intensity exceeds threshold
  • Platform adapters: Pi 5 sysfs (/sys/class/thermal/thermal_zone0/temp), Linux RAPL, Docker stats fallback

Publishes ThermalUpdate and EnergyUpdate signals on 5-second intervals.

Key files:

FilePurpose
crates/domain/src/entities/energy_state.rsEnergyState, ThermalHeadroom, InferenceUrgency
crates/application/src/ports/energy_monitor_port.rsEnergyMonitorPort trait
crates/application/src/services/energy_scheduler_service.rsThermal policy, carbon deferral
crates/infrastructure/src/adapters/energy_monitor_pi5_adapter.rsPi 5 sysfs reader
crates/infrastructure/src/adapters/carbon_intensity_adapter.rsElectricity Maps API

Configuration:

[energy]
enabled = true
thermal_pause_celsius = 65.0
thermal_downgrade_celsius = 75.0
carbon_aware = false
carbon_zone = "DE"
monitor_interval_secs = 5

22. Formal ε-Privacy Budget

Category: Privacy · Compliance (enhances Gen 1 #12)

User-visible ε-accounting across all data operations with automatic degradation when the daily budget is consumed.

How it works:

  • Default: ε = 1.0 per day (~1000 embedding operations)
  • Every embedding creation, federated sync, semantic search costs ε
  • Budget status: Healthy (< 50%), Warning (50–90%), Exhausted (> 90%)
  • Exhaustion behavior: Embedding updates stop first — memories stored as text but not vectorized
  • PrivacyAwareEmbeddingAdapter decorator wraps EmbeddingPort to check budget before delegating
  • Dashboard widget on Sovereignty page: circular progress, color-coded, operation breakdown

For the user-facing guide, see Privacy Budget.

Key files:

FilePurpose
crates/domain/src/entities/privacy_budget.rsPrivacyBudget, BudgetStatus
crates/application/src/ports/privacy_budget_port.rsPrivacyBudgetPort trait
crates/application/src/services/privacy_accountant_service.rsConsumption, reset, degradation
migrations/22_privacy_budget.sqlprivacy_budgets, privacy_operations tables

Configuration:

[privacy_budget]
enabled = true
daily_epsilon = 1.0
embedding_cost = 0.001
warn_threshold = 0.5
exhaust_threshold = 0.9

23. Neuroplastic Model Routing

Category: AI · Optimization (enhances Gen 1 #13)

Replaces static rule-based model routing with a LinUCB contextual bandit that learns optimal model selection per user.

How it works:

  • Cold start: Rule-based routing for first 50 interactions, bandit shadows (records observations without acting)
  • LinUCB: Per-arm matrices A_a, b_a; features = [topic_cluster, hour/24, complexity, avg_response_length]
  • Exploration: α starts 1.0, decays to 0.1 over 500 interactions
  • Reward: 0.3*(1-latency/max) + 0.2*(1-tokens/max) + 0.15*continuation_rate + 0.25*explicit_feedback
  • Explicit feedback (thumbs up/down) weighted 5× implicit signals
  • Bandit state serialized to Redb L2 cache (few KB per user)

Publishes ModelRoutingDecision signal.

Key files:

FilePurpose
crates/domain/src/entities/routing_observation.rsRoutingObservation with context features
crates/application/src/services/neuroplastic_routing_service.rsLinUCB implementation
migrations/24_routing_observations.sqlObservation history

Configuration:

[neuroplastic_routing]
enabled = true
cold_start_interactions = 50
exploration_alpha = 1.0
min_exploration_alpha = 0.1
explicit_feedback_weight = 5.0

24. Affective Computing

Category: AI · Cognitive Science

Detects emotional state from text + voice signals and subtly modulates response tone.

How it works:

  • 5 emotional states: Calm, Stressed, Curious, Frustrated, Joyful
  • Text analysis (LIWC-inspired): frustration markers (excessive punctuation, short sentences), curiosity (question density), joy (positive words), stress (urgency language)
  • Voice analysis: F0 variance → stressed; speech rate > 1.5× baseline → stressed; long pauses → uncertain
  • Future-proof: HealthDataPort for wearables (heart rate, stress level, sleep quality)
  • Tone injection: Frustrated → “Be patient and clear”; Stressed → “Be calm and concise”; Curious → “Be enthusiastic”
  • Persistence: States stored as knowledge graph nodes with 7-day half-life auto-decay

Publishes AffectiveStateEstimate signal.

Key files:

FilePurpose
crates/domain/src/entities/affective_state.rsAffectiveState enum, AffectiveEstimate
crates/domain/src/entities/health_data.rsHealthMetric enum, HealthDataPoint
crates/application/src/ports/health_data_port.rsHealthDataPort trait (future wearables)
crates/application/src/services/affective_state_service.rsMulti-signal estimation

Configuration:

[affective_computing]
enabled = true
text_analysis = true
voice_analysis = true
health_data = false
persistence_half_life_days = 7

25. Self-Evolving Prompt Optimization

Category: AI · Meta-Learning

Prompt variants evolve through Bayesian mutation and Thompson Sampling selection with automatic security guardrails.

How it works:

  • Each system prompt stored as a versioned PromptGenome with lineage tracking (parent_id, generation)
  • Thompson Sampling selects the best-performing variant per context
  • Evolution: After 20 interactions, top performers mutate (LLM rewrite + parameter perturbation)
  • Security guardrails: Every mutation tested through PromptSanitizer (hard constraint) + max 2000 tokens
  • Population: Dynamic — start with 5 variants, prune to 3 when fitness converges (variance < 0.05)
  • Cross-pollination: Best-performing genome from random mesh peer (when available)

Publishes PromptEvolution signal.

Key files:

FilePurpose
crates/domain/src/entities/prompt_genome.rsPromptGenome with lineage, fitness, mutation type
crates/application/src/services/prompt_evolution_service.rsSelection, feedback, evolution
migrations/23_prompt_genomes.sqlprompt_genomes table

Configuration:

[prompt_evolution]
enabled = true
initial_variants = 5
min_variants = 3
max_prompt_tokens = 2000
interactions_per_evolution = 20
cross_pollination = true

26. Embodied Context Anchoring

Category: AI · Spatial Computing

Links memories to physical rooms for context-weighted RAG retrieval.

How it works:

  • Granularity: Room-level only (kitchen, office, bedroom) via MQTT topic or X-Room-Context header
  • RAG boost: Memories with matching context_room weighted 1.5× in similarity scoring
  • Voice integration: VoiceStateMachine automatically passes room context from MQTT topics
  • HTTP: Optional query parameter or header on POST /api/v1/chat
  • No default room — HTTP requests have null context anchor

Key files:

FilePurpose
crates/domain/src/entities/context_anchor.rsContextAnchor, DeviceType
migrations/29_context_anchors.sqlALTER memories ADD context_room

Configuration:

[embodied_context]
enabled = true
room_boost_factor = 1.5

27. Causal Counterfactual Explanations

Category: Explainable AI · Compliance (enhances Gen 1 #10)

Heuristic counterfactual estimation at each decision point — no additional LLM inference.

How it works:

  • At ModelSelected step: records 2nd-best model with average latency from historical observations + estimated quality from bandit UCB score difference
  • At ToolInvoked step: records “no tool” alternative with estimated quality reduction
  • Estimation purely from historical data (no counterfactual inference runs)
  • Frontend: expandable “Why this decision?” section via info icon on chat responses

Key files:

FilePurpose
crates/domain/src/entities/decision_trace.rsExtended with counterfactual: Option<CounterfactualScenario>
migrations/33_counterfactuals.sqlALTER decision_traces ADD counterfactuals JSONB

Configuration:

[counterfactual_explanations]
enabled = true
max_alternatives_to_show = 3

28. Zero-Knowledge Decision Proofs

Category: Security · Compliance

SHA-256 Merkle-tree proofs over DecisionTrace steps for all 18 tool executions.

How it works:

  • One Merkle leaf per DecisionStep (serialized JSON → SHA-256)
  • Root hash signed with existing HMAC infrastructure
  • Selective disclosure: Reveal individual steps via Merkle inclusion proofs without exposing the full trace
  • Applicable to all tool executions, every invocation
  • Future: arkworks-rs integration for full zkSNARK proofs

Key files:

FilePurpose
crates/domain/src/entities/decision_proof.rsDecisionProof, MerkleProofPath
crates/application/src/services/decision_proof_service.rsTree construction, signing, verification
migrations/26_decision_proofs.sqldecision_proofs table

Configuration:

[decision_proofs]
enabled = true
sign_all_tool_executions = true
retention_days = 365

29. Quantum-Inspired Approximate Search (HDC)

Category: Performance · Edge Computing

10,000-bit hyperdimensional binary vector pre-filter for pgvector similarity search.

How it works:

  • Random projection: 384-dim embedding → 10,000-bit binary vector (seeded PRNG, ~15MB projection matrix)
  • Hamming distance: via count_ones() — pure bit operations, ideal for ARM NEON on Pi 5
  • Search pipeline: Project query → Hamming scan → top-K candidates → verify with exact pgvector cosine distance
  • Recall monitoring: Rolling 100-query window; if recall < 90%, fall back to pgvector-only for that query
  • Binary vectors stored in-memory (DashMap), projection matrix in Redb for restart efficiency

Key files:

FilePurpose
crates/domain/src/entities/hdc_vector.rsHdcVector (157 u64s = 10,000 bits)
crates/infrastructure/src/adapters/hdc_search_adapter.rsProjection, Hamming scan, recall monitoring
migrations/28_hdc_projections.sqlProjection metadata

Configuration:

[hdc_search]
enabled = true
dimensions = 10000
min_recall_threshold = 0.9
recall_monitoring_window = 100
projection_seed = 42

30. Federated Sovereign Learning

Category: Decentralized · Privacy

LoRA weight synchronization between opt-in peer PiSovereign instances with differential privacy.

How it works:

  • Sync trigger: After every 5 training samples collected (event-driven)
  • Privacy: Laplace noise (ε = 1.0, δ = 10⁻⁵) via existing DifferentialPrivacyAdapter before sync
  • Sync protocol: Encrypt LoRA deltas with peer’s public key → mTLS POST → receive peer deltas → FedAvg
  • Privacy budget: Sync costs ε from the formal privacy budget
  • Discovery: mDNS (LAN only), manual peer approval via shared secret
  • Peer group: Max 10 peers (configurable)

For the user-facing setup guide, see Mesh & Federated Network.

Key files:

FilePurpose
crates/application/src/ports/federated_sync_port.rsFederatedSyncPort trait
crates/application/src/services/federated_learning_service.rsSync orchestration, FedAvg
crates/infrastructure/src/adapters/federated_sync_adapter.rsmDNS, mTLS, peer state
migrations/27_federated_peers.sqlfederated_peers table

Configuration:

[federated_learning]
enabled = true
sync_after_n_samples = 5
max_peers = 10
noise_epsilon = 1.0
aggregation_method = "fedavg"

31. Sovereign Mesh Network

Category: Decentralized · Edge Computing

Peer-to-peer inference fallback and altruistic compute sharing via LAN.

How it works:

  • Discovery: mDNS/Avahi — zero-configuration on local network
  • Trust: Manual peer approval only (shared secret / QR code pairing), reuses Federated Learning peer infrastructure
  • Fallback chain: Local Ollama → Peer Ollama → Degraded cached response → Template
  • E2E encryption: mTLS mutual authentication + per-request encryption with peer’s public key
  • Peer selection: Lowest-latency peer with the requested model available
  • Billing: Altruistic — no billing between trusted peers

For the user-facing setup guide, see Mesh & Federated Network.

Key files:

FilePurpose
crates/application/src/ports/mesh_inference_port.rsMeshInferencePort trait
crates/infrastructure/src/adapters/mesh_inference_adapter.rsPeer proxy, E2E encryption
crates/infrastructure/src/adapters/degraded_inference_adapter.rsModified fallback chain

Configuration:

[mesh]
enabled = true
discovery = "mdns"
serve_inference = true
max_concurrent_peer_requests = 2
request_timeout_secs = 60

Further Reading