Changelogο
All notable changes to Ada will be documented in this file.
The format is based on Keep a Changelog, and this project adheres to Semantic Versioning.
[Unreleased]ο
[2.9.0] - 2025-12-19ο
β‘ Performance - Phase 2C: Parallel Optimizationsο
2.5x speedup in prompt building - Parallel RAG retrieval and specialist execution
3.96x speedup in RAG context retrieval (200ms β 50ms)
2.98x speedup in specialist execution (150ms β 50ms)
Real-world benchmark: 200ms β 80ms (saves 120ms per request)
ThreadPoolExecutor for concurrent RAG operations (persona, memories, FAQs, turns)
Smart specialist prioritization - HIGH/CRITICAL run parallel, MEDIUM/LOW sequential
π§ͺ Testingο
17 new tests for parallel operations (100% passing)
Performance benchmarks with ThreadPoolExecutor mocking
Real-world latency simulations (20ms-80ms range)
π Documentationο
Updated architecture diagrams with parallel flow
Performance comparison charts (before/after)
Complete release notes in
RELEASE_v2.9.0.md
[2.8.0] - 2025-12-19ο
π Performance - Response Caching Layer (Phase 2B)ο
Cache frequently asked questions - Instant responses for repeated queries
LRU eviction with configurable max size (default: 100 entries)
TTL-based expiration - Configurable cache lifetime
Cache statistics - Hit rate tracking and logging
Expected ~40% hit rate on repetitive workflows
π§ͺ Testingο
8 new cache layer tests (eviction, TTL, hit rate)
Integration with contextual router tests
Cache statistics validation
π Documentationο
Cache architecture documentation
Performance tuning guide
Complete release notes in
RELEASE_v2.8.0.md
[2.7.0] - 2025-12-19ο
π― Features - Contextual Router & Ada Log Intelligenceο
Contextual router - 22 patterns across 5 categories for intelligent routing
Trivial questions (greetings, thanks)
Fact recall (recent memories)
Analytical queries (requires reasoning)
Creative requests (needs inspiration)
Code-related (development tasks)
Dynamic specialist activation - Router determines which specialists are needed
Ada Log Intelligence - Biomimetic log analysis specialist
Minecraft crash report parser (kid-friendly explanations!)
Pattern matching for common errors (OptiFine conflicts, OutOfMemory)
Foundation for 100:1 log compression using signal weights
Standalone
ada-logsPython package
β‘ Performanceο
Contextual router reduces unnecessary specialist overhead
22 tests, 0.08s runtime, 100% passing
Smart RAG context selection based on query type
π§ͺ Testingο
22 router tests (pattern matching, specialist selection)
Minecraft crash parser tests
Integration tests with log specialist
π¦ New Packageο
ada-logs - Standalone log analysis library
Pure Python (3.11+), CC0 license
CLI tool:
ada-logs analyze crash.log --for-kidsExtensible parser system (JSON, syslog, custom formats)
π Documentationο
Router pattern documentation
Minecraft crash analysis examples
Complete release notes in
RELEASE_v2.7.0.md
[2.6.0] - 2025-12-19ο
π Major Feature - Code Completion MVPο
Native code completion in Neovim - Copilot-style autocomplete with Ada!
Press
<C-x><C-a>in insert mode for completions2.6s mean latency, 77% quality score
100% success rate across 24 diverse code scenarios
Works with Python, JavaScript, Lua, Rust, and more
10.6x speedup - Optimized for code models with FIM format (27.7s β 2.6s)
Model optimization - qwen2.5-coder:7b specialized code model (4.7GB)
MCP integration - New
complete_codetool bypasses RAG overhead
β¨ Featuresο
ada.nvim completion module -
lua/ada/completion.luaContext-aware completion - Sees code before AND after cursor
Language-aware - Auto-detects from filetype
Privacy-first - Runs entirely on your machine
π Benchmarksο
Complete latency analysis (mean, median, best, worst)
Quality scoring across 24 real-world test cases
Model comparison (DeepSeek-R1 vs Qwen2.5-Coder)
FIM format validation
π§ͺ Testingο
benchmarks/benchmark_completion.py- Reproducible test suite24 diverse code scenarios (functions, loops, error handling, etc.)
Quality scoring system with detailed metrics
π Documentationο
ada.nvim/COMPLETION_QUICKSTART.md- 5-minute setup guidebenchmarks/BENCHMARK_RESULTS_QWEN_FIM.md- Complete analysis13 real-world completion examples
Updated
DOCUMENTATION_INDEX.md
[2.5.0] - 2025-12-18ο
π§ Internal Improvementsο
Refactoring and optimization prep for Phase 2
Code quality improvements
Testing infrastructure enhancements
[2.4.0] - 2025-12-18ο
π§ Internal Improvementsο
Architecture refinements
Performance monitoring baseline
Preparation for contextual routing
[2.3.0] - 2025-12-18ο
π¬ Research & Validationο
Contextual Documentation Framework (22 research phases across 4 dimensions)
Dimension 1 (HumanβHuman): Information theory, causal discovery, noise ceiling, adversarial robustness, cross-domain transfer, sensitivity analysis, Bayesian uncertainty, meta-validation
Dimension 2 (HumanβLLM): Empathy effectiveness (effect size 3.089!), adversarial validation, contextual awareness (r=0.924)
Dimension 3 (LLMβLLM): Information density, semantic compression patterns
Dimension 4 (Cross-Model): Discovered but not yet implemented
Key findings: Context-matching beats universal approaches, empathy is quantifiable, same patterns apply to machine communication
23 tests, 2.95s runtime, 100% passing
98% replication stability, 83.3% real-world validation
π Key Numbers (Publication-Worthy!)ο
Effect size 3.089: Empathy scaffolding (0%β100% completion under stress)
r=0.924: Context-matching correlation (THE meta-principle)
83.3%: Real-world validation accuracy
98%: Replication stability
60%: Hybrid strategy win rate (humans AND LLMs!)
+53%: Query success improvement from structure
+27.8%: Comprehension recovery under cognitive load
β¨ Featuresο
New dependencies: scikit-learn>=1.3.0, scipy>=1.11.0 for scientific computing
Democratic science methodology: Rigorous research anyone can run locally in ~8 seconds
Unified communication theory: Same malleability principles across human and machine communication
π Documentationο
docs/contextual_documentation_framework.md - Complete 250+ page research document
.ai/PHASE9-22-HANDOFF.md - Session handoff summary
.ai/DEMOCRATIC-SCIENCE-MENU.md - Democratic science guide
.ai/HEAVY-MATH-ADDENDUM.md - Mathematical foundations
RELEASE_v2.3.0.md - Complete release notes
π§ Infrastructureο
Added
.hypothesis/to .gitignore for property-based test cache
π― Whatβs Nextο
Full codebase specialist for Ada (semantic code search, architecture understanding)
Cross-model communication optimization (Dimension 4)
Adaptive weighting for context-specific importance signals
[2.2.0] - 2025-12-18ο
π¬ Research & Optimizationο
Memory importance signal weight optimization (Phases 1-7)
Systematic research: property testing β synthetic data β ablation β grid search β production validation β deployment β visualization
Key discovery: Surprise-only (r=0.876) beats multi-signal baseline (r=0.869)
Optimal weights found: decay=0.10 (was 0.40), surprise=0.60 (was 0.30)
Improvement: 12-38% across synthetic datasets, +6.5% on real conversations
80 tests, 3.56s total runtime, 100% passing
Research validated: Temporal decay was overweighted 4x, surprise underweighted 2x
Same-day deployment: Research β production in <24hrs via TDD methodology
β¨ Featuresο
Optimal importance weights deployed to production
Updated
brain/config.pywith research-validated optimal configurationBackward compatible: legacy weights available via environment variables
Rollback mechanism tested and ready
Comprehensive research documentation (Phase 8: Meta-Science)
9 narrative formats, 45,000 words total documenting same research:
Machine-readable summary (
.ai/RESEARCH-FINDINGS-V2.2.md)Academic article (peer-review ready, 8,000 words)
CCRU-inspired experimental narrative (hyperstition engaged, 9,000 words)
Blog post (accessible science communication, 4,500 words)
Technical deep-dive (implementation guide, 6,000 words)
Twitter thread (15 tweets, viral-ready)
Recursion reveal README (meta-awareness, 3,500 words)
Techno-horror essay (accelerationist, 5,000 words)
Brief general audience explainer (3-minute read, 1,200 words)
New docs section:
docs/research_narratives.rstshowcasing all formatsComplete with navigation guide, verification hooks, and meta-narrative
π Visualizationsο
6 publication-quality research graphs generated (Phase 7)
Ablation study comparison (signal configurations)
Grid search heatmap (decay vs surprise landscape)
Improvement distribution (before/after comparison)
Correlation vs weights (3D surface plot)
Detail level changes (gradient efficiency)
Production validation (real conversation results)
All graphs 300 DPI, publication-ready (2.2 MB total)
π§ͺ Testingο
New research test suite
tests/test_property_based.py- 27 tests, mathematical invariantstests/test_synthetic_data.py- 10 tests, ground truth datasetstests/test_ablation_studies.py- 12 tests, signal isolationtests/test_weight_optimization.py- 7 tests, grid search (169 configurations)tests/test_production_validation.py- 6 tests, real conversation datatests/test_deployment.py- 11 tests, config validation & rollbacktests/test_visualizations.py- 7 tests, graph generationTotal: 80 new tests, all passing, <4s runtime
π Documentationο
Updated
.ai/context.mdwith research findings and optimal weightsUpdated
docs/biomimetic_features.rstwith validation resultsAdded
docs/research_narratives.rstlanding page for all narrative formatsMachine docs in
.ai/RESEARCH-FINDINGS-V2.2.mdfor AI assistant verificationResearch methodology documented for future phases (9-12 planned)
π§ Configurationο
Optimal weights now default in
brain/config.pyLegacy weights available via:
IMPORTANCE_WEIGHT_DECAY=0.40 IMPORTANCE_WEIGHT_SURPRISE=0.30All signal weights configurable via environment variables
Maintains backward compatibility with existing deployments
π― Performance Impactο
Context selection improved by +6.5% per turn on real conversations
80% of turns show positive importance prediction changes
250% increase in medium-detail memory chunks (better gradient utilization)
Token budget increase: +17.9% (acceptable trade-off for quality gain)
[2.1.0] - 2025-12-17ο
β‘ Performanceο
Multi-timescale context caching system - Dramatically reduces redundant RAG queries
Personas cached for 24 hours (identity rarely changes)
FAQs cached for 24 hours (knowledge base updates infrequently)
Memories cached for 5 minutes (balance freshness vs performance)
Conversation turns cached for 1 hour (recent context preserved)
LRU eviction prevents unbounded memory growth
Cache stats logged per request for observability
Expected ~70% reduction in ChromaDB queries for repeated context
β¨ Featuresο
New modular prompt building architecture
PromptAssembler- Clean orchestration with automatic cachingContextRetriever- Cache-aware RAG data retrievalSectionBuilder- Structured section formattingMultiTimescaleCache- Production-ready caching implementation
Cache integration transparent to specialists and adapters
Per-request cache statistics in logs
ποΈ Removed (Breaking Changes)ο
Deleted legacy
brain/_legacy_prompt_builder.py(266 lines)Removed backward-compatible
build_prompt()shimNew code must use
PromptAssemblerAPI directlyTechnical debt eliminated: -299 net lines across the codebase
π§ Fixesο
Corrected default port from 7000 to 8000 in ada-client and ada-cli
Brain runs internally on 7000, exposed via Docker on 8000
Clients now default to correct external port
Updated CLI Python requirement to >=3.13 for consistency
π Documentationο
Updated architecture.rst with caching system overview
Refreshed API usage examples for new PromptAssembler
Updated specialist RAG documentation
Token monitoring examples modernized
AI documentation (codebase-map.json) fully updated
Marked TODO-CACHE-INTEGRATION.md as complete
π§ͺ Testingο
23 new comprehensive cache tests (15 basic + 8 integration)
Tests cover TTL expiration, LRU eviction, cache stats, and retriever integration
All tests passing, cache validated in production
π¦ Dependenciesο
No new external dependencies (pure Python implementation)
Impactο
Massive performance win: Caching reduces latency on repeated queries from seconds to milliseconds. Cleaner codebase with 300 fewer lines of legacy code. Modern modular architecture ready for future optimization phases (FAQ caching, memory caching expansion).
Migration Guide:
# Old (removed):
from brain.prompt_builder import build_prompt
prompt, context = await build_prompt(...)
# New (required):
from brain.prompt_builder import PromptAssembler
assembler = PromptAssembler()
prompt = assembler.build_prompt(
user_message="...",
conversation_id="...",
notices=get_active_notices()
)
[1.8.0] - 2025-12-16ο
β¨ Featuresο
Complete Nix flake with development shell, package build, and NixOS module
Python version fallback (python313 β python312 β python3) for compatibility
Automatic locale fixes (LC_ALL=C.UTF-8) in Nix environment
direnv integration for automatic environment activation
Validation script (scripts/test_nix_setup.sh) with 10-point health check
ada doctor command now detects Python version mismatches and suggests Nix
π Documentationο
New primary entry point: docs/zero_to_ada.rst with decision tree
Comprehensive Nix guide: docs/nix.rst (537 lines)
Real-world troubleshooting: docs/nix_troubleshooting.md
Updated README.md to prioritize Nix for users without Python 3.13
π§ Maintenanceο
Version management automation (scripts/version.sh)
Updated .gitignore for Nix artifacts
Impactο
Solves Python 3.13 availability on Ubuntu/Debian. Makes Ada accessible to users on older distros without requiring Docker or building Python from source.
[1.7.0] - 2025-12-16ο
β¨ Featuresο
Kubernetes deployment manifests (k8s/deployment.yaml, service.yaml, configmap.yaml)
Helm chart for one-command deployment (helm/)
Nomad job specification for HashiCorp Nomad (nomad/ada.nomad)
Multi-cloud deployment support (AWS EKS, GCP GKE, Azure AKS, local k3s/k0s)
π Documentationο
Comprehensive orchestration guides (docs/k8s.rst, docs/helm.rst, docs/nomad.rst)
Hardware requirements and scaling guidance
Multi-cloud deployment strategies
Production deployment checklist
π§ Maintenanceο
Updated README.md with orchestration quick links
Fixed markdown formatting in documentation
Impactο
Enterprise-ready deployment options. Ada can now run in production Kubernetes clusters, scale horizontally, and integrate with cloud infrastructure.
[1.6.0] - 2025-12-15ο
β¨ Featuresο
Matrix bridge for chat integration (matrix-bridge/)
Multiple interface adapters (CLI, Web, Matrix, MCP)
Bidirectional specialists (LLM can invoke tools mid-response)
Wikipedia and Fandom wiki lookup specialist
Documentation specialist (Ada can read her own docs)
Web search specialist with DuckDuckGo integration
π Bug Fixesο
Fixed OCR specialist activation
Improved memory consolidation reliability
Fixed streaming response handling
π Documentationο
Architecture documentation overhaul
Adapter pattern documentation
Specialist system documentation
Matrix integration guide
π§ Maintenanceο
Improved error handling in specialists
Better logging throughout system
Code cleanup and refactoring
[1.5.0] - 2025-12-10ο
β¨ Featuresο
MCP (Model Context Protocol) server implementation
Ada CLI tool (ada-cli) for terminal interaction
Streaming chat responses via Server-Sent Events
RAG-based memory system with ChromaDB
π Documentationο
Getting started guide
API documentation
Specialist development guide
Earlier Versionsο
See git history for changes before v1.5.0.
Version Formatο
Major.Minor.Patch (Semantic Versioning)
Major: Breaking changes, architectural shifts
Minor: New features, backwards-compatible changes
Patch: Bug fixes, documentation updates
Commit Types (Conventional Commits):
feat: New feature β Minor version bumpfix: Bug fix β Patch version bumpdocs: Documentation β Patch version bumpperf: Performance improvement β Minor version bumprefactor: Code refactoring β Patch version bumptest: Test changes β Patch version bumpchore: Maintenance β Patch version bumpBREAKING CHANGE: in commit body β Major version bump
Example:
feat(specialists): add Wikipedia search
fix(matrix): handle rate limiting errors
docs: update changelog automation guide