ARIA Platform v3.0
ARIA

ARIA — Adaptive Responsive
Intelligence Assistant

A personal AI assistant that has analyzed 567,090 iMessages, 139,000 photos, and 1,002 Google Voice records — built across 5 repos, 18+ integrations, and 9 device data streams.

567K
iMessages Analyzed
139K
Photos
131K
Knowledge Facts
133
AI Tools
18+
Integrations

Platform at a Glance

Five repositories, each with a focused responsibility. A shared PostgreSQL database ties them together. The iOS app connects via API. External AI clients connect via MCP.

aria

The main web application and API layer. Chat UI with SSE streaming, 21 tool modules dispatching 133 tools, system prompt construction, memory extraction, and all integration endpoints.

Next.js 14 · React 18 · TypeScript · Tailwind
35 API route groups · 21 tool modules · SSE chat streaming

aria-tempo

The background job worker. Polls the shared database for queued jobs and processes them with a 3-provider LLM router. Runs the proactive intelligence pipeline, journal generation, report compilation, and 38 distinct job handlers.

Node.js · TypeScript · PostgreSQL polling
38 job handlers · 3-provider LLM router · 5s poll interval
📦

aria-tempo-client

Shared TypeScript library defining job type contracts. Type-safe payload interfaces, Zod validation schemas, and the TempoClient class for enqueueing jobs. Used by both aria and aria-tempo.

TypeScript · Zod · npm package
38 job types · Type-safe payloads · Shared contracts
🔌

aria-mcp-server

Model Context Protocol server exposing ARIA's capabilities to external AI clients like Claude.ai, Claude Desktop, and Claude Code. Memory CRUD, conversation history, owner context, and Tempo job scheduling.

Node.js · TypeScript · MCP SDK · Express
18 MCP tools · OAuth 2.1 with PKCE · stdio + HTTP
📱

aria-ios

Native iOS client built with Swift 6 and SwiftUI. Syncs 9 device data streams to the server, handles push notifications, and provides a polished chat interface with SSE streaming. Background fetch, geofencing, Spotlight indexing, app shortcuts, and a watchOS companion app.

Swift 6 · SwiftUI · iOS 17+ · XcodeGen
9 data streams · 12 widgets · APNs push · watchOS companion

Core Systems & Engines

Five named systems give ARIA her cognitive architecture. Each handles a distinct dimension of intelligence — from personality to prediction, from classification to knowledge.

🎭

PRISM

Persona Role & Interaction Style Manager

Two-layer prompt architecture separating immutable capabilities (kernel) from swappable personality overlays (personas). Seven distinct cognitive modes — from Chief of Staff to Provocateur — with seamless mid-conversation switching and team-style handoff patterns.

7 personas kernel + overlay
🌱

PIE

Proactive Intelligence Engine

Autonomous three-gate pipeline monitoring 15+ data sources against watermarks. Detects changes, classifies significance, performs deep analysis, and delivers actionable insights via push notifications or email — all without being asked.

3-gate pipeline Gate 1–2 free

CAFT

Context-Adaptive Fast-Track

Significance classification layer powered by Gemini Flash-Lite. Sits between raw data changes and expensive analysis. Filters noise from signal with near-zero cost, ensuring only meaningful changes trigger Claude Sonnet analysis.

free tier
🔬

PKE

Personal Knowledge Engine

Knowledge graph storing structured facts about people, places, events, and preferences. Entity resolution, confidence scoring, temporal validity, and evidence tracking. Continuously enriched from conversations, device data, and proactive analysis.

5,518 entities 26,940 facts
📐

KG

Knowledge Graph & Owner Profile

Structured owner profile with 131,016 facts across 9+ domains: events, people, preferences, lifestyle, work, health, places, media, and more. Every fact carries confidence scores, source attribution, and temporal validity. The foundation ARIA uses to truly know her owner.

131K+ facts 9+ domains confidence scoring

How ARIA Thinks

Every message traverses a carefully orchestrated pipeline — from authentication through context assembly to streaming response. Here's the path a single chat request takes.

iOS / Web Client
POST /api/chat
Auth
Build Context
12 parallel queries
Load PRISM Persona
Construct System Prompt
~12K tokens
Claude Sonnet 4.6
streaming
Tool Dispatch
up to 10 rounds
Memory Extraction
<memory_update> blocks
SSE Response
start → status → chunk → done

Context includes: core memory · calendar · gmail · ARIA email · contacts · health · location · weather · social · skills · PIE insights · KG facts
12 parallel context queries. When a message arrives, ARIA simultaneously fetches core memory, upcoming calendar events, recent emails, contact data, health metrics, location, weather, social interactions, active skills, proactive insights, knowledge graph facts, and conversation history. All run concurrently to minimize latency.

Intelligence Pipeline

ARIA doesn't wait to be asked. The Proactive Intelligence Engine runs continuously, monitoring data sources and surfacing insights before the owner needs them.

Gate 1

Context Accumulate

Monitors 15+ data sources against watermarks. Detects meaningful changes. Runs every 5 minutes.

FREE
Gate 2

Significance Check

Gemini Flash-Lite classifies changes as significant or noise. Filters ~80% of events. Auto-chained from Gate 1.

FREE TIER
Gate 3

Anticipation Analyze

Claude Sonnet performs deep analysis on significant changes. Generates actionable insights with reasoning and evidence.

Deliver

Insight Deliver

Routes insights to configured notification methods: push notifications, email, or in-app display.

PUSH / EMAIL
Cost-optimized by design. Gates 1 and 2 are free or near-free, filtering out noise before expensive Claude analysis. Pattern maintenance runs daily to decay, promote, or deactivate behavioral patterns, keeping the system sharp without manual tuning.
ARIA interacting with a holographic interface

The Integration Ecosystem

ARIA connects to 18+ services and APIs. Each integration follows a consistent two-file pattern: tool definitions with input schemas, and an API wrapper with service logic. All dispatched by prefix in the chat route.

📅
Google Calendar
calendar_*
📧
Gmail
gmail_*
ARIA Email
aria_email_*
👥
Contacts
contacts_*
📞
Twilio SMS/Voice
twilio_*
🎤
ElevenLabs Voice
voice calls
Looki Wearable
looki_*
HealthKit
health_*
📷
Photos
photo_*
🏠
HomeKit
home_*
🎵
Music
music_*
📋
Reminders
reminders_*
💬
iMessage
imessage_*
Weather
weather_*
🐙
GitHub
github_*
📝
Notion
notion_*
🌐
Social Network
social_*
👁
Gemini Vision
gemini_*
Consistent two-file pattern. Every integration follows the same structure: *-tools.ts defines tool names, descriptions, and input schemas. *.ts wraps the actual API. Tools are conditionally registered based on connection status. All non-read-only calls are logged to event_log.

Device Data Streams

The iOS app syncs 9 distinct data streams to the server in near real-time. Each stream uses sync hash optimization for efficient change detection — only deltas are transmitted.

📅

Calendar Events

EventKit → /api/device-calendar/

👥

Contacts

ContactsFramework → /api/device-contacts/

📋

Reminders

EventKit → /api/device-reminders/

📍

Location

CoreLocation → /api/device-location/

🏃

Activity / Motion

CoreMotion → /api/device-activity/

Health

HealthKit → /api/health-data/

🎵

Music

MusicKit → /api/device-music/

🏠

HomeKit

HomeKit → /api/device-homekit/

📷

Photos

PhotosFramework → /api/photos/metadata

3-Tier LLM Router

Every LLM call in ARIA is routed through a tiered provider chain. If the primary model fails or is unavailable, the request automatically falls through to the next provider. Each tier is optimized for its use case.

Tier Use Case Primary Fallback 1 Fallback 2
Reasoning Chat, deep analysis, tool use, memory extraction Claude Sonnet 4.6 GPT-4o Claude Haiku 4.5
Generation Journal entries, briefings, reports, summaries Claude Haiku 4.5 Gemini 2.0 Flash GPT-4o-mini
Classification Significance checks, quality scoring, fast-track gating Gemini 2.5 Flash-Lite Claude Haiku 4.5 GPT-4o-mini

Automatic Fallback

If a provider returns an error or times out, the router transparently retries with the next provider in the chain. No user-visible interruption.

Cost Tracking

Every LLM call is logged to the llm_usage table with model, tokens, cost, and latency. Monthly cost reports are generated automatically.

Configurable

Model assignments are stored in llm_model_config and can be changed without code deploys. 9 chain entries cover all use cases.

ARIA giving a presentation

Deployment & Infrastructure

Production infrastructure on AWS. Push-to-deploy for server-side services. Manual upload to TestFlight for iOS. All secrets managed centrally via AWS Secrets Manager.

ECS Fargate

aria + aria-tempo
Auto-deploy on push

🗃

Aurora PostgreSQL

Shared database
48+ tables, 39 migrations

🚀

App Runner

aria-mcp-server
OAuth 2.1 + PKCE

📱

TestFlight

aria-ios
Manual upload via script

Push to branch
GitHub Actions
OIDC auth
Build Docker Image
3-stage build
Push to Registry
Rolling Update
zero-downtime

Secrets fetched from secrets manager at build time (aria) and runtime (aria-tempo)

Database Migrations

Schema lives in aria/sql/ with 39+ migration files. Applied manually via psql before deploys. Migrations must land before new code goes live, or API routes fail on missing columns.

Schema Registry

The schema_registry table declares required tables and columns for each job type. Pre-flight validation prevents runtime failures from missing migrations. Every new job handler must register its schema requirements.