Skip to main content
Normalized for Mintlify from knowledge-base/neurigraph-memory-architecture/legacy-brain-api-pricing-strategy.mdx.

aiConnected Brain API

Strategic Product Document


Executive Summary

Brain is a persistent memory layer that sits above all AI platforms. Users maintain continuous context across ChatGPT, Claude, Gemini, and any MCP-compatible service. Brain is the foundation of aiConnected’s cognitive operating system for AI and future robotics. The name earns its meaning: aiConnected = All AIs connected through persistent memory.

The Problem

Every person using AI today has fragmented conversations:
  • ChatGPT knows about your project but Claude doesn’t
  • You explained your preferences to Gemini last week, now you’re starting over with GPT
  • Switch models and lose everything
  • Even within the same platform, context disappears after the conversation ends
The major platforms (OpenAI, Anthropic, Google) will never build cross-platform memory. Walled gardens serve their business interests. That interoperability gap is aiConnected’s entire market.

The Solution

Brain is a persistent memory layer that:
  • Stores conversation context across all AI platforms
  • Enables continuous memory regardless of which AI you use
  • Learns and accumulates knowledge over time
  • Creates switching costs that compound with usage

Target Markets

Consumer Market

  • Power users juggling multiple AI platforms
  • Developers and creators needing continuity
  • Professionals wanting AI that learns their preferences
  • Teams needing shared context

B2B Market (via Agency Platform)

  • Marketing agencies deploying AI for clients
  • Service businesses using AI-powered chat and voice
  • Enterprises wanting persistent AI relationships with customers

Pricing Structure

Consumer Tiers

TierBase PriceIncludesOverage
Starter$1/mo10 free actions$0.50/action
Personal$9/mo100 actions$0.25/action
Pro$19/moUnlimited actions-
Teams$39/seatUnlimited + shared memory-

Storage Add-On

Add-OnPriceIncludes
Storage expansion$5/mo+100,000 memories

Included Storage by Tier

TierIncluded Memories
Starter1,000
Personal10,000
Pro100,000

Conversion Economics

The pricing structure drives natural upgrades:
Monthly ActionsStarter CostPersonal CostPro Cost
10$1 (free actions)$9$19
15$3.50$9$19
20$6$9$19
25$8.50$9 ✓$19
50$21$14$19 ✓
100$46$34$19 ✓
  • At 25 actions/month, Personal becomes the better value
  • At 50 actions/month, Pro becomes the better value
  • The $1 entry point filters out non-serious users while remaining accessible
  • The 10 free actions (a $5 value) let users experience the product before paying overage

Revenue Projections

Distribution Assumptions

SegmentPercentageAverage Monthly Revenue
Starter50%4(4 (1 + ~6 paid actions)
Personal30%11(11 (9 + overage)
Pro15%$19
Teams5%$39

Brain Revenue by Scale

UsersMonthly RevenueAnnual Revenue
1,000$10,100$121,200
10,000$101,000$1,212,000
100,000$1,010,000$12,120,000
1,000,000$10,100,000$121,200,000

Personas Revenue (Additional)

Average 1.5 personas per user at $9/month average:
UsersPersona Revenue/MonthPersona Revenue/Year
1,000$13,500$162,000
10,000$135,000$1,620,000
100,000$1,350,000$16,200,000
1,000,000$13,500,000$162,000,000

Storage Upsell Revenue

Assuming 10% of users purchase average 2 storage add-ons:
UsersStorage Revenue/MonthStorage Revenue/Year
100,000$100,000$1,200,000
1,000,000$1,000,000$12,000,000

Combined Revenue (Brain + Personas + Storage)

UsersMonthly RevenueAnnual Revenue
1,000$23,600$283,200
10,000$236,000$2,832,000
100,000$2,460,000$29,520,000
1,000,000$24,600,000$295,200,000

Infrastructure Costs

Compute Strategy

Self-hosted LLM (Ollama or similar) rather than paying per-token to third-party providers. Rationale:
  • Summarization and keyword extraction don’t require frontier models
  • Chinese models (DeepSeek, Qwen) cost ~$0.01 per 1M tokens
  • Self-hosted reduces this to fixed infrastructure cost
  • Break-even vs API pricing occurs around 200-400 active businesses

Compute Costs by Scale

UsersMonthly Infrastructure
1,000$3,000
10,000$8,000
100,000$50,000
1,000,000$150,000

Storage Architecture

StateWhat’s StoredSize
Active (recent)Full conversation, uncompressed~2MB
Archived (after X days)Full conversation, compressed~200KB (90% compression)
Search indexSummary + keywords + embeddings (internal only)~10KB

Storage Costs at 100,000 Users

Assuming 10,000 memories average per user, 90% archived:
TypeSizeCost/GB/MonthMonthly Cost
Hot (active + index)210 TB$0.02$4,200
Cold (archived)180 TB$0.004$720
Total390 TB$4,920

Storage Costs at Scale

UsersTotal StorageMonthly Cost
100,000~400 TB$5,000
1,000,000~4 PB$50,000

Profit Projections

At 100,000 Users

ItemMonthly
Revenue (Brain + Personas + Storage)$2,460,000
Compute$50,000
Storage$5,000
Total Cost$55,000
Profit$2,405,000
Annual Profit$28,860,000

At 1,000,000 Users

ItemMonthly
Revenue (Brain + Personas + Storage)$24,600,000
Compute$150,000
Storage$50,000
Total Cost$200,000
Profit$24,400,000
Annual Profit$292,800,000

Gross Margins

ScaleGross Margin
1,000 users~70%
100,000 users~98%
1,000,000 users~99%

Technical Architecture

Integration Model

Brain operates as an MCP (Model Context Protocol) server:
  • Integrates the same way as Google Calendar, Gmail, GitHub MCPs
  • Works with Claude natively
  • Other platforms integrate as MCP adoption grows
  • No proprietary protocol needed; uses existing standard

Memory Operations

OperationWhat Happens
StoreFull conversation saved, compressed after X days, summary and keywords generated internally for search
SearchInternal summaries and keywords searched, relevant memories identified
RetrieveFull conversation decompressed and returned to active context

What Users See (Metadata Only)

  • Memory exists
  • Token count
  • Attachments
  • Timestamp
  • Source conversation

What Stays Hidden (Trade Secret)

  • Summaries (internal indexing only)
  • Keywords (internal indexing only)
  • Ranking/relevance algorithm
  • Compression method
  • Retrieval logic
  • Storage architecture

Competitive Protection

Strategy: Trade Secret Over Patent

ApproachTrade-Off
PatentRequires full public disclosure; 20-year protection in theory
Trade SecretNo disclosure; protection lasts as long as secret is kept
Patents are not pursued because:
  • Enforcement against well-funded competitors costs millions
  • International coverage is limited (China doesn’t honor US patents)
  • Competitors can design around patents
  • Software patents are increasingly weak in courts
  • Filing requires disclosing exactly what we’re protecting

Protection Layers

ComponentStrategy
Core memory mechanismTrade secret
Brand (aiConnected, Brain)Trademark
CodeCopyright (automatic)
UI/UX innovationsPossibly patent if unique enough

What Protects Us Over Time

TimeframePrimary Moat
Year 1Secrecy + speed to market
Year 2User data gravity (memories accumulate)
Year 3+Ecosystem lock-in (personas, integrations, agencies)

Why Competitors Can’t Easily Replicate

FactorProtection
No public spec6-12 months of guessing for competitors
Hidden summarization logicThey build inferior version first
No insight into search qualityTrial and error required
Data gravityEven if replicated, user memories don’t transfer

Strategic Position

The Gap We Fill

PlatformTheir MemoryThe Limitation
OpenAIChatGPT memoryOnly works in ChatGPT
AnthropicClaude memoryOnly works in Claude
GoogleGemini memoryOnly works in Gemini
MicrosoftCopilot memoryOnly works in Microsoft ecosystem
aiConnectedBrainWorks everywhere
The major platforms have zero incentive to make memory portable. Their business model depends on keeping users inside their ecosystem. This interoperability gap is permanent and is our entire market.

Brain Within aiConnected

Brain is not a standalone business. Brain is the foundation of the cognitive operating system.
aiConnected Cognitive Operating System

    ├── Brain (memory, continuity, learning)

    ├── Knowledge (retrieval, expertise)

    ├── Voice (verbal communication)

    ├── Chat (text communication)

    ├── Personas (identity, behavior)

    └── Future: Vision, Motor Control (robotics)

Long-Term Vision

aiConnected is building the cognitive operating system for the physical AI era. Today’s products (Brain, Knowledge, Voice, Chat) generate revenue and establish the memory layer. Tomorrow, this same infrastructure powers robotics:
ComponentRole in Robotics
BrainRemembers tasks, learns from experience, maintains continuity
KnowledgeAccesses manuals, procedures, domain expertise
VoiceVerbal interaction with humans
ChatText-based commands, logging, reporting
PersonaConsistent personality, appropriate behavior per context
Any robotics company can build a body. Any robotics company can license an LLM. Nobody else is building the integrated cognitive stack that connects persistent memory, retrievable knowledge, multi-modal communication, and consistent identity.

Revenue Comparison

Brain vs Agency Platform

BusinessRevenue at Scale
Agency platform (100 agencies × $500/mo)$600,000/year
Brain (100,000 users)$29,520,000/year
Brain is 49x larger at modest scale. However, Brain remains one product within aiConnected because the long-term goal requires the complete cognitive stack.

Revenue Layers

LayerRole
BrainFoundation, entry point, highest retention
PersonasMonetization layer on top of Brain
Agency PlatformCash flow engine, funds Brain development
Knowledge/Voice/Chat APIsDeveloper and platform revenue
Robotics Cognitive LayerFuture enterprise and manufacturer revenue

The Moat

  1. Data Gravity: User memories accumulate in Brain. Switching means losing them.
  2. Persona Relationships: A Writing Partner with 8 months of context on your book can’t be replicated elsewhere.
  3. Cross-Platform Freedom: Ironically, Brain lets users use any LLM they want, but Brain itself becomes irreplaceable.
  4. Network Effects: Teams tier means shared memories. A company’s institutional knowledge lives in Brain.
  5. Ecosystem: Agencies building on aiConnected infrastructure aren’t switching because OpenAI added memory to ChatGPT.

Summary

Brain is the persistent memory layer for all AI. It solves the fragmentation problem that major platforms will never solve because doing so conflicts with their business model. At scale, Brain alone generates $290M+ annual profit. Combined with Personas and the broader aiConnected product suite, the business supports the long-term vision of becoming the cognitive operating system for AI and robotics. The protection strategy is trade secret over patent. Secrecy buys time; data gravity makes the moat permanent. By the time anyone reverse-engineers a comparable system, users have years of memories they won’t abandon. aiConnected: All AIs connected through persistent memory.
Document Version: 1.0 Last Updated: April 2026
Last modified on April 18, 2026