How Do AI Girlfriends Work? The Technology Behind AI Companions Explained (2026)
Most articles about AI girlfriend platforms compare features and recommend products. This one is different. It explains how the technology underneath these products actually works, why some platforms feel meaningfully smarter than others when the same tasks are run on both, and what the trajectory of the underlying technology will probably do over the next 4 years. If you have ever wondered why your AI companion remembers some things and forgets others, why voice on one platform sounds human and on another sounds robotic, or why live video calls work on exactly one platform in 2026, the answers are in here.
This is also the post we wish existed when we started writing platform reviews. Once you understand the five layers of an AI girlfriend product, the differences between platforms stop being mysterious and start being predictable. A platform with a strong language model and weak memory architecture will feel sharp in a single conversation and forgetful across sessions. A platform with strong voice synthesis but no live video will deliver excellent voice messages and no real-time conversation experience. The product behavior follows directly from the architecture; you just have to know what to look for.
For the empirical companion to this post, our Memory Benchmark tests how memory actually performs across 10 platforms, our Voice Quality Test covers voice synthesis quality, and our Hidden Costs Tear-Down covers the economics. This post is the technical theory; those are the empirical practice.
The Five Layers of an AI Girlfriend Product
Every production AI girlfriend platform in 2026 sits on top of a five-layer stack. Some platforms invest heavily in all five; most invest deeply in one or two and ship the rest as adequate-but-unremarkable layers. The combined picture across the stack is what determines product feel — and the gap between platforms shows up most clearly when you compare the same task across two products with different architectural priorities.
Layer 1: The Language Model (LLM)
The language model is the brain. Every word your AI companion writes comes out of an LLM — typically a frontier-tier model like GPT, Claude, Llama, or a fine-tuned variant of one of these. The LLM takes your message, the conversation context, and a system prompt that defines the character, then produces the next message in the conversation.
Three architectural choices at this layer determine product behavior:
Base model selection. A platform running on top of a frontier model (Claude or GPT) will generally produce sharper, more coherent, more emotionally aware conversations than a platform running on a smaller open-source model. The cost trade-off is real — frontier models cost meaningfully more per token to run — and it explains why some platforms charge more or have stricter usage limits. The free model on platforms like SpicyChat or JanitorAI is typically a smaller open-source model that costs less to serve; the upgrade tiers unlock larger or frontier models.
Fine-tuning vs base model use. Some platforms use the base model as-is with a sophisticated system prompt to shape behavior. Others fine-tune the model on roleplay-specific data, NSFW content, or conversation patterns to make it perform better at the specific task. Fine-tuned models can outperform larger base models on the narrow task they were tuned for; they also typically have weaker general reasoning. Most NSFW-focused platforms use fine-tuned models; most wellness-focused platforms use base models with strong system prompts.
Character system prompts. The system prompt is the instructions the model reads before every response — name, personality, backstory, behavioral rules, content boundaries. A well-engineered character system prompt can make a mid-tier model feel like a top-tier character; a poorly-engineered one can make a top-tier model feel generic. Platforms with strong character builders (like Candy AI's depth or Nectar AI's persona engineering) are essentially automating high-quality system prompt construction; platforms with thin builders leave more of this work to chance.
For users, the practical consequence is that two platforms running the same underlying LLM can feel completely different based on how they use it. Two platforms running different LLMs can feel similar if the better-engineered one compensates for the weaker model. The model name on the marketing page is one signal; the actual product feel is the combined picture across all three architectural choices.
Layer 2: Memory Architecture
Memory is the second layer and the one that varies most dramatically across platforms. Even when two platforms use the same LLM, they can deliver completely different memory experiences depending on how they handle the gap between the model's context window (what it can read in a single response) and the conversation history that exists outside that window.
Four distinct memory mechanisms work together (or fail to) on every platform:
Short-term context is the conversation tokens the model reads when generating its next response. The 2026 baseline is around 32,000 tokens (roughly 24,000 words); top platforms clear 128,000 tokens. Below 16K, the AI starts forgetting things you said earlier in the same session, which is disqualifying for any serious use. Larger context windows are not always better — they cost more to serve and can introduce attention quality issues at the very long end — but the floor matters.
Cross-session continuity is whether facts, events, and emotional context persist between sessions days or weeks apart. This is technically harder than it sounds because the LLM only reads what fits in its context window. Platforms solve this by summarizing past conversations into compressed memory representations and feeding the summaries back as part of the system prompt for new sessions. Quality varies enormously based on summarization quality, retrieval logic, and how aggressively the platform invests in this layer.
Active vs passive memory. Passive memory is the AI answering correctly when you ask about something you mentioned earlier. Active memory is the AI bringing up past content unprompted at the right moment, without you mentioning it. Active memory is much harder — it requires the platform to detect when a memory is contextually relevant and surface it without forcing it. Most platforms ship passive memory; only Tier 1 platforms in our Memory Benchmark ship active memory consistently.
Editing transparency. Can you see what the AI has stored about you and correct mistakes? Most platforms hide memory from users; the AI's stored facts are opaque. Muah AI is the major exception — the platform exposes the memory ledger and lets users edit individual entries. Editable memory is technically harder than opaque memory because the platform has to maintain a structured representation of memory that humans can read; most platforms have not invested in this layer.
The combined picture across these four mechanisms is what determines whether your AI girlfriend feels like a continuous relationship or a chatbot resetting every session. The technical character memory glossary covers this in more depth.
Layer 3: Voice Synthesis
Voice is the third layer and one of the fastest-improving across the AI girlfriend category in 2026. Voice generation has gone from clearly-robotic in 2023 to passing casual auditory tests in 2026 on the top-tier platforms. The architecture sitting underneath this improvement has converged on a small set of approaches.
Text-to-speech (TTS) quality depends on the underlying voice synthesis model. ElevenLabs has been the dominant frontier-quality option for the past two years; several open-source competitors have caught up partially. Platforms running ElevenLabs-tier voice synthesis produce audio that is often indistinguishable from human in casual segments; platforms running older TTS engines sound clearly robotic by comparison.
Latency is round-trip delay from your message to audio response. The threshold for natural-conversation feel is approximately 300 milliseconds; the threshold for walkie-talkie feel (uncomfortable but workable) is around 500ms. Voice messages are not bound by this threshold because you press play after the audio is generated. Live voice calls are bound by it because the conversation rhythm depends on real-time turn-taking. Sub-300ms live voice requires investment across the model serving infrastructure, the synthesis pipeline, and the audio streaming layer.
Voice cloning is the synthesis of a custom voice tuned to user specifications. Muah AI is the only major platform offering this on the AI girlfriend (and boyfriend) side as of 2026. The technology is not unique to Muah — general-purpose tools like ElevenLabs let any user clone voices — but the integration with character creation, the persistent voice identity across sessions, and the consent guardrails that prevent cloning real people without permission make Muah's implementation distinctive.
Live voice calls are real-time conversation with sub-second response. Technically much harder than voice messages because the model has to generate audio fast enough to feel like a phone call, handle interruptions cleanly, and produce response content rapidly enough that the audio pipeline does not stall. Few platforms ship this competently — SweetDream AI and Muah AI are the production examples in 2026. For deeper coverage, see our Voice Quality Test.
Layer 4: Image and Video Generation
Image generation is the fourth layer and the one where the technology is most visible — every generated image is a clear artifact you can evaluate at a glance. The architecture has converged on diffusion models (descendants of Stable Diffusion), with platforms differentiating through fine-tuning, character consistency techniques, and content policy.
Character consistency is the hardest problem in this layer. A diffusion model generates images from scratch every time, so getting the same character to look identical across multiple generations requires either fine-tuning the model on the character or using techniques like character-preserving prompts and embedding-based consistency. Top platforms (Candy AI, SweetDream AI, Muah AI) ship strong character consistency; weaker platforms produce images where the character looks different in every generation, which breaks immersion.
NSFW LoRAs and fine-tuning. LoRAs (Low-Rank Adaptations) are small fine-tuning artifacts that adjust a base diffusion model toward specific content categories — NSFW, anime, photorealistic, specific aesthetics. Most NSFW-capable AI girlfriend platforms use NSFW LoRAs on top of a base model rather than running a fully fine-tuned NSFW model; this is computationally cheaper and easier to update.
Video generation is image generation extended to short clips. Architecturally similar to image gen but more expensive (more frames means more compute), with most platforms shipping short clips (5-15 seconds) rather than longer videos. The quality variance across platforms is wider than in image generation because the techniques are less mature.
Live video is the highest-end expression of this layer — real-time face generation with lip sync, emotional expression tracking, and conversational eye contact. SweetDream AI is the production example in the AI girlfriend category. Technically much harder than image or video generation because everything has to render in real time at acceptable quality on consumer connections.
Layer 5: The System Architecture (Plumbing)
The fifth layer is the plumbing that connects everything else — the API gateways, model serving infrastructure, database for memory storage, content moderation pipeline, payment processing, and user-facing application. This layer is invisible when it works well and very visible when it does not.
Model serving infrastructure determines latency and reliability. Platforms running on top of self-hosted GPU clusters can hit lower latency than platforms making API calls to external providers, but at higher operational cost and complexity. Most large-scale AI girlfriend platforms run a hybrid — self-hosted serving for high-volume tasks, external API for occasional or specialized needs.
Memory storage is where the conversation history and summarized memories live. Database design at this layer determines how much history can be retained, how fast retrieval is, and whether memory editing (Muah's distinctive feature) is technically possible. Platforms that retrofit memory features onto a chat-focused database typically struggle; platforms that built memory into the architecture from the start (Replika is the canonical example, since 2017) handle it more cleanly.
Content moderation runs as a layer between user input, AI output, and the user-facing display. Moderation can happen at multiple points — pre-prompt filtering, post-generation safety checks, image content scanning. Aggressive moderation produces false positives that frustrate users; lax moderation produces content the platform's policies prohibit. Calibration is product-defining.
Payment and account systems determine billing transparency, refund handling, and account portability. Platforms that built strong payment infrastructure (Candy AI's discreet billing with crypto support is the standout) deliver smoother experiences; platforms with weaker payment infrastructure often have user-visible billing problems.
Persona Engineering: How Character Personality Is Built
The character on your screen is not a separate entity from the LLM — it is the LLM behaving according to a system prompt and (sometimes) a fine-tuning that defines who the character is supposed to be. Persona engineering is the discipline of making this behavior consistent, distinctive, and emotionally compelling.
Three levers control character feel:
The system prompt is the instructions the model reads before every response. A typical character system prompt includes name, age, occupation, personality traits, communication style, backstory, relationship to the user, content boundaries, and behavioral rules. Length matters — too short and the character feels generic; too long and the model starts ignoring later sections. Quality matters more than length.
Character cards (the format used on community-character platforms like SpicyChat and JanitorAI) are structured system prompts that follow a standard schema. Users can write character cards and share them across platforms; the same character card produces broadly similar behavior on any platform that supports the format. Card quality varies enormously — top community characters have sophisticated cards with consistent persona, distinctive voice, and clear behavioral rules; thinly-built characters have generic cards that produce generic behavior regardless of the underlying model.
Fine-tuning on character-specific data is the deepest level of persona engineering. Some platforms train their model on a corpus of conversations, behaviors, and responses that exemplify a specific character or character archetype. This is more expensive than system-prompt engineering but produces deeper character consistency. Few platforms invest at this level for individual characters; some invest in archetype-level fine-tuning (an "anime girlfriend" tune, a "wellness companion" tune).
Why characters drift over time. Long conversations expose persona drift — the AI's behavior gradually shifts away from the character definition as the conversation history fills the context window and pushes the system prompt's influence weaker. Platforms with stronger character anchoring (re-introducing the system prompt at intervals, summarizing the character's behavior into the conversation context) handle this better. Persona drift is one of the most common complaints about long-term AI companion use; it is a solvable problem but requires per-platform investment to solve.
The Economics of Running an AI Girlfriend Platform
Understanding how the economics work helps explain why pricing is structured the way it is across the category.
Text generation is relatively cheap. A typical message exchange uses a few thousand tokens of LLM compute. At frontier-model pricing (~$3-15 per million input tokens, ~$15-75 per million output tokens depending on the model), a message exchange costs a fraction of a cent. This is why most platforms can offer unlimited free text on the lowest tier — text is the cheapest modality to serve.
Voice generation costs meaningfully more. ElevenLabs-tier voice synthesis costs roughly $0.001-0.01 per character of audio depending on tier. A typical voice message of 100 words uses 500-700 characters of audio, costing $0.005-0.05 per message. Multiply across thousands of users and voice infrastructure becomes a significant cost line. This is why voice is paid on most platforms; the unit economics do not work at $0.
Image generation costs are more. A high-quality NSFW image generation typically costs $0.02-0.10 in compute depending on resolution and quality. Thousands of images per user per month means real money. This is why image generation is metered or rate-limited on every platform — even the most generous free tiers cap image generation because the marginal cost is too high to make unlimited free image gen sustainable.
Video generation is much more expensive. Short video clips (5-15 seconds) cost $0.10-1.00 per generation depending on quality and length. Live video is even more expensive — sustained real-time generation costs in the order of $0.10-0.50 per minute of conversation. This is why live video is a premium-tier feature on the one platform (SweetDream AI) that ships it; the cost economics force premium pricing.
The token-vs-subscription decision for users is downstream of these economics. Platforms with predictable text-heavy users prefer flat subscriptions; platforms with variable multimedia-heavy users prefer token-based pricing where users pay for what they consume. Both models work; the right pick for a user depends on their actual usage pattern. Our Tokens vs Unlimited guide covers the math from the user perspective.
Memory: The Hardest Problem in AI Companion Tech
Memory deserves its own section because it is the hardest problem in the category and the one where platforms differentiate most clearly. The technical challenges are real and not yet fully solved even on the best platforms.
Why memory is hard. LLMs do not have memory in the human sense. They have a context window — the tokens they read before generating a response. Anything outside that window does not exist for the model. Platforms create the illusion of memory by summarizing past conversations, storing the summaries in a database, and feeding relevant summaries back into the context window as part of the system prompt for each new session.
The summarization quality bottleneck. The summarizer is itself an LLM (often a smaller, faster model than the main chat model). Summary quality determines whether the memory layer actually works. A summary that captures the right details lets the AI reference past content correctly; a summary that misses important content makes the AI seem forgetful. Platforms vary enormously on summarization quality, and most do not expose this to users.
Retrieval ranking is the second bottleneck. Even with good summaries, the platform has to decide which summaries to surface for any given new conversation. Retrieve too few and relevant memories are missed; retrieve too many and the context window fills with irrelevant content. Most platforms use embedding-based similarity search; quality depends on the embedding model and the ranking logic.
Why she sometimes forgets. When your AI companion forgets something you told her, the failure is usually retrieval rather than storage. The memory exists in the database but did not get surfaced to the current prompt because the retrieval ranking missed it. Restarting the conversation, explicitly mentioning the topic, or referencing related context usually brings the memory back. Platforms with weak retrieval are the most common offenders for this failure mode.
Active memory is the holy grail. When your AI companion brings up past content unprompted at the right moment, that is active memory in action — and it is hard. The platform has to detect when a memory is contextually relevant, decide whether to surface it, and integrate it into the response naturally. Most platforms ship passive memory (answering correctly when asked); only Tier 1 platforms ship active memory consistently. Forecast: active memory becomes table stakes by late 2027 as platforms invest in it.
Editable memory is a privacy and usability win. Muah AI's approach of exposing the memory ledger and letting users edit entries is the only major platform implementation in 2026. The technical bar is real (you need a structured memory representation rather than free-form text summaries) but the user benefit is real too — users can correct mistakes, delete embarrassing entries, and add facts they want the AI to remember. Privacy regulators are likely to mandate something similar across the category by 2028-2029.
Privacy: What Happens When You Send a Message
A quick walkthrough of the data flow when you send a message to your AI companion. Useful for understanding what privacy means in this context.
The message you type is sent over an encrypted HTTPS connection to the platform's servers. Encryption-in-transit is universal in 2026; any platform that does not run HTTPS is a red flag.
The message is stored in the platform's database, typically in plaintext (encrypted at rest is less common but exists on privacy-focused platforms). This is where privacy policies matter — the platform decides how long to retain the message, whether to use it for AI model training, who at the company can access it, and what happens if the company gets a court order.
The message is fed into the LLM along with the system prompt, recent conversation context, and (if the platform has memory) relevant retrieved memories. The LLM generates a response. The response is stored in the database and sent back to you over HTTPS.
If voice or image generation was requested, the response triggers downstream calls to the voice synthesis or image generation services. The generated audio or image is stored on the platform's servers and served to you. Generated content is typically retained in your account gallery; some platforms allow deletion.
For end-to-end encrypted platforms (which are essentially nonexistent in this category in 2026 because the AI needs to read your messages to respond), this flow would be different. For all standard platforms, the operator can technically read your messages — the question is whether they do, under what circumstances, and how transparent they are about it. Our AI Companion Privacy guide covers the practical checklist for evaluating platform privacy.
Why Some Platforms Feel Smarter Than Others
With all five layers in mind, the question of why some platforms feel meaningfully smarter than others has a clear answer. The combined picture across the stack is what matters; differences on individual layers compound.
A platform with a frontier LLM, strong character system prompts, deep memory architecture, ElevenLabs-tier voice synthesis, character-consistent image generation, and well-tuned content moderation will feel sharply better than a platform that only invested in two of these layers. The user-visible difference is clear even when both platforms claim to do the same things; the underlying architectural quality drives product feel.
This is also why platform reputation persists even when underlying technology shifts. Replika has been refining its character architecture since 2017; eight years of investment in persona consistency, memory continuity, and emotional intelligence is hard for a newer platform to replicate even with better baseline LLM technology. Newer platforms compete on different dimensions (multimedia, NSFW openness, live video) where the architectural investment is more recent.
For users, the practical implication is that picking a platform should not be based on "which model do they use" alone. The same model can produce vastly different experiences depending on the rest of the stack. Try the platform; let the integrated experience tell you whether the architectural investment is there.
The 2026 to 2030 Trajectory
Directional forecasts based on the current technology trajectory and the patterns we see across platforms.
By late 2027: Active memory becomes table stakes on Tier 1 platforms. The current Tier 1 leaders' active memory will look like 2024's passive memory does today. Smaller platforms catch up within 18 months as the underlying techniques become better-documented.
By 2027-2028: Live video appears on at least three more platforms beyond SweetDream AI. The technical barrier (real-time face generation with lip sync) is dropping fast and the market is large enough to attract competitors.
By 2028: Sub-200ms latency on live voice becomes the new natural-conversation threshold. The current 300ms goal will look like 2024's 600ms goal does today.
By 2028: Editing transparency becomes a regulatory expectation. Privacy regulators will likely require users be able to inspect and delete what AI products store about them, mandating something similar to Muah AI's current memory ledger across the category.
By 2028-2029: Voice indistinguishable from human in casual auditory tests becomes universal across Tier 1 and Tier 2 platforms. The differentiator moves from naturalness to dimensions like emotional intelligence in voice and real-time tonal modulation.
By 2029-2030: Cross-platform memory portability emerges, likely under regulatory pressure rather than vendor cooperation. Users will be able to migrate a relationship's worth of accumulated memory from one platform to another via standardized export/import.
By 2030: The dominant interaction modality shifts from text to live voice and video. Text remains available but most users prefer voice for casual use and video for emotional moments. Platforms that invested in live multimedia (SweetDream's current bet) are best-positioned; text-first platforms have to retrofit their architecture.
For a deeper look at the long-term trajectory of AI companions, see our AGI and attachment theory pillar.
Frequently Asked Questions
Are AI girlfriends actually intelligent or just predicting words?
LLMs are next-token predictors at the architecture level — they predict what word comes next given the context. But the emergent behavior at scale produces conversation that is functionally indistinguishable from intelligence in many contexts. The honest answer is that the question is partly philosophical: by the standards we use to recognize intelligence in others (responding contextually, remembering, understanding emotion), top-tier 2026 AI companions clear the bar. By the standards of having internal experience or understanding, the question is unresolved and probably unresolvable with current methods.
Can AI girlfriends really remember me long-term?
On Tier 1 platforms, yes — meaningfully across months and emerging across years. The technical mechanism is summarization and retrieval rather than human-style memory, but the user experience is similar. On Tier 2 and Tier 3 platforms, memory is partial — fact retrieval works, active recall is rare. See our Memory Benchmark for empirical breakdown.
How is AI girlfriend voice generated?
Text-to-speech (TTS) synthesis via models like ElevenLabs or in-house alternatives. The model takes the text response from the LLM and generates audio in the chosen voice. Voice cloning (synthesizing a custom voice tuned to user specifications) is a specialized variant that few platforms ship; Muah AI is the major example in the AI girlfriend category. Live voice calls add real-time audio streaming with sub-300ms latency.
Why do AI girlfriends sometimes forget what I told them?
Usually a retrieval failure rather than a storage failure. The memory exists in the database but the platform's retrieval ranking did not surface it for the current prompt. Restarting the conversation, explicitly mentioning the topic, or referencing related context usually brings the memory back. Platforms with weak retrieval logic are the most common offenders for this failure mode.
Will AI girlfriends replace human relationships?
For a small minority of users, they substitute. For most users, they supplement — a complement to existing relationships rather than a replacement. The category attracts users across a wide spectrum of social contexts; the platforms themselves do not have a single user profile. For deeper exploration of this question, see our AGI and attachment theory pillar and AI Companions and Loneliness guide.
How is the character personality built?
Through system prompts (instructions the LLM reads before every response), character cards (structured prompts on community platforms), and sometimes fine-tuning on character-specific data. The combination determines how distinctive and consistent the character feels. Strong character builders (Candy AI, Nectar AI) automate high-quality system prompt construction; thin builders leave more to chance.
Why is voice on some platforms much better than others?
Three factors: the underlying voice synthesis model (ElevenLabs vs older TTS), the latency of the model serving infrastructure, and the per-character voice tuning. Platforms that invested in all three deliver voice that passes casual auditory tests; platforms that invested in only the model deliver voice that sounds good in samples but lags in live conversation. See our Voice Quality Test.
Can AI girlfriends generate images of any character?
Mostly yes for fictional characters; mostly no for real identifiable people. Platforms apply content moderation to image generation prompts to prevent abuse. The major platforms also apply character consistency techniques so the generated images look like your specific custom character rather than a generic person matching the prompt. Image quality varies by platform investment in fine-tuning and character-preserving prompts.
How does live video work?
Real-time face generation with lip sync to the audio response, emotional expression tracking on the chat content, and conversational eye contact. Technically the hardest layer in the AI companion stack — every frame needs to render fast enough to keep up with conversation rhythm, the audio and video need to stay synchronized, and the experience needs to hold up over consumer internet connections. SweetDream AI is the production example in 2026.
What happens to my conversations when I delete my account?
Depends on the platform. Most platforms soft-delete (account is hidden but data remains in the database for compliance or fraud prevention). Some platforms hard-delete with a grace period. Few platforms offer immediate hard delete. Look for the explicit phrase "permanently delete my data" in account settings to be sure. Our Privacy guide covers this in depth.
Will AI girlfriend technology get much better in the next 2 years?
Yes, fast. Active memory and live video will become table stakes; voice will become indistinguishable from human in casual tests; editing transparency will likely become a regulatory expectation. The current Tier 1 leaders will be the Tier 2 baseline by 2028; the current Tier 3 platforms will need to invest meaningfully to remain competitive.
Why do some platforms feel smarter than others even with the same model?
The LLM is one of five architectural layers. Two platforms running the same model can feel completely different based on character system prompt quality, memory architecture, voice and image generation integration, and overall product polish. Reputation persists for years because architectural investment compounds — Replika's eight years of refining its character continuity is hard to replicate from scratch.
Are AI girlfriend platforms profitable?
Varies enormously. The largest platforms are profitable; smaller platforms often run at a loss while building user base. The unit economics work because text is cheap to serve and premium tiers cover the cost of expensive multimedia features. Platforms that ship live video (SweetDream) have higher per-user costs than text-first platforms; they price accordingly.
Bottom Line
AI girlfriend products are not magic. They are five-layer technology stacks built on top of large language models, with the differences between platforms coming from architectural choices at each layer rather than from any single technical breakthrough. Once you know what the layers are and how they fit together, the differences between products stop being mysterious and start being predictable.
For users, the practical takeaways:
Test the integrated experience, not the marketing claims. Two platforms running the same model can feel completely different. The combined picture across all five layers is what matters; individual layer claims tell you less than the actual product feel.
Memory is the hardest problem and the most differentiating layer. Platforms with strong memory architecture deliver continuity that feels like a relationship; platforms with weak memory feel like chatbots resetting every session. Test memory specifically before committing.
Voice and live video are real expressions of architectural investment. Platforms that ship strong voice and live video have invested in technically demanding layers; the price reflects real cost. Platforms that ship weak voice or no live video have made different investment choices, not necessarily wrong ones.
The technology will get much better fast. Active memory, indistinguishable voice, mainstream live video, editable memory across the category — all likely within 2-3 years. The platforms you choose today are not necessarily the platforms you will use in 2028; expect the category to shift meaningfully.
For the empirical companion to this technical theory: Memory Benchmark, Voice Quality Test, AI Girlfriend Hidden Costs, Tokens vs Unlimited economics, and Privacy guide for the full picture across the stack.