All articles
Voice Matching

Embeddings vs. Prompts: What 'AI That Learns You' Actually Means

Everyone claims their AI email tool 'learns your voice.' Here's the technical difference between prompt engineering, fine-tuning, and RAG, and which one actually works for email style.

4 min read·

Every AI email tool claims to "learn your voice." The claim is not always false, but it describes very different things depending on which tool is making it.

Here is the actual technical breakdown of what each approach does, what it can and cannot achieve, and why it matters for whether your emails actually sound like you.

Three approaches to "learning your voice"

Approach 1: Prompt engineering (most ChatGPT users)

This is what you are doing when you write a system prompt that describes your writing style. "Write in a warm, direct tone. Keep it under 100 words. Avoid em-dashes."

What it does: Shapes the statistical output of the base model using natural language instructions.

What it can do well: Eliminate obvious tells (em-dashes, filler phrases), set general tone, establish length constraints.

What it cannot do: Capture the fine-grained idiosyncratic patterns that are actually specific to you, because you cannot describe them in words. You do not know your average sentence length. You do not know whether you use em-dashes more with warm contacts than cold ones. You do not know the exact vocabulary that appears in your email more than in average professional email.

The ceiling: You will get generic professional email with your preferences applied. You will not get you.

Approach 2: Fine-tuning

Fine-tuning means updating a model's weights by training it further on your specific data. If you could fine-tune GPT-4o on your sent email history, the model would literally update its learned patterns toward yours.

What it does: Actually updates the model to match your patterns, not just apply instructions at inference time.

What it can do well: Potentially high-fidelity voice matching if the training set is large enough and representative.

What it cannot do: Scale affordably for individual users. Fine-tuning GPT-4o-scale models costs significant compute. It also does not update dynamically as your writing evolves.

The ceiling: Not viable for individual users. Enterprise deployments with very large email corpora can make it work, but this is not what consumer email tools are doing when they say "learns your voice."

Approach 3: RAG (Retrieval-Augmented Generation)

RAG is the approach that actually scales for individual voice matching. Here is how it works:

  1. Your sent emails are converted into vector embeddings: numerical representations of their semantic content that can be searched by similarity.
  2. When you draft a new email, the system searches your embeddings for your most relevant previous emails (similar recipient type, similar topic, similar relationship context).
  3. Those emails are injected into the context window as examples before generation begins.
  4. The model generates a new draft that is conditioned on your actual past writing, not just a style description.

What it does: Provides the model with concrete evidence of how you write in similar situations, not an abstract description of your style.

What it can do well: Captures the idiosyncratic patterns that prompt engineering cannot describe. Your opener habits per relationship type. Your specific vocabulary. Your length patterns across contexts. The structural choices you make for difficult asks vs. casual notes.

What it cannot do: Work without a sufficient sample of your sent emails, and work perfectly for novel situations that have no precedent in your history.

The ceiling: Currently the best available approach for individual email voice matching. Quality scales with the size and diversity of your email corpus.

Why this distinction matters

When you use a ChatGPT Custom GPT for email and it "learns your voice" from a knowledge file, you are using a degraded version of RAG: it retrieves from a small, curated, static sample rather than a large, updated, real-time corpus of your actual sent email.

When FinalDraft says it learns your voice, it is using proper RAG: your real sent email history, stored as embeddings, updated as you send more email, retrieved and injected per draft.

The difference in output is real. Generic-prompt email averages across millions of users. RAG-trained email is conditioned on your specific writing history.

The manual version of this insight

You cannot run embeddings on your own sent folder without infrastructure. But you can build a first-person prompt that captures more of your specifics than a generic style description. The Persona Prompt Generator is the manual version of this process.

It asks you the questions that surface your idiosyncratic patterns: how do you open different types of emails? What do your closers look like? What phrases do you avoid? What is your typical length for different relationship types?

The resulting prompt is not as powerful as a full RAG pipeline, but it is meaningfully better than "be warm and professional," and you can use it in any AI tool you already have.

Build your persona prompt →

More in Voice Matching

Free tool

Ready to stop sounding like everyone else?

Build a first-person persona prompt that captures your voice in under 5 minutes. No account required.

Generate my prompt