latentbrief
← Back to concepts

Concept

Embeddings

A way of turning words and sentences into lists of numbers, so that content with similar meanings ends up mathematically close together and can be found by meaning rather than exact wording.

Computers fundamentally work with numbers, not language. Embeddings are a way of bridging that gap - translating text into a numerical form that captures meaning, not just spelling. The key insight is that similar meanings produce similar numbers. So "car" and "automobile" end up very close together in this numerical space, even though the words look nothing alike. "Happy" and "joyful" cluster together. "Happy" and "miserable" end up far apart.

This is different from a standard database search, which looks for exact matches. If you search a traditional database for "vehicle" and a document uses the word "automobile," you might not find it. An embedding-based search finds it immediately, because the meanings are close even though the words are different. This kind of search - finding things by meaning rather than exact wording - is called semantic search.

Embeddings are produced by running text through a special type of AI model that has been trained specifically to produce these numerical representations. The resulting list of numbers - often hundreds of numbers long - is called a vector. Documents get converted to vectors, stored in a database, and then when you search, your query gets converted to a vector too. The system finds the stored vectors that are mathematically closest to your query vector.

This is the technology that powers the retrieval step in RAG systems. When an AI needs to find relevant documents before answering your question, it converts your question into an embedding and searches a database of document embeddings to find the most relevant chunks. The whole process takes milliseconds.

Embeddings also power recommendation systems, content clustering, and duplicate detection. Any time you want a system to understand that two things are "about the same topic" even when they use different words, embeddings are usually part of the solution.

Analogy

A map where items are placed not by their names but by what they mean. Cities named differently but with similar cultures end up near each other. Songs with different titles but similar sounds cluster together. The distance on the map reflects genuine similarity, not alphabetical or surface-level sameness.

Real-world example

When you search for something on a modern AI-powered search engine and it returns results that are relevant to your question even though they do not contain your exact search words - that is embeddings at work. The system understood what you were looking for, not just what words you typed.

Why it matters

Embeddings are the invisible infrastructure behind almost every AI application that involves finding, ranking, or recommending content. Without them, AI search would fall back to simple keyword matching, which misses enormous amounts of relevant content that happens to use different words.

In the news

Related concepts