BERT algorithm(BERT)
BERT stands for Bidirectional Encoder Representations from Transformers. It is the language model Google rolled out to its search ranking and snippet systems in October 2019. BERT improved Google's understanding of prepositions, conjunctions, and overall query context — especially for conversational queries.
Long definition
BERT was published as a research paper by Google AI in October 2018 and rolled into Google Search exactly one year later (October 25, 2019), starting with English in the United States and expanding to 70+ languages by December 2019. At launch, Google said BERT affected roughly 10% of all search queries — the largest single change to Search since RankBrain in 2015.
The technical innovation is bidirectional context. Earlier language models (including the unidirectional precursors of BERT) read sentences left-to-right or right-to-left and built representations word-by-word. BERT processes the whole sentence at once via the Transformer architecture's self-attention mechanism, so the meaning of "bank" can be informed by the word "river" appearing later in the same sentence.
Concrete impact on search:
- Queries like "2019 brazil traveler to usa need a visa" — BERT correctly understands the directionality (Brazilian going to USA, not American going to Brazil). Pre-BERT, the "to" was often ignored.
- Conversational long-tail queries — multi-word questions with prepositions and modifiers improved noticeably.
- Featured snippets — BERT improved snippet selection by better matching question intent to passage meaning.
BERT didn't replace prior ranking signals; it added a layer of natural-language understanding on top. RankBrain (2015) was Google's first major ML signal in ranking; BERT was the first major application of Transformer-based language models. MUM (announced 2021) is the next-generation model claimed to be 1000x more capable than BERT and forms the foundation of features like AI Overviews and Things to Know.
Google initially described BERT as a "way of understanding queries", not a single algorithm replacement. It runs on the query side (parsing what the user asked) and on the content side (parsing passages in the index). This is what enabled the passage indexing rollout in February 2021 — Google can now rank specific passages of long pages for queries even when the whole page isn't optimized for them.
Common misconceptions
- "BERT replaced RankBrain." It didn't. They run side-by-side. RankBrain helps interpret novel queries; BERT helps with linguistic comprehension of the query and content. Both feed into ranking signals.
- "You can optimize content for BERT." Not in any specific way. BERT understands natural language, so writing naturally and answering questions clearly is the only "optimization." There is no BERT keyword strategy.
- "BERT is the AI behind ChatGPT/AI Overviews." No. BERT is an encoder model; it represents text but doesn't generate it. Generative features like AI Overviews use successor models (PaLM, Gemini) and the MUM-derived stack. BERT is purely for understanding/ranking.
- "BERT applies to every query." At launch, Google said it affected ~10% of queries — typically conversational and long-tail queries. Short keyword-style queries see less BERT effect because there's less natural-language context to parse.
Continue exploring