Technical SEO
48 terms in this domain · show all 220
C
-
Canonical chain
A canonical chain occurs when URL A canonicalizes to B, and B canonicalizes to C. Google may follow chains for a few hops, but treats long chains as signal noise and may pick a different canonical entirely. Always point each canonical directly at the final intended destination.
-
Canonical tag
A canonical tag (`<link rel="canonical" href="...">`) tells search engines which URL is the preferred version when near-duplicate content exists on multiple URLs. It is a hint, not a directive: Google may override it when its own signals point elsewhere.
-
Canonicalization
Canonicalization is the process by which Google selects one URL from a set of duplicate or near-duplicate URLs to represent the cluster in search results. The chosen URL is the canonical; the rest are alternates. Google weighs multiple signals — rel=canonical, redirects, sitemaps, internal links — to decide.
-
Crawl budget
Crawl budget is the combination of crawl capacity (how much Googlebot can crawl without overloading a server) and crawl demand (how much it wants to). Under about 10,000 URLs it rarely matters; above that threshold, it decides which pages ever reach the index.
-
Crawl trap
A crawl trap is any URL pattern that generates an unbounded or pathologically large set of URLs. Common sources: infinite calendar archives, faceted navigation combinatorics, session IDs in query strings, and recursive relative-link bugs. Crawl traps drain crawl budget and frustrate indexing of pages that actually matter.
-
Crawlability
Crawlability is whether a URL can be discovered and fetched by a search engine crawler. It depends on internal linking exposure, robots.txt rules, DNS and server availability, and HTTP response codes. A URL can be crawlable without being indexable — and the reverse is also true.
D
E
F
H
-
HSTS (HSTS)
HSTS (HTTP Strict Transport Security) is a response header that tells browsers to load a domain only over HTTPS for a declared duration. Defined in RFC 6797. It eliminates the HTTP-to-HTTPS redirect on repeat visits and blocks downgrade attacks. Sites can submit to the HSTS preload list for first-visit protection.
-
HTTP 301 redirect
A 301 is the HTTP "Moved Permanently" status code. It signals that a URL has changed location for good and clients should update bookmarks and references. Since 2016 Google has treated 301s as passing effectively all link equity to the target. The default redirect for permanent moves.
-
HTTP 302 redirect
A 302 is the HTTP "Found" status — a temporary redirect. It tells clients the original URL is the canonical one and they should keep using it. Google historically treated 302s differently from 301s, but since 2016 a long-lived 302 is consolidated to the target like a 301. Use only for true temporary moves.
-
HTTP 308 redirect
A 308 is the HTTP "Permanent Redirect" status code, defined in RFC 7538 (2015). It behaves like a 301 — permanent move, full link equity passed — except it preserves the original HTTP method. A POST stays a POST. The right choice for permanent redirects on APIs and form endpoints.
-
HTTP 404 error
A 404 is the HTTP "Not Found" status code. It tells clients the URL doesn't exist on this server right now and may or may not return. Google treats 404 as a soft signal — pages are eventually deindexed, but slowly. Healthy sites have some 404s; the problem is patterns, not individual hits.
-
HTTP 410 Gone
A 410 is the HTTP "Gone" status code — an explicit "this URL was here, has been intentionally removed, and is not coming back". Stronger than 404, which is ambiguous about future availability. Google has confirmed it deindexes 410s faster than 404s in practice.
-
HTTP 503 error
A 503 is the HTTP "Service Unavailable" status code. It tells clients the server is up but cannot handle the request right now and to try again later. Paired with a Retry-After header, it's the SEO-safe maintenance signal — Googlebot pauses crawling without deindexing.
-
HTTP status codes
HTTP status codes are three-digit responses servers return for every request, grouped into five classes: 1xx informational, 2xx success, 3xx redirection, 4xx client errors, 5xx server errors. Googlebot reads them on every fetch to decide whether to index, recrawl, drop, or back off.
I
-
Indexability
Indexability is whether a crawlable URL is eligible to appear in a search engine's index. Blocked by noindex directives, canonical tags pointing elsewhere, thin or duplicate content, or quality filters. Crawlability is a prerequisite but not a guarantee.
-
IndexNow
IndexNow is an open push-based indexing protocol launched by Microsoft and Yandex in October 2021. Sites publish a key file at the root and POST changed URLs to IndexNow endpoints; participating search engines fetch them on notification rather than waiting for the next crawl cycle.
-
Infinite scroll SEO
Infinite scroll SEO is the problem of making continuously-loading list pages crawlable. Googlebot doesn't scroll — it requests URLs. A pure JS infinite scroll where new items load on scroll without changing the URL is invisible to crawlers beyond the first page. The fix: paginated alternative URLs.
J
-
JavaScript SEO
JavaScript SEO is the practice of making JS-rendered content discoverable, crawlable, and indexable by search engines. It covers server-side rendering, client-side rendering, hydration, dynamic rendering, and the trade-offs each rendering strategy creates for crawlers like Googlebot, Bingbot, and AI bots.
-
JSON-LD
JSON-LD (JSON for Linked Data) is a JSON-based format for machine-readable annotations, standardized by W3C in 2014. In SEO it's the preferred way to deliver Schema.org structured data — a `<script type="application/ld+json">` block, typically in the HTML `<head>`, decoupled from the visible DOM.
L
-
Last-Modified header
Last-Modified is an HTTP response header containing the date a resource last changed (RFC 7232 format). On subsequent requests, clients send `If-Modified-Since` with that date; the server returns 304 Not Modified if nothing has changed. Googlebot honors this and saves crawl capacity on stable URLs.
-
Log file analysis
Log file analysis is the parsing of web server access logs to see exactly which URLs search crawlers (Googlebot, Bingbot, others) have requested, at what rate, and with what response codes. It's the only ground-truth source for what crawlers actually do on your site.
M
-
Mixed content
Mixed content is an HTTPS page that loads subresources — images, scripts, stylesheets, iframes — over plain HTTP. Browsers block active mixed content by default since Chrome 80 (Feb 2020) and progressively block passive mixed content too. It breaks the HTTPS lock icon and degrades the security signal Google has used in ranking since 2014.
-
Mobile-first indexing
Mobile-first indexing means Google primarily uses the mobile version of a page for indexing and ranking. Announced 2016, default for new sites since July 2019, full rollout completed in 2023. Content, structured data, and links must exist on the mobile version or Google won't see them.
N
O
P
-
Pagination SEO
Pagination SEO is the handling of paginated series — category pages, archives, search results — split across `/page/2`, `/page/3`, etc. Google deprecated `rel=prev/next` for indexing in 2019. Each page should self-canonical. View-all is an option only when the full list is reasonably small.
-
Parameter handling
Parameter handling is how Google decides which URL parameters change content meaningfully versus which are noise. Sort orders, filters, session IDs, tracking codes, and pagination all use parameters. The GSC parameter tool was deprecated in April 2022 — Google now relies on its own algorithms.
R
-
Redirect chain
A redirect chain is two or more sequential HTTP redirects — URL A → 301 → B → 301 → C. Each hop adds latency, consumes crawl budget, and risks clients stopping mid-chain. Most crawlers abandon the chain after 5-10 hops; users wait through every one.
-
rel=canonical
`rel="canonical"` is the HTML attribute declaring which URL is the preferred version among duplicates or near-duplicates. It's the technical mechanism behind the canonical-tag concept. Available as `<link rel="canonical">`, as an HTTP `Link` header, and implicitly via XML sitemap entries.
-
Robots.txt
Robots.txt is a plain-text file at the root of a host (`/robots.txt`) that tells crawlers which paths they may or may not fetch. It controls crawling, not indexing — a blocked URL can still appear in search results if it has inbound links.
S
-
Schema.org
Schema.org is the shared vocabulary — types and properties — maintained by Google, Microsoft, Yahoo, and Yandex to describe entities on the web. Launched 2011. Machine-readable annotations using this vocabulary are what search engines parse into rich results.
-
Sitemap index
A sitemap index is an XML file that lists multiple individual sitemap files rather than URLs directly. Required when a single sitemap would exceed the limits of 50,000 URLs or 50 MB uncompressed. The mandatory pattern for any site whose URL count puts it past those caps.
-
Soft 404
A soft 404 is a URL that returns HTTP 200 but the content tells the user the page doesn't exist — empty product page, "no results found", a generic template with no real content. Google detects soft 404s heuristically and flags them in Search Console under "Page indexing".
-
SPA SEO (SPA)
SPA SEO is the practice of making Single-Page Applications (React, Vue, Angular, Svelte) discoverable and indexable. The rendering strategy is the primary decision: SSR, SSG, hybrid, or dynamic rendering. Pure client-side rendering forces Google's WRS into a second-wave fetch, slowing indexing.
-
Structured data
Structured data is machine-readable metadata embedded in a web page that describes its content in a standardized vocabulary (usually Schema.org). It enables rich results — stars, prices, events, FAQs — in the SERP and helps search engines understand the page's entities.
-
Subdomain vs subdirectory
The choice between hosting content at `blog.example.com` (subdomain) or `example.com/blog` (subdirectory). Google states it treats both equivalently. Practitioners report subdirectories consistently inherit site authority more directly. For most SEO-driven decisions, default to subdirectory.
T
-
TLS / SSL (TLS)
TLS (Transport Layer Security) is the cryptographic protocol that secures HTTPS connections. SSL (Secure Sockets Layer) is the legacy term still used colloquially — actual SSL is deprecated; modern sites use TLS 1.2 or 1.3. HTTPS has been a confirmed (minor) Google ranking signal since August 2014.
-
Trailing slash
The trailing slash is the `/` character at the end of a URL path (`/about/` vs `/about`). To Google, the two forms are different URLs and can serve different content. Pick one canonical form site-wide and 301 the other. Mixing both produces duplicate-content signals and splits internal link equity.
U
V
-
Video schema
Video schema is the schema.org/VideoObject structured-data type. It tells Google what a video is, when it was uploaded, how long it runs, what thumbnail to show, and where the playable file lives. Required for video rich-result eligibility, key-moments, and live-stream badges.
-
Video sitemap
A video sitemap is an XML sitemap extension that lists each page on the site that contains a video, along with the video's metadata: thumbnail, title, description, duration, content URL, player URL, age restriction, and tags. It's the single cheapest way to surface video content to Google.
-
Viewport meta tag
The viewport meta tag is the single HTML element that tells mobile browsers how to scale and lay out a page: `<meta name="viewport" content="width=device-width, initial-scale=1">`. Without it, the page renders at a desktop default width (~980px) and gets flagged as not mobile-friendly under mobile-first indexing.
W
X
-
X-Robots-Tag
X-Robots-Tag is an HTTP response header that delivers crawler directives at the server level. Functionally equivalent to a meta robots tag but works on any resource, including PDFs, images, and other non-HTML files where meta tags cannot be embedded. Supports the same directives: noindex, nofollow, etc.
-
XML sitemap
An XML sitemap is a machine-readable file listing URLs that a site wants search engines to discover. A single sitemap holds up to 50,000 URLs or 50 MB uncompressed; larger sites use a sitemap index that references multiple sitemap files.