SEO Dictionary: The Terms That Matter
An SEO dictionary is a reference guide to the terminology used in search engine optimisation, covering everything from technical concepts to content strategy to link building. This one is built differently. Rather than listing every term alphabetically, it groups them by how they function in practice, so you can see how the concepts connect rather than treating each one as an isolated fact.
If you have ever sat in a meeting where someone dropped “topical authority” or “crawl budget” without anyone questioning whether they were using the term correctly, this guide is for you. Knowing what these terms mean is one thing. Understanding when they matter, and when they are being used to obscure rather than illuminate, is something else entirely.
Key Takeaways
- SEO terminology is often used loosely in agency and client conversations. Precise definitions protect you from bad decisions built on vague language.
- Many SEO concepts only make sense in relation to each other. Crawlability means nothing if your content has no authority. Authority means nothing if your pages cannot be indexed.
- Technical SEO terms are frequently overcomplicated. Most sites do not have crawl budget problems. Most do not need schema on every page. Knowing the term does not mean the tactic is relevant to your situation.
- The terms that matter most in commercial SEO are the ones tied to intent: what a searcher wants, what stage they are at, and whether your page actually delivers it.
- SEO vocabulary shifts faster than SEO fundamentals. The words change. Relevance, authority, and user satisfaction have been the core variables for over a decade.
In This Article
- Why SEO Terminology Gets Misused So Often
- Crawling, Indexing, and Rendering: The Foundation Layer
- Keyword and Search Intent Terminology
- On-Page SEO Terminology
- Authority, Links, and Off-Page Terminology
- Technical SEO Terminology
- Content and Architecture Terminology
- Measurement and Reporting Terminology
- Terms That Are Frequently Misused or Overstated
Why SEO Terminology Gets Misused So Often
I spent years running agencies where SEO was a core service line. One thing I noticed consistently: the more jargon an SEO practitioner used with a client, the less confident I became in their actual grasp of the subject. Complexity is sometimes necessary. But in SEO, complexity is often a performance, not a reflection of the work required.
When I grew the iProspect team from around 20 people to close to 100, one of the things we had to work hard on was plain language. Not dumbing things down, but being precise. There is a difference between a term that carries real meaning and a term that sounds authoritative without committing to anything. “We are improving your domain authority” is the latter. Domain Authority is a Moz metric. It is not a Google ranking factor. Using it as if it were is either sloppy or deliberately misleading.
This dictionary aims to be precise. Where a term is commonly misused, it says so. Where a concept sounds more technical than it is, it says that too. If you want a deeper look at how these concepts fit into a full search strategy, the Complete SEO Strategy hub covers the mechanics in more detail.
Crawling, Indexing, and Rendering: The Foundation Layer
Before Google can rank your page, it needs to find it, process it, and store it. These three steps are distinct, and confusing them leads to misdiagnosis when something goes wrong.
Crawling is the process by which Googlebot, Google’s automated web crawler, discovers and fetches pages. It follows links from known pages to new ones. If a page has no inbound links and is not in your sitemap, it is unlikely to be crawled.
Crawl budget refers to the number of pages Google will crawl on your site within a given timeframe. This is frequently cited as a concern when it rarely is for most sites. Unless you have hundreds of thousands of pages or significant crawl errors, crawl budget is probably not your problem. I have seen agencies bill for crawl budget audits on 200-page sites. That is not an SEO service. That is theatre.
Rendering is the step where Google processes the HTML, CSS, and JavaScript of your page to understand what a user would actually see. This matters because JavaScript-heavy sites can present content that Googlebot cannot immediately read. Google does render JavaScript, but it can be delayed, which is why server-side rendering is generally preferred for content you want indexed quickly.
Indexing is the process of storing a crawled and rendered page in Google’s index, making it eligible to appear in search results. A page being crawled does not guarantee it will be indexed. Google may choose not to index thin, duplicate, or low-quality content.
robots.txt is a text file at the root of your domain that tells crawlers which pages or directories they should not crawl. It does not prevent indexing. A page blocked in robots.txt can still be indexed if it has inbound links. This is one of the most commonly misunderstood distinctions in technical SEO.
Noindex is a directive placed in a page’s meta tags or HTTP headers that tells Google not to index the page. Unlike robots.txt, this directly controls indexation rather than crawling.
Canonical tag is an HTML element that signals to Google which version of a page should be treated as the primary version. It is used to manage duplicate content across multiple URLs. Google treats it as a hint, not an instruction, which is worth knowing before you rely on it too heavily.
XML sitemap is a file that lists the URLs you want Google to know about. It is a discovery tool, not a ranking tool. Submitting a sitemap does not improve your rankings. It helps ensure your important pages are known to Google.
Keyword and Search Intent Terminology
Keyword is the word or phrase a user types into a search engine. In modern SEO, the keyword is less important than the intent behind it. Two keywords can be nearly identical in wording but represent completely different user needs.
Search intent (also called user intent) refers to the underlying goal of a search query. Google categorises intent broadly into four types: informational (the user wants to learn something), navigational (the user wants to find a specific site), commercial (the user is researching before a purchase), and transactional (the user is ready to buy or act). Matching your content to the correct intent type is more important than keyword density.
Search volume is the estimated number of times a keyword is searched per month. It is a directional indicator, not a precise figure. Tools like Ahrefs, Semrush, and Google Keyword Planner all produce different estimates for the same keyword. Use volume to prioritise, not to forecast traffic with false precision.
Keyword difficulty is a metric used by SEO tools to estimate how hard it would be to rank for a given keyword. It is calculated differently by each tool and is a rough proxy, not a reliable prediction. A keyword with high difficulty may be achievable if you have strong topical authority. A keyword with low difficulty may still be out of reach if the top results are established brands with deep content.
Long-tail keywords are more specific, lower-volume search phrases. They typically convert better than head terms because the intent is clearer. “Running shoes” is a head term. “Best trail running shoes for wide feet under £100” is long-tail. The latter searcher knows what they want.
Head terms are short, high-volume, competitive keywords. They are often navigational or broad informational queries. Ranking for head terms is possible but rarely the most efficient use of SEO resource for most businesses.
Semantic search refers to Google’s ability to understand the meaning and context of a query, not just the literal words. This is why you do not need to stuff exact-match keywords into your content. Google understands synonyms, related concepts, and topic clusters. Writing for humans and writing for search engines has become the same thing, mostly.
Keyword mapping is the process of assigning target keywords to specific pages on your site. It prevents keyword cannibalism (multiple pages competing for the same term) and ensures each page has a clear search purpose.
On-Page SEO Terminology
Title tag is the HTML element that specifies the title of a page. It appears in the browser tab and, usually, as the clickable headline in search results. It is one of the clearest on-page signals to Google about what a page covers. Google rewrites title tags when it decides yours does not match the page content, which is worth knowing if you have ever wondered why your title looks different in results than what you wrote.
Meta description is the short summary that appears beneath the title in search results. It is not a ranking factor, but it influences click-through rate. Google also rewrites meta descriptions frequently, pulling text directly from the page when it judges that to be more relevant to the query.
H1, H2, H3 are HTML heading tags used to structure page content. The H1 is the main page title. H2s are section headers. H3s are subsections. Google uses heading structure to understand page hierarchy and topic coverage. Having one H1 per page is standard practice.
Alt text is the text description applied to images. It serves two purposes: accessibility for users who cannot see the image, and a signal to Google about what the image contains. Keyword-stuffed alt text is counterproductive. Descriptive, accurate alt text is the standard.
Internal linking is the practice of linking from one page on your site to another. It distributes authority across your site, helps Google understand your site structure, and guides users to related content. It is one of the most underused tactics in SEO, particularly on content-heavy sites.
Anchor text is the visible, clickable text in a hyperlink. Descriptive anchor text gives Google context about the linked page’s content. Over-optimised anchor text (using exact-match keywords in every link) is a historical spam signal that Google has become adept at identifying.
Featured snippet is the boxed result that appears above the standard organic listings, sometimes called “position zero.” It pulls a direct answer from a page, usually in response to a question-based query. Formatting content clearly, with direct answers to specific questions, increases the likelihood of earning a featured snippet.
Schema markup (also called structured data) is code added to a page to help Google understand its content more precisely. It powers rich results such as FAQ accordions, review stars, and recipe cards in search. Schema is useful for specific content types. It is not a ranking booster for general content, despite what some SEO vendors imply.
Authority, Links, and Off-Page Terminology
Backlink is a link from another website to your site. Backlinks remain one of the most significant ranking signals in Google’s algorithm. Not all backlinks are equal. A link from a relevant, authoritative site carries far more weight than one from a low-quality directory.
PageRank is Google’s original algorithm for measuring the importance of a page based on the quantity and quality of links pointing to it. Google no longer publicly shares PageRank scores, but the underlying concept still influences how authority flows through the web.
Domain Authority (DA) is a metric created by Moz to estimate how well a domain might rank in search results. It is not a Google metric and is not used in Google’s ranking systems. The same applies to Domain Rating (DR) from Ahrefs and Authority Score from Semrush. These are useful for benchmarking and comparison, but treating them as Google signals is a mistake I have seen made repeatedly, including by people who should know better.
Link equity (sometimes called “link juice,” though that term has aged poorly) refers to the value passed from one page to another through a hyperlink. Pages with more inbound links from authoritative sources pass more equity. Internal links also distribute equity across a site.
Nofollow is an attribute added to a link that signals to Google not to pass ranking credit to the linked page. Originally introduced to combat comment spam, nofollow links now carry some weight as a hint rather than a strict directive. Sponsored and UGC (user-generated content) attributes have since been introduced for more specific use cases.
Link building is the process of acquiring backlinks from other websites. It remains one of the most time-consuming and commercially significant activities in SEO. The best link building produces links that would earn coverage regardless of SEO, through genuinely useful content, original research, or tools. The worst link building produces links that exist purely to manipulate rankings and eventually attract a penalty.
Topical authority refers to how comprehensively a site covers a subject area. Google increasingly rewards sites that demonstrate deep expertise on a topic rather than sites that produce isolated high-ranking pages. Building topical authority requires a coherent content strategy, not just individual articles optimised for individual keywords. Moz has written thoughtfully about how content strategy connects to authority in ways that go beyond simple link counting.
E-E-A-T stands for Experience, Expertise, Authoritativeness, and Trustworthiness. It is a framework from Google’s Search Quality Rater Guidelines, used by human quality raters to assess search results. It is not a direct ranking algorithm, but it describes the qualities Google’s systems are trying to reward. The first E (Experience) was added to reflect the value of first-hand, real-world knowledge, which is why author credentials and demonstrated experience matter more than they once did.
I judged the Effie Awards for several years, and one thing that experience sharpened was my ability to spot the difference between claimed expertise and demonstrated expertise. The same distinction applies in SEO. A site that says it is authoritative is not. A site that consistently produces content that others reference, link to, and return to is.
Technical SEO Terminology
Core Web Vitals are a set of page experience metrics defined by Google: Largest Contentful Paint (LCP), which measures loading performance; Interaction to Next Paint (INP), which measures responsiveness; and Cumulative Layout Shift (CLS), which measures visual stability. They are ranking signals, but relatively modest ones. A site with excellent Core Web Vitals scores and thin content will not outrank a site with strong content and average scores.
Page speed is the time it takes for a page to load. It affects both user experience and, to a degree, rankings. The relationship is not linear. Improving load time from 8 seconds to 4 seconds may have a meaningful impact. Improving from 2.1 seconds to 1.9 seconds probably will not. Chasing marginal speed gains is a form of complexity that delivers diminishing returns.
Mobile-first indexing means Google primarily uses the mobile version of a page for indexing and ranking. This has been the default for new sites for several years. If your mobile experience is significantly worse than your desktop experience, that is the version Google is evaluating.
HTTPS is the secure version of HTTP. Google confirmed HTTPS as a ranking signal years ago. If your site is still on HTTP, fix it. This is not a competitive advantage. It is table stakes.
Redirect is a method of sending users and search engines from one URL to another. A 301 redirect is permanent and passes link equity. A 302 redirect is temporary and, in theory, does not pass equity. Redirect chains (where a URL redirects to another URL that redirects again) should be cleaned up, as they slow crawling and dilute equity.
404 error occurs when a page cannot be found. It is not inherently harmful in small numbers. A site with hundreds of broken internal links pointing to 404 pages has a structural problem worth addressing. A handful of 404s on pages with no inbound links is not worth losing sleep over.
Log file analysis is the process of examining server logs to see exactly how Googlebot is crawling your site, which pages it visits, how frequently, and which it ignores. It is one of the most underused diagnostic tools in technical SEO and one of the most revealing. Most crawl issues that agencies claim to have identified through tools become much clearer through log data.
Content and Architecture Terminology
Content cluster (also called a topic cluster) is a content architecture model where a central “pillar” page covers a broad topic and links to supporting “cluster” pages that cover subtopics in depth. The cluster pages link back to the pillar. This structure signals topical authority and helps Google understand the relationship between your pages.
Pillar page is the central, comprehensive page in a content cluster. It covers a topic broadly and links out to more detailed cluster content. The article you are reading is part of a cluster. The Complete SEO Strategy hub at themarketingjuice.com/seo-strategy is the pillar that this and related articles support.
Thin content refers to pages with little substantive value: pages that are too short, duplicate other content, or exist primarily for SEO rather than user value. Google’s Panda algorithm update targeted thin content at scale. The concept remains relevant. Content that exists to rank rather than to answer a real question is increasingly unlikely to rank at all.
Duplicate content occurs when the same or substantially similar content appears at multiple URLs. It can happen across domains (plagiarism or scraped content) or within a single site (parameter-based URLs, printer-friendly pages, session IDs). Google handles most duplicate content algorithmically, selecting one version to index. Canonical tags help guide that selection.
Content decay is the gradual decline in traffic and rankings that affects older content as it becomes outdated or is overtaken by fresher, more comprehensive alternatives. Regularly auditing and updating existing content is often more efficient than producing new content, particularly for established sites.
Search landscape refers to the mix of result types Google shows for a given query: organic results, featured snippets, local packs, image results, video results, People Also Ask boxes, and so on. Understanding the search landscape for your target keywords tells you what type of content Google judges to be most relevant, and therefore what you should produce.
Measurement and Reporting Terminology
Organic traffic is traffic that arrives at your site through unpaid search results. It is the primary output metric for SEO. It is also, on its own, an incomplete measure of SEO success. Organic traffic that does not convert, does not lead to enquiries, and does not contribute to revenue is vanity. I have seen agencies report organic traffic growth of 40% while the client’s pipeline from search declined. The number looked good. The business outcome did not.
Click-through rate (CTR) is the percentage of users who see your result in search and click on it. Google Search Console provides CTR data at the page and query level. Low CTR on high-impression queries often indicates a title tag or meta description problem rather than a ranking problem.
Impressions in the context of Google Search Console refers to the number of times your page appeared in search results, regardless of whether it was clicked. A page with high impressions and low CTR is visible but not compelling. A page with low impressions may have a ranking problem or a keyword targeting problem.
Google Search Console (GSC) is Google’s free tool for monitoring how your site appears in search results. It shows performance data (clicks, impressions, CTR, average position), index coverage, Core Web Vitals, and manual actions. It is the most reliable source of search data available because it comes directly from Google. Third-party rank trackers and traffic estimators are approximations. GSC is the source.
Rank tracking is the process of monitoring where your pages appear in search results for target keywords over time. Position tracking tools provide useful directional data but should not be treated as precise. Rankings vary by location, device, personalisation, and the time of day. The trend matters more than any individual data point.
Share of voice in SEO refers to the proportion of available organic clicks your site captures for a given set of keywords relative to competitors. It is a more commercially meaningful metric than raw rankings because it contextualises your visibility against the competitive landscape.
Having managed hundreds of millions in media spend across more than 30 industries, one thing I feel strongly about is that SEO reporting needs to connect to business outcomes, not just search metrics. Organic sessions are a leading indicator. Revenue, pipeline, and customer acquisition are the outcomes. Forrester has written about the gap between marketing capability and commercial accountability, and it is a gap I recognise from years of client work. The vocabulary of SEO measurement should include business language, not just search language.
Terms That Are Frequently Misused or Overstated
A few terms deserve particular scrutiny because they are regularly used in ways that mislead clients or inflate the perceived complexity of SEO work.
Domain Authority: As noted above, this is a Moz metric, not a Google ranking factor. Agencies that report DA as a measure of SEO progress are either confused or are using a proxy because genuine progress is harder to demonstrate. Use it for competitive benchmarking. Do not treat it as a performance KPI.
Algorithm update: Google runs thousands of updates per year, most of them minor. Major named updates (Panda, Penguin, Helpful Content, Core Updates) have genuine significance. Attributing every traffic fluctuation to an “algorithm update” is a common agency deflection. Sometimes traffic drops because a competitor published better content. Sometimes it is seasonal. Sometimes it is a tracking issue. The algorithm is not always the culprit.
Keyword ranking: Ranking first for a keyword does not guarantee traffic. If the SERP for that keyword is dominated by ads, featured snippets, and People Also Ask boxes, the organic result in position one may receive very little of the available clicks. Ranking is an intermediate metric. Clicks and conversions are the outcomes.
Bounce rate: A high bounce rate is not inherently bad. A user who reads an entire article and leaves satisfied has technically “bounced.” Engagement rate (introduced in GA4 as a replacement for bounce rate) is a more nuanced measure, but it too requires context. What matters is whether users are getting value from your content, not whether they click to a second page.
Moz has explored how community signals and brand behaviour connect to search performance in ways that go beyond standard link metrics, and their Whiteboard Friday on community and SEO is worth watching for anyone who wants to think beyond keyword optimisation. The broader point is that SEO has never been purely technical. The signals Google values most, authority, relevance, trust, are earned through genuine quality, not manufactured through tactics.
About the Author
Keith Lacy is a marketing strategist and former agency CEO with 20+ years of experience across agency leadership, performance marketing, and commercial strategy. He writes The Marketing Juice to cut through the noise and share what works.
Frequently Asked Questions
If you found this useful and want to see how these concepts fit into a broader framework, the Complete SEO Strategy hub brings together the mechanics of search in one place, from technical foundations to content architecture to measurement.
