SEO Glossary: 60 Terms That Matter
An SEO glossary is a reference guide to the terminology used in search engine optimisation, covering everything from technical infrastructure to content signals and link authority. This one is built differently: instead of listing every term alphabetically, it organises concepts by how they connect to each other and to commercial outcomes.
Most SEO glossaries read like a dictionary someone compiled to fill a content calendar. This one is written for marketers who need to understand what these terms mean in practice, how they relate to ranking decisions, and where the industry has a habit of overcomplicating things that are fundamentally straightforward.
Key Takeaways
- SEO terminology is only useful when you understand how terms relate to each other, not just what they mean in isolation.
- Most ranking factors operate as thresholds, not linear scales. Being good enough on technical signals frees you to compete on content quality.
- E-E-A-T is not a checklist. It is a framework Google uses to assess whether a page is a credible source for a given query type.
- Link metrics like Domain Authority are third-party proxies, not Google measurements. Treat them as directional indicators, not ground truth.
- Search intent is the single most important variable in keyword strategy. A page optimised for the wrong intent will not rank, regardless of its technical quality.
In This Article
- Why SEO Terminology Gets Misused
- Crawling, Indexing, and Rendering: The Foundation
- On-Page Signals: What Google Reads
- Search Intent: The Term That Explains Most Ranking Decisions
- Authority, Links, and the Metrics That Surround Them
- E-E-A-T: What It Is and What It Is Not
- Technical SEO: The Infrastructure Terms
- Keyword Research Terms: The Vocabulary of Demand
- Content and Architecture Terms
- Measurement Terms: What the Numbers Actually Mean
- Algorithm Terms: What Google Is Actually Doing
- Local SEO Terms
- A Note on Tools and Metrics
Why SEO Terminology Gets Misused
When I was running an agency, one of the most reliable signals that a client relationship was going to be difficult was the misuse of SEO terminology in briefings. Not because precision in language is a virtue in itself, but because imprecise language usually signals imprecise thinking. Someone who conflates crawling and indexing probably also conflates traffic and revenue. These are different problems with different solutions.
The SEO industry has a specific version of this problem. It generates terminology at a rate that outpaces its ability to define terms clearly. New concepts get named before they are properly understood, old terms get redefined as the algorithm changes, and third-party tools introduce their own metrics that get treated as if Google invented them. The result is a vocabulary that sounds technical but often obscures more than it reveals.
This glossary is part of a broader resource on building a complete SEO strategy. If you want the full framework, the SEO strategy hub covers how these terms connect to real decisions: keyword selection, content architecture, link building, and measurement.
Crawling, Indexing, and Rendering: The Foundation
Crawling is the process by which search engine bots (Googlebot being the most commercially significant) discover and retrieve pages from the web. A bot follows links, reads page content, and passes that content back to Google’s servers for processing. Crawling is not the same as indexing.
Indexing is the process of storing and organising crawled content so it can be retrieved in response to a search query. A page can be crawled and not indexed. Google makes that decision based on quality signals, duplicate content assessments, and crawl budget allocation.
Rendering is the step between crawling and indexing where Google executes the JavaScript on a page to understand what a user would actually see. This matters because a significant proportion of modern websites deliver content via JavaScript frameworks. If Google cannot render a page correctly, it may index an empty shell rather than the actual content.
Crawl budget refers to the number of pages Googlebot will crawl on a given site within a given period. For large sites with thousands of URLs, crawl budget management becomes a real concern. For most sites under a few hundred pages, it is not worth worrying about.
Robots.txt is a text file that instructs crawlers which parts of a site they are allowed to access. It does not prevent indexing. A page blocked by robots.txt can still appear in search results if other sites link to it. Many marketers conflate robots.txt with noindex directives, which are different tools for different purposes.
Noindex is a directive placed in a page’s HTML that tells search engines not to include that page in their index. Unlike robots.txt, noindex operates at the page level and is processed after the page has been crawled. Use it for thin pages, duplicate content, and internal search results you do not want consuming index space.
Canonical tag is a signal placed in a page’s HTML that tells Google which version of a URL should be treated as the definitive one. Canonicals are advisory, not directives. Google may choose to ignore them if it disagrees with your assessment of which page is the primary version.
On-Page Signals: What Google Reads
Title tag is the HTML element that specifies the title of a page. It appears in search results as the clickable headline of a listing. It is one of the clearest signals of page topic available to Google and one of the few on-page elements where precise keyword placement still has a measurable effect on rankings. Google rewrites title tags when it judges them to be misleading or poorly matched to the page content.
Meta description is the HTML element that provides a summary of a page’s content. It does not directly influence rankings. It does influence click-through rate, which means it has an indirect commercial effect. Write meta descriptions for users, not for algorithms.
H1 tag is the primary heading on a page. There should be one per page. It signals the main topic of the content and should align with the title tag without being identical to it. H2 and H3 tags structure the content hierarchy beneath it.
Keyword density is a metric that measures how frequently a target keyword appears in a page relative to total word count. It was a meaningful signal in early search algorithms. It is not a meaningful optimisation target today. Pages that are written naturally for a specific topic will use relevant terms at appropriate frequencies without counting them.
Semantic relevance describes the degree to which a page’s vocabulary and topic coverage align with a query beyond exact keyword matching. Modern search algorithms assess meaning, not just word frequency. A page about mortgage refinancing that never uses the phrase “mortgage refinancing” can still rank for it if the content is semantically coherent and comprehensive.
Schema markup is structured data added to a page’s HTML that helps search engines understand the content type and its properties. Schema does not directly improve rankings, but it enables rich results in search listings, including FAQ dropdowns, review stars, and event information, which can improve click-through rates.
Core Web Vitals are a set of user experience metrics Google uses as ranking signals. They measure loading performance (Largest Contentful Paint), interactivity (Interaction to Next Paint), and visual stability (Cumulative Layout Shift). They matter more as tiebreakers between otherwise comparable pages than as primary ranking drivers. Sites that perform poorly on Core Web Vitals often have broader UX problems worth fixing regardless of their SEO impact.
Search Intent: The Term That Explains Most Ranking Decisions
Search intent is the underlying goal a user has when they type a query into a search engine. Google classifies intent into four broad categories: informational (the user wants to learn something), navigational (the user wants to find a specific site or page), commercial (the user is researching a purchase decision), and transactional (the user is ready to take an action).
Intent is the single most important variable in keyword strategy, and it is the one most often ignored in favour of search volume. I have seen clients invest months building content around high-volume keywords only to discover that the search results for those terms are dominated by a content type they cannot or will not produce. Understanding intent before you commit to a keyword is not optional work.
SERP features are the non-standard elements that appear in search results beyond the ten blue links. They include featured snippets, People Also Ask boxes, image carousels, video results, local map packs, and knowledge panels. SERP features are intent-driven. Google shows them when it judges that a particular format better serves the user’s goal than a standard listing. Understanding which features appear for a target keyword tells you a great deal about how Google interprets the intent behind that query.
Featured snippet is the boxed result that appears at the top of some search listings, displaying a direct answer extracted from a web page. Winning a featured snippet does not require ranking first. Pages ranking in positions two through five frequently earn featured snippets. what matters is structuring content so that it directly and concisely answers a specific question.
Zero-click search refers to a search session where the user gets their answer directly from the SERP without clicking through to any website. Featured snippets, knowledge panels, and calculator tools all drive zero-click behaviour. For informational queries, zero-click rates are high. For commercial and transactional queries, they are much lower. This distinction matters for how you prioritise content investment.
Authority, Links, and the Metrics That Surround Them
Backlink is a link from one website to another. Backlinks remain one of Google’s most important ranking signals because they function as editorial endorsements. A link from a credible, relevant site signals that the linked page has value worth referencing. Not all backlinks carry equal weight.
PageRank is Google’s original algorithm for measuring the authority of a page based on the quantity and quality of links pointing to it. PageRank is still a core part of how Google assesses authority, though the public toolbar version was discontinued in 2016. The underlying concept, that links transfer authority and that not all links are equal, remains valid.
Domain Authority (DA) is a metric developed by Moz that predicts how likely a domain is to rank in search results, expressed on a scale of one to one hundred. Domain Authority is a third-party metric. Google does not use it. It is a useful directional indicator for comparing the relative strength of competing domains, but it is not a target to optimise for directly. Treating DA as if it were a Google metric is one of the more persistent category errors in the industry.
Domain Rating (DR) is Ahrefs’ equivalent metric, measuring the strength of a domain’s backlink profile relative to others in their index. The same caveat applies. DR is a proxy, not a ground truth.
Anchor text is the visible, clickable text of a hyperlink. Anchor text signals relevance to search engines. A link with anchor text “mortgage refinancing guide” carries a different signal than the same link with anchor text “click here.” Over-optimised anchor text profiles, where a high proportion of inbound links use exact-match keyword anchor text, are a signal of manipulative link building and can trigger algorithmic penalties.
Nofollow is a link attribute that instructs search engines not to pass authority through a link. Google introduced nofollow in 2005 to combat comment spam. In 2019, Google reclassified nofollow as a hint rather than a directive, meaning it may choose to count nofollow links in some circumstances. Sponsored and UGC attributes were introduced at the same time to provide more specific signals for paid and user-generated content.
Link velocity refers to the rate at which a site acquires new backlinks over time. A sudden, unnatural spike in link acquisition can be a signal of manipulative behaviour and may attract algorithmic scrutiny. Steady, gradual link growth from credible sources is the pattern that correlates with sustainable ranking improvement.
For a broader view of how link signals fit into a complete SEO approach, the Complete SEO Strategy hub covers the full picture, from technical foundations through to content and authority building.
E-E-A-T: What It Is and What It Is Not
E-E-A-T stands for Experience, Expertise, Authoritativeness, and Trustworthiness. It is a framework from Google’s Search Quality Rater Guidelines, a document used to train human evaluators who assess search result quality. E-E-A-T is not a direct ranking factor in the sense that there is no E-E-A-T score in Google’s algorithm. It is a set of qualities that Google’s systems attempt to assess through proxies: author credentials, site reputation, link signals, content accuracy, and user engagement patterns.
I judged the Effie Awards for a period and spent a lot of time evaluating campaigns against criteria that were clear in principle but required significant interpretation in practice. E-E-A-T works similarly. The criteria are coherent. The application requires judgment. Anyone who tells you they have a precise formula for optimising E-E-A-T is selling something.
Experience was added to the framework in 2022, distinguishing between expertise (knowing about a topic) and first-hand experience (having done the thing). A review written by someone who has used a product carries a different signal than one written by someone who has researched it. This distinction matters most for YMYL content.
YMYL stands for Your Money or Your Life. It describes content categories where inaccurate information could cause real harm: medical, financial, legal, and safety-related topics. Google applies higher quality thresholds to YMYL content. The E-E-A-T bar is higher in these categories, which is why thin, anonymous content rarely ranks for competitive health or finance queries.
Technical SEO: The Infrastructure Terms
XML sitemap is a file that lists the URLs on a website and provides metadata about each one, including when it was last updated and how frequently it changes. Sitemaps help search engines discover content more efficiently, particularly on large sites with complex link structures. Submitting a sitemap to Google Search Console does not guarantee indexing.
301 redirect is a permanent redirect from one URL to another. It passes the majority of link authority from the old URL to the new one. Use 301s when you permanently move content, consolidate duplicate pages, or change URL structures. A 302 redirect signals a temporary move and does not pass authority in the same way.
404 error occurs when a server cannot find the requested page. 404s are a normal part of the web and do not inherently harm rankings. What matters is whether the 404 pages are ones that previously had authority and inbound links. Those should be redirected rather than left to return errors.
HTTPS is the secure version of the HTTP protocol, using SSL/TLS encryption. Google confirmed HTTPS as a ranking signal in 2014. In practice, it functions more as a baseline requirement than a competitive advantage. Sites still running on HTTP in 2025 have bigger problems than their search rankings.
Page speed refers to how quickly a page loads for a user. It is a ranking factor, but its weight varies by query type and competitive context. The practical priority is ensuring pages load quickly enough not to frustrate users, rather than chasing marginal speed improvements for their own sake.
Mobile-first indexing means Google primarily uses the mobile version of a page for indexing and ranking. Google completed the transition to mobile-first indexing for all sites in 2024. If your mobile and desktop versions serve significantly different content, Google will index the mobile version. This has implications for sites that strip content from their mobile experience.
Structured data is code added to a page to provide explicit information about its content to search engines. Schema markup is the most common implementation. Structured data supports rich results but does not directly improve standard rankings.
Keyword Research Terms: The Vocabulary of Demand
Search volume is the average number of times a keyword is searched per month in a given market. Search volume is an estimate, not a precise measurement. Different tools produce different estimates for the same keyword, and the actual volume for any given month will vary. Treat search volume as a relative indicator of demand, not an absolute number.
Keyword difficulty is a metric used by SEO tools to estimate how hard it would be to rank for a given keyword. Like Domain Authority, it is a third-party approximation. The inputs vary by tool, but most combine the authority of currently ranking pages with the number of backlinks those pages have. Use keyword difficulty to shortlist, not to make final decisions.
Long-tail keyword refers to a longer, more specific search phrase that typically has lower search volume but higher conversion intent. The term comes from the statistical distribution of search demand, where a small number of head terms account for a large proportion of searches and a very large number of specific queries each account for a small proportion. Long-tail keywords are often less competitive and more commercially precise than head terms.
Head term is a short, high-volume keyword, typically one to two words. Head terms are usually highly competitive and often have mixed or ambiguous intent. Ranking for head terms requires significant authority and is not a realistic near-term goal for most sites.
Keyword clustering is the practice of grouping related keywords that share the same search intent and should be targeted by a single page, rather than creating separate pages for each variation. Keyword clustering reduces the risk of keyword cannibalisation and helps ensure that content is comprehensive enough to rank for a topic rather than a single phrase.
Keyword cannibalisation occurs when multiple pages on a site compete for the same keyword, confusing search engines about which page to rank and diluting the authority of both. It is a common problem on sites that have grown organically without a content architecture strategy. The fix is usually consolidation or canonical signals, not deletion.
Content and Architecture Terms
Content hub is a structured collection of content organised around a central topic, with a pillar page covering the topic broadly and cluster pages covering subtopics in depth. Hubs create internal linking structures that signal topical authority to search engines and improve user navigation. They are also a sensible way to organise editorial planning.
Pillar page is the central page of a content hub, covering a broad topic comprehensively and linking out to more detailed cluster content. Pillar pages are typically longer than standard articles and are designed to rank for competitive head terms while supporting the ranking of cluster pages through internal linking.
Internal linking is the practice of linking between pages on the same website. Internal links distribute authority across a site, help search engines understand content relationships, and guide users to relevant content. Internal linking is consistently undervalued relative to external link building, despite being entirely within the site owner’s control.
Thin content refers to pages that offer little value to users: very short pages, pages that duplicate content from elsewhere, or pages that exist primarily to target a keyword rather than to serve a user need. Google’s Panda algorithm update, first released in 2011, targeted thin content at scale. Sites with large proportions of thin content can see broad ranking suppression rather than page-level penalties.
Content decay is the gradual decline in traffic and rankings that affects content as it ages, becomes outdated, or is overtaken by newer, more comprehensive pages from competitors. Content decay is predictable and manageable. Regular content audits, with a focus on updating or consolidating declining pages, are more efficient than continuously producing new content.
Measurement Terms: What the Numbers Actually Mean
Organic traffic refers to visits to a website that originate from unpaid search results. It is one of the primary metrics used to assess SEO performance, but it is not the only one that matters commercially. Organic traffic that does not convert to enquiries, leads, or sales is an activity metric, not a business outcome.
Early in my career running agency P&Ls, I learned to be suspicious of metrics that looked good in isolation. Organic traffic is one of them. A site can grow organic traffic significantly by ranking for informational queries that have no commercial value to the business. The question is always whether the traffic is the right traffic.
Click-through rate (CTR) is the percentage of users who see a search listing and click on it. CTR varies significantly by position, query type, and SERP layout. Position one listings typically achieve higher CTRs than position two, but the presence of SERP features can compress the difference or suppress organic CTR entirely for some query types.
Impressions in the context of Google Search Console refers to the number of times a page URL appeared in search results, regardless of whether it was clicked. Impressions data is useful for identifying keywords where a site has visibility but low CTR, suggesting title tag or meta description improvements may be warranted.
Average position is the mean ranking position of a page across all queries for which it appears. It is a useful directional metric but can be misleading when averaged across very different query types. A page ranking first for low-volume queries and tenth for high-volume queries may show an average position of five, which tells you very little about its actual commercial performance.
Google Search Console is Google’s free tool for monitoring a site’s presence in search results. It provides data on impressions, clicks, average position, index coverage, Core Web Vitals, and manual actions. It is the closest thing to a primary source that SEO practitioners have for understanding how Google sees a site. Every site should have it configured.
Google Analytics provides traffic, engagement, and conversion data. It complements Search Console but measures different things. Search Console shows what happens in Google’s systems. Google Analytics shows what happens after a user arrives on the site. Both are perspectives on reality, not reality itself. Sampling, attribution models, and cookie consent all affect the accuracy of what you see.
The Search Engine Journal covers ongoing developments in how Google’s systems evolve, which is worth monitoring alongside your own measurement data.
Algorithm Terms: What Google Is Actually Doing
Core algorithm update is a broad change to Google’s ranking systems that can significantly affect rankings across many queries and industries simultaneously. Google releases several core updates per year. They do not target specific sites or tactics. They recalibrate how Google assesses quality across the board. Sites that see ranking drops after a core update have usually not been penalised. They have been reassessed.
Helpful Content System is a Google system introduced in 2022 and subsequently integrated into the core algorithm that assesses whether content is created primarily to serve users or primarily to rank in search engines. It applies a site-wide signal, meaning that large amounts of unhelpful content on a site can suppress the performance of otherwise good pages. The practical implication is that content auditing matters more than most SEO practitioners have historically treated it.
Manual action is a penalty applied to a site by a human reviewer at Google, rather than by an algorithm. Manual actions are applied for clear violations of Google’s spam policies: paid link schemes, cloaking, thin content at scale. They appear in Google Search Console and require a reconsideration request to resolve. Algorithmic demotions are more common than manual actions and do not appear as notifications.
Penguin was a Google algorithm update first released in 2012, targeting manipulative link building practices. Penguin was integrated into the core algorithm in 2016 and now operates in real time. Sites with toxic backlink profiles are assessed continuously rather than at update intervals.
Panda was a Google algorithm update first released in 2011, targeting low-quality and thin content. Like Penguin, it was integrated into the core algorithm. Its legacy is the principle that content quality is assessed at the site level, not just the page level.
Understanding algorithm history matters because it explains why certain SEO practices that worked a decade ago are now either ineffective or actively harmful. The industry has a long memory for tactics and a short memory for why those tactics stopped working.
Local SEO Terms
Google Business Profile (formerly Google My Business) is the free listing tool that controls how a business appears in Google Maps and local search results. For businesses with a physical location or a defined service area, Google Business Profile is often the highest-return SEO activity available. It is also one of the most neglected.
Local pack is the map-based result block that appears for location-specific queries, typically showing three business listings. Ranking in the local pack is driven by proximity, relevance, and prominence. Prominence is influenced by the same signals that affect organic rankings: links, reviews, and content quality.
NAP consistency stands for Name, Address, Phone number. Consistent NAP information across directories, social profiles, and the business website is a foundational local SEO signal. Inconsistencies confuse search engines about which information is authoritative and can suppress local rankings.
Citation is any online mention of a business’s name, address, and phone number. Citations from local directories, industry associations, and chamber of commerce listings contribute to local authority signals. The quality of citations matters more than the quantity.
A Note on Tools and Metrics
A significant proportion of the terms in this glossary are tool-generated metrics rather than Google measurements. Domain Authority, Domain Rating, keyword difficulty scores, and spam scores are all third-party approximations of signals that Google measures differently and does not make public. They are useful as relative indicators. They are not ground truth.
When I was growing an agency from twenty to a hundred people, one of the discipline problems I had to manage was the tendency to treat tool outputs as facts. A client’s Domain Authority drops three points and suddenly there is a crisis meeting. The Domain Authority had not changed. Moz’s model had. These are different things.
The same principle applies to ranking tools, traffic estimators, and backlink databases. They are perspectives on reality, built from crawled samples of the web. They are useful. They are not the same as Google’s actual data, which you can only access through Search Console and Analytics, imperfect as those are.
The Forrester perspective on digital marketing measurement is worth reading for a broader view of how measurement frameworks shape the decisions marketers make, often in ways that are not commercially rational.
For the full strategic framework that connects these terms to real decisions, the Complete SEO Strategy hub is where to go next. It covers how to build a keyword strategy, structure content for authority, and measure what actually matters.
About the Author
Keith Lacy is a marketing strategist and former agency CEO with 20+ years of experience across agency leadership, performance marketing, and commercial strategy. He writes The Marketing Juice to cut through the noise and share what works.
