What You'll Learn
- How search engines discover and crawl content
- How indexing works and what can prevent it
- How ranking algorithms evaluate pages
- What signals matter most in 2025
- How AI is changing search and what to do about it
Step 1: Crawling
Search engines use automated programs called crawlers (or spiders) to discover content on the web. Googlebot, Google's primary crawler, follows links from page to page, starting from a list of known URLs and expanding outward through every link it encounters.
Crawl budget — the number of pages Google crawls on your site per day — is finite. Large sites with poor architecture waste crawl budget on low-value pages, leaving important content undiscovered. Managing crawl efficiency is a core technical SEO concern.
What Can Block Crawling
- Robots.txt rules that accidentally block important paths
- Noindex meta tags on pages you want indexed
- Login walls or JavaScript-heavy content that Googlebot can't render
- Crawl traps (infinite loops from faceted navigation, calendars, etc.)
Step 2: Indexing
After crawling, Google processes and stores page content in its index — a massive database of all the pages it has evaluated. Only indexed pages can appear in search results. Processing involves understanding the page's topic, entities, quality, and how it relates to other pages.
Factors that affect indexing include content quality (thin or duplicate content may not be indexed), page experience signals, and whether the page has been crawled recently. Google Search Console's URL Inspection tool is the best way to check indexing status for specific pages.
Step 3: Ranking
When a user enters a query, Google's ranking algorithm evaluates all relevant indexed pages and sorts them by estimated usefulness for that specific query. This process happens in milliseconds and considers hundreds of signals.
Core Ranking Signals
- Relevance: how well the page's content matches the query's intent
- Authority: the quality and quantity of links pointing to the page
- Experience: page speed, mobile-friendliness, Core Web Vitals
- Content quality: depth, accuracy, freshness, and E-E-A-T signals
- User signals: click-through rate, dwell time, pogo-sticking
How AI Is Changing Search
Google's Search Generative Experience (SGE), Perplexity, and ChatGPT Search are increasingly surfacing AI-generated answers above traditional blue links. These systems pull from pages they consider authoritative, well-structured, and clearly written.
Optimising for AI-powered search means structuring content with clear headings, concise answer blocks, FAQ sections, and schema markup. Pages that answer questions directly and demonstrate expertise are more likely to be cited in AI-generated summaries.
Frequently Asked Questions
How long does it take for a new page to rank?
New pages on established domains can rank within days or weeks for low-competition queries. New domains and competitive terms typically take 3–6 months to see meaningful rankings. Consistency of publishing and quality of backlinks are the primary accelerators.
Why does Google rank pages differently in different countries?
Google personalises results based on location, language, search history, and device. A page that ranks #1 in the UK may rank #5 in the US due to different competitor strength, localisation, and relevance signals.
Can I tell Google which pages to crawl?
Yes, through your robots.txt file (for general crawl control), XML sitemap (to highlight important pages), and canonical tags (to tell Google the preferred version of duplicate pages). Google Search Console also allows you to request indexing of specific URLs.

