The Technical SEO Audit Checklist: What Every Agency Should Check Before You Sign
You've hired an SEO agency, or you're about to. They promise rankings, traffic, and revenue. But before you hand over your site credentials, there's a fundamental truth you need to understand: if your website's technical foundation is broken, no amount of content or backlinks will fix it. Technical SEO is the plumbing of your online presence—invisible when it works, catastrophic when it fails.
Most agencies will pitch you on keyword research and content strategy because those are easy to sell. But the real work—the work that separates professionals from amateurs—happens in the crawl logs, the server response codes, and the Core Web Vitals report. This checklist is your roadmap for evaluating whether an agency actually knows what they're doing, or if they're just going to break things faster than they fix them.
What Is Technical SEO, Really?
Technical SEO refers to the optimization of your website's infrastructure to help search engines crawl, index, and render your pages efficiently. It's not about writing better meta descriptions or adding keywords to headings—that's on-page optimization. Technical SEO is about ensuring that Googlebot can actually find your content, understand its structure, and deliver it to users without errors or delays.
Think of it this way: if your site is a library, technical SEO is the building's foundation, the lighting, the signage, and the catalog system. On-page optimization is the books themselves. Link building is the foot traffic. You can have the best books in the world, but if the library is dark, has no signs, and the catalog is missing every other card, nobody is going to find anything.
The Core Components
| Component | What It Does | What Happens If It's Broken |
|---|---|---|
| Crawl budget | Determines how many pages Googlebot will scan per visit | Important pages never get indexed; thin pages waste resources |
| XML sitemap | Tells search engines which pages exist and how often they change | New content may take weeks or months to appear in search results |
| robots.txt | Instructs crawlers which parts of the site to ignore | Sensitive content may be indexed; critical pages may be blocked |
| Canonical tags | Prevents duplicate content by specifying the preferred URL | Search engines see multiple versions of the same page, diluting ranking signals |
| Core Web Vitals | Measures real-user experience metrics (LCP, CLS, INP) | Poor scores can affect rankings, especially on mobile |
| Site structure | Organizes URLs and internal linking for logical hierarchy | Users and crawlers get lost; authority doesn't flow to important pages |
The Crawl Budget Trap
Here's a scenario that plays out more often than agencies will admit: a client with a large e-commerce site hires an agency. The agency runs a crawl, finds many issues, and immediately starts "optimizing" everything. Months later, organic traffic declines. What happened?
They wasted the crawl budget.
Crawl budget is the number of URLs Googlebot will request from your server during a given crawl session. It's not infinite, and it's not equal for every site. Smaller sites rarely worry about it, but for large sites—think thousands or millions of pages—crawl budget is a finite resource that must be managed carefully.
When an agency starts changing redirects, adding parameters, or restructuring URLs without understanding your crawl budget, they can accidentally cause Googlebot to waste its limited requests on low-value pages (filter pages, parameterized URLs, old product variants) while high-value pages (category pages, best-selling products, cornerstone content) go uncrawled for weeks.
What to ask your agency: "Show me your crawl budget analysis. How many pages does Googlebot crawl per visit? Which pages are being crawled most frequently? Are we wasting budget on thin or duplicate content?"
Core Web Vitals: The User Experience Audit
Google's Core Web Vitals are a ranking factor. The three metrics—Largest Contentful Paint (LCP), Cumulative Layout Shift (CLS), and Interaction to Next Paint (INP)—measure how real users experience your site.
Here's the uncomfortable truth: most agencies don't know how to fix Core Web Vitals issues. They'll run a Lighthouse report, see a red score, and tell you to "compress images" or "use a CDN." Those are surface-level fixes. The real problems are often deeper:
- LCP issues caused by slow server response times, render-blocking JavaScript, or lazy-loading the hero image
- CLS issues caused by dynamically injected ads, web fonts that shift layout, or images without explicit dimensions
- INP issues caused by heavy JavaScript frameworks, unoptimized third-party scripts, or poor event handler performance

Red flag: If an agency says "we'll fix Core Web Vitals in a week" without first auditing your third-party scripts, server configuration, and front-end framework, they're oversimplifying the problem.
The Duplicate Content Minefield
Duplicate content is one of those problems that sounds simple but rarely is. You might think, "I don't have duplicate content—I wrote everything myself." But search engines don't see it that way.
Consider these scenarios:
- Your site is accessible at both `https://example.com` and `https://www.example.com` without a redirect
- Product pages are available with and without tracking parameters (`?color=red`, `?size=large`)
- Blog posts have tag pages that show the same content as the main category page
- Session IDs are added to URLs for logged-in users
- Print-friendly versions of pages exist without a canonical tag
A thorough technical SEO audit will identify every instance of duplicate content and recommend a canonicalization strategy. The agency should explain whether they're using 301 redirects, rel=canonical tags, or parameter handling in Google Search Console—and why each decision makes sense for your specific site structure.
On-Page Optimization: Beyond the Meta Description
On-page optimization is where most agencies start and, unfortunately, where many stop. It's the process of optimizing individual pages to rank for specific keywords and satisfy user intent. But there's a right way and a wrong way.
The wrong way: stuffing keywords into title tags, headers, and body copy regardless of whether they make sense for the reader. This approach might work briefly, but Google's algorithms are sophisticated enough to detect keyword stuffing and may penalize you for it.
The right way: mapping each page to a specific search intent, then optimizing for that intent. This is where keyword research and intent mapping come together.
Keyword Research and Intent Mapping
Keyword research isn't just about finding high-volume terms. It's about understanding what users actually want when they search for those terms. The four types of search intent are:
- Informational: The user wants to learn something ("how to fix a leaky faucet")
- Navigational: The user wants to find a specific site ("Facebook login")
- Commercial: The user wants to research before buying ("best SEO tools 2025")
- Transactional: The user wants to buy ("buy organic coffee beans online")
What to look for: The agency should provide a keyword-to-intent mapping document that shows which keywords they're targeting, which intent each keyword represents, and which page on your site is the best match. If they can't explain why a particular keyword belongs on a particular page, they're guessing.
Content Strategy: The Long Game
Content strategy is where technical SEO meets editorial planning. It's not about writing blog posts randomly; it's about creating a structured plan that supports your business goals, targets the right keywords, and builds topical authority over time.

A good content strategy includes:
- Topic clusters: A pillar page that covers a broad topic, supported by cluster pages that cover specific subtopics. Internal links connect the cluster pages to the pillar, creating a topic hub that signals expertise to search engines.
- Content gap analysis: Identifying keywords your competitors rank for that you don't. This isn't about copying competitors; it's about finding opportunities they've missed.
- Content calendar: A schedule that accounts for seasonality, product launches, and industry events. The calendar should prioritize pages that have the highest potential for traffic and conversions.
- Measurement framework: Clear KPIs for each piece of content—not just rankings, but organic traffic, engagement metrics, and conversion rates.
Link Building: The Riskiest Part of SEO
Link building is the most controversial and risk-prone area of SEO. Done right, it builds authority and drives referral traffic. Done wrong, it can harm your site's performance.
White-Hat vs. Black-Hat Link Building
| Approach | Methods | Risk Level | Long-Term Value |
|---|---|---|---|
| White-hat | Guest posting on authoritative sites, digital PR, broken link building, resource page outreach | Low | High; builds sustainable authority |
| Gray-hat | Paid links, link exchanges, blog comment spam, forum signatures | Medium | Moderate; may work temporarily but can be risky |
| Black-hat | Private blog networks (PBNs), hacked links, automated link farms, link injection | High | None; will likely harm your site |
Any agency that promises "guaranteed first page ranking" through link building is likely overselling. No legitimate agency can guarantee rankings because they don't control Google's algorithm.
What a good agency does: They audit your existing backlink profile to identify toxic links that should be disavowed. They develop a link building strategy based on your industry, competitors, and available resources. They focus on earning links through quality content, digital PR, and genuine relationships—not buying them.
What to ask: "Show me examples of links you've earned for other clients. Can I see the outreach emails? How do you vet the sites you're targeting for link placement?"
The Technical SEO Audit Checklist
When you're evaluating an agency or conducting your own audit, use this checklist to ensure nothing is missed:
Crawl and Indexation
- Run a full site crawl using Screaming Frog or Sitebulb
- Check for 4xx and 5xx errors
- Verify XML sitemap is submitted to Google Search Console
- Check robots.txt for accidental blocking of important pages
- Review crawl stats in Google Search Console for anomalies
- Identify orphan pages (pages with no internal links)
Site Structure and URLs
- Ensure URL structure is logical and contains target keywords where appropriate
- Check for URL parameters that create duplicate content
- Verify canonical tags point to the correct preferred URL
- Review internal linking structure for logical hierarchy
- Check pagination for proper rel=next/prev implementation
Core Web Vitals
- Run Lighthouse report for mobile and desktop
- Analyze LCP: check server response time, image optimization, render-blocking resources
- Analyze CLS: check for layout shifts caused by ads, fonts, or images
- Analyze INP: check JavaScript execution time and third-party script impact
- Verify real user monitoring data in Google Search Console
On-Page Elements
- Review title tags for length, keyword placement, and uniqueness
- Review meta descriptions for compelling copy and keyword inclusion
- Check H1 tags for proper structure and keyword targeting
- Verify image alt text is descriptive and includes relevant keywords
- Ensure schema markup is implemented correctly (if applicable)
Content and Keywords
- Conduct keyword research for target topics
- Map keywords to search intent (informational, navigational, commercial, transactional)
- Identify content gaps compared to competitors
- Review existing content for quality, accuracy, and freshness
- Plan topic clusters and pillar pages
Backlink Profile
- Run a backlink audit using Ahrefs or Majestic
- Identify toxic links (spammy sites, PBNs, irrelevant directories)
- Disavow toxic links in Google Search Console
- Analyze competitor backlink profiles for opportunities
- Develop a white-hat link building strategy
What Can Go Wrong: Risk Awareness
Technical SEO isn't just about doing things right—it's about avoiding catastrophic mistakes. Here are the most common risks and how to mitigate them:
Wrong Redirects
Using 302 (temporary) redirects when you should use 301 (permanent) redirects, or vice versa. 301 redirects are generally preferred for passing link equity permanently, while 302 redirects may be used for temporary moves. Avoid redirect chains (A → B → C) that slow down page load and waste crawl budget.Poor Core Web Vitals Implementation
Rushing to fix Core Web Vitals without understanding the root cause can make things worse. For example, lazy-loading the hero image to improve LCP might actually increase LCP if the lazy-load script is render-blocking. Always test changes before deploying.Black-Hat Links
Even if you don't buy links, competitors might use negative SEO to point spammy links at your site. Regular backlink audits are essential to catch this early. If you find toxic links, disavow them immediately.Duplicate Content from CMS Issues
Many content management systems create duplicate content automatically. WordPress can create duplicate versions of posts through tags, categories, and archives. E-commerce platforms like Shopify and Magento often generate parameterized URLs for filters and sorting. These must be handled with canonical tags or robots.txt directives.When to Walk Away from an Agency
Not all agencies are created equal. Here are signs that an agency doesn't know what they're doing:
- They promise "guaranteed first page rankings"
- They use black-hat link building techniques (PBNs, link farms)
- They can't explain what crawl budget is or how it affects your site
- They don't audit Core Web Vitals before starting optimization
- They recommend keyword stuffing or other outdated tactics
- They don't provide detailed reports with actionable recommendations
- They blame Google algorithm updates for every ranking drop
Final Checklist: Before You Sign the Contract
Before you commit to any SEO agency, make sure they've addressed these points:
- They've conducted a technical SEO audit and shared the results
- They've explained the crawl budget implications for your site
- They've identified and prioritized Core Web Vitals improvements
- They've mapped keywords to search intent, not just volume
- They've developed a content strategy with topic clusters
- They've audited your backlink profile and disavowed toxic links
- They've set realistic expectations (no guaranteed rankings)
- They've provided a timeline with measurable milestones
- They've explained how they'll report progress (metrics, not vanity numbers)
Now go find an agency that actually understands crawl budget and Core Web Vitals. Your website—and your rankings—will thank you.

Reader Comments (0)