The Technical SEO Checklist: What a Top-Tier Agency Should Audit for Real Site Health

The Technical SEO Checklist: What a Top-Tier Agency Should Audit for Real Site Health

You’ve hired an SEO agency, or you’re about to. The pitch deck looked slick—promises of “traffic growth,” “Core Web Vitals fixes,” and “crawl optimization.” But here’s the uncomfortable truth: many agencies treat technical SEO as a one-time checkbox. They run a tool, send a PDF, and move on to content writing. Real technical SEO isn’t a report; it’s a continuous diagnostic process that involves your server configuration, your development team, and your content strategy. If the agency isn’t digging into the specifics of how search engines actually interact with your site’s infrastructure, you’re paying for surface-level polish, not structural health.

This article breaks down the specific services a top-tier technical SEO agency should deliver. We’re not talking about generic “site audits.” We’re talking about the granular checks that separate a site that competes from a site that merely exists. Use this as your checklist when evaluating proposals or reviewing deliverables. If an agency skips these steps, ask why.

1. Crawl Budget Analysis: Not Just a Number

Most site audits will tell you how many pages Googlebot crawled. A top-tier audit explains why the crawl happened that way and what it means for your site’s indexation. Crawl budget is the allocation of resources Google gives to crawl your site. For small sites (under a few thousand pages), it’s rarely an issue. For large e-commerce platforms, news sites, or directories with hundreds of thousands of URLs, it’s critical.

What a technical audit should cover:

  • Crawl frequency vs. crawl depth: Are bots spending time on thin pages (tag archives, filtered parameter URLs) instead of your money pages?
  • Server response times: Slow server response (over 200ms) directly reduces crawl rate. The agency should correlate server logs with crawl data.
  • Redundant URL patterns: Session IDs, tracking parameters, and infinite filter combinations that waste crawl budget.
  • The robots.txt file: Is it accidentally blocking important sections? Is it allowing crawlers into areas that should be excluded?
The checklist item: Ask the agency to provide a crawl budget analysis that includes server log file data, not just third-party tool screenshots. If they can’t access your server logs, they’re guessing.

2. Core Web Vitals: Beyond the Lab Data

Core Web Vitals (LCP, FID/INP, CLS) are now ranking signals. But many agencies treat them as a dashboard metric to “improve” without understanding the underlying architecture. A genuine audit doesn’t just flag “LCP too high”—it identifies whether the issue is render-blocking JavaScript, unoptimized images, or slow server TTFB.

What a technical audit should cover:

  • Field data vs. lab data: The agency should analyze Chrome User Experience Report (CrUX) data (real user experiences) alongside Lighthouse scores (simulated conditions). These often disagree.
  • Specific resource analysis: Which third-party scripts are causing layout shifts? Which image format change (WebP, AVIF) would reduce LCP by the largest margin?
  • INP (Interaction to Next Paint): For sites with heavy JavaScript, this is the new metric that matters. The agency should identify long tasks and suggest code splitting or deferral.
The checklist item: Request a prioritized list of Core Web Vitals fixes ordered by estimated impact on user experience, not just by ease of implementation.

3. Duplicate Content and Canonicalization: The Silent Leak

Duplicate content isn’t a penalty—it’s a dilution problem. When Google sees the same content on multiple URLs, it splits ranking signals across them. The canonical tag is your primary tool to consolidate signals, but it’s often misconfigured.

What a technical audit should cover:

  • Cross-domain duplication: If you syndicate content or have multiple subdomains, the agency should verify that canonical tags point to the original source.
  • Parameter handling: Sites with faceted navigation (e.g., color filters, size filters) often generate thousands of near-duplicate URLs. The audit should recommend either noindex, canonical consolidation, or JavaScript-based filtering.
  • HTTP vs. HTTPS and www vs. non-www: Still a common issue. The audit should confirm that all versions redirect cleanly to a single preferred domain.
The checklist item: Look for a duplicate content analysis that quantifies the percentage of your indexed URLs that are duplicates. If the number exceeds 5%, it’s a problem.

4. On-Page Optimization: The Structure of Relevance

On-page optimization is often reduced to “write meta descriptions and include keywords.” That’s table stakes. A top-tier agency treats on-page as a structural exercise that aligns with search intent and technical constraints.

What a technical audit should cover:

  • Keyword research and intent mapping: The agency should map keywords to specific stages of the user journey (informational, navigational, commercial, transactional). This isn’t just a list of terms—it’s a strategic document that tells you which pages should target which queries.
  • Header hierarchy: H1s should be unique per page and describe the page’s primary topic. H2s should support the H1. The audit should flag missing or duplicated headers.
  • Internal linking structure: Are your most important pages getting enough internal link juice? The audit should identify orphan pages (pages with no internal links) and suggest a logical linking hierarchy.
  • Schema markup: The agency should check for appropriate structured data (Product, FAQ, Article, LocalBusiness) and validate it using Google’s Rich Results Test.
The checklist item: Ask for a sample page analysis that shows how keyword research informed the header structure, internal links, and content outline. If the agency can’t explain the “why” behind the optimization, it’s not strategic.

5. Link Building and Backlink Profile: Quality Over Quantity

Link building is the riskiest part of SEO. A bad link profile can trigger manual actions or algorithmic demotions. A top-tier agency doesn’t just build links—it manages your backlink profile as a risk portfolio.

What a technical audit should cover:

  • Backlink profile audit: The agency should analyze your existing backlinks for toxic signals (spammy domains, irrelevant anchor text, sudden spikes). They should use tools like Majestic (Trust Flow, Citation Flow) or Ahrefs (Domain Rating) to assess domain quality.
  • Disavow strategy: If toxic links exist, the agency should recommend a disavow file and explain how to submit it via Google Search Console.
  • Outreach strategy: Legitimate link building involves creating linkable assets (original research, comprehensive guides, interactive tools) and reaching out to relevant publishers. The agency should provide a sample outreach email and a list of target domains.
The checklist item: Ask the agency to show you their process for vetting potential link targets. If they can’t articulate how they distinguish a good domain from a bad one, they’re likely building low-quality links.

6. XML Sitemaps and robots.txt: The Foundation

These two files are the first things Googlebot looks at when it visits your site. They’re also the most commonly misconfigured.

What a technical audit should cover:

  • XML sitemap: Should only include indexable, canonical URLs. It should exclude paginated pages, filter pages, and duplicate content. The agency should check for errors (e.g., URLs returning 404 or 301, URLs blocked by robots.txt).
  • robots.txt: Should not block CSS, JavaScript, or image files (this can break rendering). It should block areas you don’t want indexed (admin panels, staging environments, duplicate content sections). The audit should verify that the file is accessible and returns a 200 status.
The checklist item: Request a screenshot of the sitemap submission in Google Search Console and a copy of the robots.txt file with annotations explaining each directive.

7. Site Performance and Core Web Vitals: The Ongoing Commitment

Site performance isn’t a one-time fix. It’s a continuous process that requires monitoring after every code deployment, plugin update, or content addition.

What a technical audit should cover:

  • Performance baseline: The agency should establish a baseline for LCP, FID/INP, CLS, and TTFB using real user monitoring (RUM) data.
  • Performance budget: A set of thresholds (e.g., LCP under 2.5 seconds, CLS under 0.1) that the site must maintain. The agency should recommend tools (e.g., Lighthouse CI, WebPageTest) for automated testing.
  • Ongoing monitoring: The agency should set up alerts for performance regressions and provide a monthly performance report.
The checklist item: Ask the agency how they handle performance regressions. Do they have a process for rolling back changes that hurt Core Web Vitals? If not, your site’s performance will degrade over time.

Summary Checklist for Choosing an Agency

Service AreaWhat to Look ForRed Flags
Crawl BudgetServer log analysis, crawl frequency vs. depthOnly third-party tool data
Core Web VitalsField data analysis, specific resource fixesLab data only, vague recommendations
Duplicate ContentQuantified duplicate percentage, canonical verificationNo mention of canonical tags
On-Page OptimizationIntent mapping, header hierarchy, internal linkingGeneric keyword lists, no structural analysis
Link BuildingBacklink profile audit, disavow strategy, outreach samplesGuaranteed links, no vetting process
XML Sitemaps & robots.txtAnnotated files, error checkingMissing files, blocking CSS/JS
Site PerformanceBaseline, performance budget, monitoring alertsOne-time fix, no ongoing monitoring

A top-tier technical SEO agency doesn’t just tell you what’s broken—it explains why it matters, how to fix it, and how to keep it fixed. Use this checklist to separate the real experts from the report-generators. Your site’s health depends on it.

Ready to dive deeper? Check out our guides on technical SEO audits and Core Web Vitals optimization.

Wendy Garza

Wendy Garza

Technical SEO Specialist

Elena focuses on site architecture, crawl efficiency, and structured data. She breaks down complex technical issues into clear, actionable steps.

Reader Comments (0)

Leave a comment