The Technical SEO Checklist: What a Top-Tier Agency Should Audit for Real Site Health
You’ve hired an SEO agency, or you’re about to. The pitch deck looked slick—promises of “traffic growth,” “Core Web Vitals fixes,” and “crawl optimization.” But here’s the uncomfortable truth: many agencies treat technical SEO as a one-time checkbox. They run a tool, send a PDF, and move on to content writing. Real technical SEO isn’t a report; it’s a continuous diagnostic process that involves your server configuration, your development team, and your content strategy. If the agency isn’t digging into the specifics of how search engines actually interact with your site’s infrastructure, you’re paying for surface-level polish, not structural health.
This article breaks down the specific services a top-tier technical SEO agency should deliver. We’re not talking about generic “site audits.” We’re talking about the granular checks that separate a site that competes from a site that merely exists. Use this as your checklist when evaluating proposals or reviewing deliverables. If an agency skips these steps, ask why.
1. Crawl Budget Analysis: Not Just a Number
Most site audits will tell you how many pages Googlebot crawled. A top-tier audit explains why the crawl happened that way and what it means for your site’s indexation. Crawl budget is the allocation of resources Google gives to crawl your site. For small sites (under a few thousand pages), it’s rarely an issue. For large e-commerce platforms, news sites, or directories with hundreds of thousands of URLs, it’s critical.
What a technical audit should cover:
- Crawl frequency vs. crawl depth: Are bots spending time on thin pages (tag archives, filtered parameter URLs) instead of your money pages?
- Server response times: Slow server response (over 200ms) directly reduces crawl rate. The agency should correlate server logs with crawl data.
- Redundant URL patterns: Session IDs, tracking parameters, and infinite filter combinations that waste crawl budget.
- The robots.txt file: Is it accidentally blocking important sections? Is it allowing crawlers into areas that should be excluded?
2. Core Web Vitals: Beyond the Lab Data
Core Web Vitals (LCP, FID/INP, CLS) are now ranking signals. But many agencies treat them as a dashboard metric to “improve” without understanding the underlying architecture. A genuine audit doesn’t just flag “LCP too high”—it identifies whether the issue is render-blocking JavaScript, unoptimized images, or slow server TTFB.
What a technical audit should cover:
- Field data vs. lab data: The agency should analyze Chrome User Experience Report (CrUX) data (real user experiences) alongside Lighthouse scores (simulated conditions). These often disagree.
- Specific resource analysis: Which third-party scripts are causing layout shifts? Which image format change (WebP, AVIF) would reduce LCP by the largest margin?
- INP (Interaction to Next Paint): For sites with heavy JavaScript, this is the new metric that matters. The agency should identify long tasks and suggest code splitting or deferral.
3. Duplicate Content and Canonicalization: The Silent Leak
Duplicate content isn’t a penalty—it’s a dilution problem. When Google sees the same content on multiple URLs, it splits ranking signals across them. The canonical tag is your primary tool to consolidate signals, but it’s often misconfigured.

What a technical audit should cover:
- Cross-domain duplication: If you syndicate content or have multiple subdomains, the agency should verify that canonical tags point to the original source.
- Parameter handling: Sites with faceted navigation (e.g., color filters, size filters) often generate thousands of near-duplicate URLs. The audit should recommend either noindex, canonical consolidation, or JavaScript-based filtering.
- HTTP vs. HTTPS and www vs. non-www: Still a common issue. The audit should confirm that all versions redirect cleanly to a single preferred domain.
4. On-Page Optimization: The Structure of Relevance
On-page optimization is often reduced to “write meta descriptions and include keywords.” That’s table stakes. A top-tier agency treats on-page as a structural exercise that aligns with search intent and technical constraints.
What a technical audit should cover:
- Keyword research and intent mapping: The agency should map keywords to specific stages of the user journey (informational, navigational, commercial, transactional). This isn’t just a list of terms—it’s a strategic document that tells you which pages should target which queries.
- Header hierarchy: H1s should be unique per page and describe the page’s primary topic. H2s should support the H1. The audit should flag missing or duplicated headers.
- Internal linking structure: Are your most important pages getting enough internal link juice? The audit should identify orphan pages (pages with no internal links) and suggest a logical linking hierarchy.
- Schema markup: The agency should check for appropriate structured data (Product, FAQ, Article, LocalBusiness) and validate it using Google’s Rich Results Test.
5. Link Building and Backlink Profile: Quality Over Quantity
Link building is the riskiest part of SEO. A bad link profile can trigger manual actions or algorithmic demotions. A top-tier agency doesn’t just build links—it manages your backlink profile as a risk portfolio.
What a technical audit should cover:
- Backlink profile audit: The agency should analyze your existing backlinks for toxic signals (spammy domains, irrelevant anchor text, sudden spikes). They should use tools like Majestic (Trust Flow, Citation Flow) or Ahrefs (Domain Rating) to assess domain quality.
- Disavow strategy: If toxic links exist, the agency should recommend a disavow file and explain how to submit it via Google Search Console.
- Outreach strategy: Legitimate link building involves creating linkable assets (original research, comprehensive guides, interactive tools) and reaching out to relevant publishers. The agency should provide a sample outreach email and a list of target domains.
6. XML Sitemaps and robots.txt: The Foundation
These two files are the first things Googlebot looks at when it visits your site. They’re also the most commonly misconfigured.

What a technical audit should cover:
- XML sitemap: Should only include indexable, canonical URLs. It should exclude paginated pages, filter pages, and duplicate content. The agency should check for errors (e.g., URLs returning 404 or 301, URLs blocked by robots.txt).
- robots.txt: Should not block CSS, JavaScript, or image files (this can break rendering). It should block areas you don’t want indexed (admin panels, staging environments, duplicate content sections). The audit should verify that the file is accessible and returns a 200 status.
7. Site Performance and Core Web Vitals: The Ongoing Commitment
Site performance isn’t a one-time fix. It’s a continuous process that requires monitoring after every code deployment, plugin update, or content addition.
What a technical audit should cover:
- Performance baseline: The agency should establish a baseline for LCP, FID/INP, CLS, and TTFB using real user monitoring (RUM) data.
- Performance budget: A set of thresholds (e.g., LCP under 2.5 seconds, CLS under 0.1) that the site must maintain. The agency should recommend tools (e.g., Lighthouse CI, WebPageTest) for automated testing.
- Ongoing monitoring: The agency should set up alerts for performance regressions and provide a monthly performance report.
Summary Checklist for Choosing an Agency
| Service Area | What to Look For | Red Flags |
|---|---|---|
| Crawl Budget | Server log analysis, crawl frequency vs. depth | Only third-party tool data |
| Core Web Vitals | Field data analysis, specific resource fixes | Lab data only, vague recommendations |
| Duplicate Content | Quantified duplicate percentage, canonical verification | No mention of canonical tags |
| On-Page Optimization | Intent mapping, header hierarchy, internal linking | Generic keyword lists, no structural analysis |
| Link Building | Backlink profile audit, disavow strategy, outreach samples | Guaranteed links, no vetting process |
| XML Sitemaps & robots.txt | Annotated files, error checking | Missing files, blocking CSS/JS |
| Site Performance | Baseline, performance budget, monitoring alerts | One-time fix, no ongoing monitoring |
A top-tier technical SEO agency doesn’t just tell you what’s broken—it explains why it matters, how to fix it, and how to keep it fixed. Use this checklist to separate the real experts from the report-generators. Your site’s health depends on it.
Ready to dive deeper? Check out our guides on technical SEO audits and Core Web Vitals optimization.

Reader Comments (0)