1. The Technical SEO Audit: What It Must Cover

When you engage an SEO agency for technical audits and site health, you are not buying a one-time report. You are investing in a systematic process that diagnoses how search engines discover, crawl, render, and index your web pages. The approach of a competent agency often differs from others in the rigor of their technical foundation—and in their willingness to tell you what you don't want to hear about your site's infrastructure.

This checklist is designed for marketing directors, product owners, and technical leads who need to brief an agency or evaluate their output. It covers the core pillars of technical SEO: crawl budget management, Core Web Vitals optimization, content duplication control, and safe link acquisition. Each section includes both the educational rationale and the actionable steps you should expect from a professional partner.

1. The Technical SEO Audit: What It Must Cover

A proper technical SEO audit is not a list of broken links and missing alt tags. It is a deep inspection of how your site's architecture interacts with Google's crawling and rendering pipeline. The agency should begin by analyzing your server logs—not just crawling your site with a tool. Server log analysis can reveal which pages Googlebot actually visits, how often, and with what HTTP response codes. Without this, you may be guessing at crawl behavior.

The audit must also assess JavaScript rendering. Many modern single-page applications (SPAs) and React-based sites may fail to deliver content to Googlebot because they rely on client-side rendering without proper hydration. An expert agency will test your critical pages using Google's URL Inspection Tool and a headless browser to verify that text, links, and structured data are visible after JavaScript execution.

Key deliverables from a technical audit:

  • Crawl budget analysis: identification of wasted crawl activity on low-value pages (e.g., parameterized URLs, infinite scroll, faceted navigation).
  • Rendering audit: confirmation that all meaningful content is accessible without user interaction (e.g., no lazy-loaded content hidden behind clicks).
  • Indexation coverage report: a breakdown of indexed vs. crawled pages, with reasons for exclusion (e.g., noindex directives, canonicalization issues, blocked resources).
  • Core Web Vitals baseline: field data from the Chrome User Experience Report (CrUX) for LCP, FID/INP, and CLS, segmented by device and connection type.
What to watch for: An agency that only runs Screaming Frog and hands you a 200-page PDF without server log analysis may not be performing a thorough technical audit—they are performing a crawl. Consider requesting server-side data to ensure depth.

2. Crawl Budget & XML Sitemap Management

Crawl budget is the finite number of URLs Googlebot can and will crawl on your site within a given timeframe. For large sites (over 10,000 pages), mismanagement of crawl budget can potentially prevent important pages from being discovered for extended periods. The agency should prioritize crawl efficiency by:

  • Consolidating duplicate or thin pages via canonical tags or 301 redirects.
  • Blocking low-value URL parameters (e.g., sorting, filtering, session IDs) in robots.txt or via the `?` parameter handling in Google Search Console.
  • Ensuring your XML sitemap contains only canonical, indexable URLs with a lastmod date that reflects actual content changes. Do not include paginated pages or filtered views.
Common crawl budget mistakes:
  • Allowing infinite crawl paths through faceted navigation without `noindex` or `nofollow` controls.
  • Serving 200 HTTP status codes on error pages (soft 404s) instead of 404 or 410.
  • Using JavaScript redirects instead of server-side 301s, which wastes crawl time.
The agency should also configure your robots.txt to allow Googlebot access to critical resources (CSS, JavaScript, images) while disallowing low-value directories (e.g., `/search/`, `/tag/`, `/user/`). A common mistake is blocking JavaScript files entirely, which can break rendering for Googlebot.

Crawl Budget FactorImpactAgency Action Required
Parameterized URLsHigh wasteSet parameter handling in GSC or block via robots.txt
Soft 404sMedium wasteFix server response codes
JavaScript redirectsHigh wasteReplace with 301 redirects
Orphan pagesCannot be crawledAdd to sitemap or internal links
Infinite scroll without crawlable paginationLow discoveryImplement `rel="next"` / `rel="prev"` or load-more buttons with unique URLs

3. Core Web Vitals: Beyond the Surface Metrics

Core Web Vitals are not just a ranking factor—they are a proxy for user experience. However, many agencies treat them as a checklist: "LCP under 2.5 seconds, CLS under 0.1, FID under 100ms." That is insufficient. A professional engagement will dig into the root causes of poor metrics.

Largest Contentful Paint (LCP) is often caused by slow server response times (TTFB), render-blocking resources, or unoptimized images. The agency should:

  • Measure TTFB from multiple geographic locations using a tool like WebPageTest.
  • Identify the LCP element (usually a hero image or heading) and optimize its delivery: compress images, use modern formats (WebP, AVIF), lazy-load below-the-fold content, and implement resource hints (preload, prefetch).
  • Evaluate your hosting infrastructure—shared hosting or poor CDN configuration can cap your LCP improvement regardless of on-page changes.
Cumulative Layout Shift (CLS) is frequently caused by ads, embeds, or web fonts loading after the initial paint. The agency should audit for:
  • Images and embeds without explicit width and height attributes.
  • Late-loading ads that push content down.
  • Web fonts that cause invisible text (FOIT) or flash of unstyled text (FOUT). Using `font-display: swap` with a fallback font is standard, but the fallback must match the final font's metrics to avoid layout shift.
Interaction to Next Paint (INP) is the newer metric replacing FID. It measures the time from a user interaction (click, tap, keypress) to the next paint. Poor INP often stems from heavy JavaScript execution on event handlers. The agency should profile your site's JavaScript bundles using Chrome DevTools' Performance tab and identify long tasks (>50ms). Solutions include code splitting, deferring non-critical scripts, and using web workers for heavy computations.

Warning: If an agency promises to "fix Core Web Vitals in one week" without first analyzing your server logs and field data, they may be applying superficial fixes that might not hold under real user conditions. Core Web Vitals improvement is often an iterative process that may require backend changes.

4. Duplicate Content & Canonicalization Strategy

Duplicate content is not a penalty—it is a dilution of ranking signals. When multiple URLs serve the same or very similar content, Google must choose which version to index. If it picks the wrong one, your traffic suffers. The agency must implement a clear canonicalization strategy that covers:

  • www vs. non-www: Choose one and 301 redirect the other.
  • HTTP vs. HTTPS: Ensure all traffic redirects to HTTPS with HSTS headers.
  • Trailing slashes: Decide on a standard and enforce it via redirects or canonical tags.
  • Parameter handling: Use `rel="canonical"` on paginated pages to point to the main listing page (or use self-referencing canonicals with `rel="next"` / `rel="prev"`).
  • Internationalization: For multilingual sites, use `hreflang` tags correctly. A common error is using `x-default` incorrectly or mixing `hreflang` with `rel="canonical"` in conflicting ways.
The agency should also audit for "near-duplicate" content—pages that are not identical but share 80-90% of the same text (e.g., product descriptions with minor variations). For e-commerce sites, this is a major issue. Solutions include:
  • Writing unique descriptions for each product.
  • Using canonical tags to point to the primary product page.
  • Implementing `noindex` on thin affiliate or landing pages.
Case in point: A site with many product pages that all use the same manufacturer description may see those pages compete with each other in search results, potentially resulting in none ranking well. The agency should flag this and recommend a content generation strategy (human-written or AI-assisted, with careful quality control).

Duplicate Content TypeRisk LevelCanonicalization Approach
Exact duplicate (e.g., HTTP vs HTTPS)Low301 redirect to preferred version
Near-duplicate product descriptionsHighUnique content or `noindex` on thin pages
Paginated content (e.g., /page/2/)MediumSelf-referencing canonicals with `rel=next/prev`
URL parameter variationsMediumParameter handling in GSC + canonical tag
Syndicated contentMediumLink back to original source + noindex on syndicated copy

5. On-Page Optimization & Intent Mapping

On-page optimization is where technical SEO meets content strategy. The agency should not simply stuff keywords into title tags and meta descriptions. Instead, they should map keywords to search intent and structure pages accordingly.

Intent mapping categories:

  • Informational: User wants to learn (e.g., "how to fix slow website"). Optimize for blog posts, guides, videos. Focus on clear headings, structured data (FAQ, HowTo), and internal links to related content.
  • Navigational: User wants to find a specific site (e.g., "SearchScope technical SEO"). Ensure your brand page ranks first with a clear title and sitelinks.
  • Commercial: User is comparing options (e.g., "best SEO agency for e-commerce"). Optimize product comparison pages, case studies, and reviews. Include schema markup for Product or Review.
  • Transactional: User wants to buy or sign up (e.g., "hire SEO agency pricing"). Optimize landing pages with clear CTAs, trust signals, and fast load times.
The agency should also audit your heading structure (H1-H6). A common mistake is using multiple H1 tags on a single page or skipping heading levels (e.g., H1 to H3 without an H2). This does not cause a penalty but can confuse search engines about content hierarchy. The fix is straightforward: ensure each page has one descriptive H1 that matches the primary keyword intent, and use H2s for major sections.

Title tags and meta descriptions remain important for click-through rates (CTR). The agency should write unique title tags for every page (no templated "Page Title | Site Name" without variation) and keep them under a reasonable length. Meta descriptions can be longer and should include the target keyword naturally while compelling a click. Do not duplicate meta descriptions across pages—this is a sign of lazy on-page work.

6. Link Building: Safe Acquisition & Profile Auditing

Link building is the most risk-prone part of any SEO campaign. Black-hat techniques—such as buying links from private blog networks (PBNs), using automated link exchanges, or spamming forum comments—can lead to manual penalties or algorithmic demotions (e.g., Google's Penguin update). A professional agency will never guarantee a specific number of backlinks or a fixed Domain Authority increase. Instead, they will focus on quality and relevance.

Safe link building strategies:

  • Content-based outreach: Creating genuinely useful resources (original research, data studies, comprehensive guides) and reaching out to relevant sites for natural citations.
  • Digital PR: Securing coverage from news outlets, industry publications, and blogs through newsworthy stories or expert commentary.
  • Broken link building: Finding broken resources on authoritative sites and offering your content as a replacement.
  • Guest posting on reputable sites: Writing high-quality articles for sites with editorial standards and relevant audiences. Avoid sites that accept any guest post with a link—these are often low-quality or part of a PBN.
  • Unlinked mentions: Finding brand mentions without a link and requesting the addition of a hyperlink.
The agency should also audit your existing backlink profile to identify toxic links that could harm your site. Tools like Majestic, Ahrefs, or Moz can flag links from spammy domains, link farms, or sites with low Trust Flow. If the agency finds a high volume of toxic links, they should:
  • Disavow them via Google's Disavow Tool (only after careful review—disavowing good links can hurt).
  • Document the disavowal process and rationale.
  • Monitor the profile monthly for new toxic links.
What to avoid: Any agency that offers "guaranteed backlinks from DA 50+ sites" or "100 links in 30 days" is likely using methods that may be against search engine guidelines. Real link building is slow, requires relationship building, and produces variable results. If an agency claims they can get you a link from The New York Times or Wikipedia without a compelling story or content, that claim may be unrealistic.

Link Building MethodRisk LevelExpected Timeline
Content-based outreachLowSeveral months for first results
Digital PRLow-MediumSeveral months for sustained coverage
Broken link buildingLowSeveral months
Guest posting on reputable sitesLow-MediumSeveral months
Buying links from PBNsHigh (penalty risk)Immediate traffic drop if caught
Automated link exchangesHighImmediate risk

7. Final Checklist: How to Brief Your Agency

When you brief an SEO agency for technical audits and site health, use this checklist to ensure alignment:

  1. Request a server log analysis as part of the initial audit. If they cannot provide this, ask why.
  2. Demand Core Web Vitals field data from CrUX, not just lab data from Lighthouse. Lab data is useful for debugging, but field data reflects real user experiences.
  3. Ask for a crawl budget optimization plan that includes parameter handling, robots.txt updates, and sitemap refinement.
  4. Require a duplicate content audit with specific recommendations for canonicalization and content consolidation.
  5. Insist on intent mapping for every target keyword, not just keyword stuffing.
  6. Get a backlink profile audit with a disavow plan if needed. Avoid agencies that dismiss toxic links as irrelevant.
  7. Set realistic timelines: Technical SEO improvements often take weeks to show impact on crawling and indexing. Core Web Vitals fixes may require development sprints. Link building results are typically visible over several months.
  8. Agree on reporting cadence: Monthly reports should include crawl stats, indexation changes, Core Web Vitals trends, and backlink acquisition. Avoid vanity metrics like "Domain Authority increased by 5 points"—focus on organic traffic and conversion rates.
A professional technical SEO engagement is a partnership, not a transaction. The agency should educate you on the why behind their recommendations, not just deliver a list of tasks. If they cannot explain how a canonical tag works or why server logs matter, they may lack the depth required for meaningful site health improvement.

For further reading, explore our guides on technical SEO audits and Core Web Vitals optimization.

Tyler Alvarado

Tyler Alvarado

Analytics and Reporting Reviewer

Jordan audits tracking setups and interprets SEO data to inform strategy. He focuses on actionable insights from analytics platforms.

Reader Comments (0)

Leave a comment