Your Technical SEO Health Checklist: What Every Site Owner Should Audit Right Now
You’ve probably heard that “content is king” and “links are the backbone of SEO.” Both statements are true, but neither matters if Google’s crawlers can’t access your pages, your site loads like a dial-up connection, or your redirects are pointing visitors into a 404 abyss. Technical SEO is the foundation that makes every other optimization possible—and it’s the area where small mistakes can quietly bleed your traffic for months.
This checklist walks you through the critical technical health checks every site needs, explains why each one matters, and shows you how to brief an SEO agency when you need expert hands. We’ll cover crawlability, site performance, content duplication, and link profile hygiene. Let’s start with the machine that decides whether your site even gets seen.
1. Crawl Budget & Site Architecture: Are You Wasting Google’s Time?
Every time Googlebot visits your site, it has a limited number of pages it can crawl in a single session. That’s your crawl budget. If your site has 10,000 pages but only 200 are valuable, you’re wasting budget on thin content, paginated archives, or duplicate URLs. The result? Your money pages get crawled less frequently, and rankings stagnate.
What to check:
- Crawl rate in Google Search Console. Look for spikes or drops. A sudden decrease often signals a technical issue (server errors, blocked resources).
- Server response time. If your server responds slowly, Googlebot may slow down or abandon the crawl. Aim for fast response times.
- Number of crawled vs. indexed pages. A huge gap means Google is finding pages it doesn’t consider valuable enough to index.
- Consolidate thin content. Merge similar pages into one comprehensive resource.
- Use `noindex` tags on low-value pages (tag archives, filter pages, staging environments).
- Ensure your XML sitemap only contains canonical, indexable URLs.
2. Core Web Vitals: The User Experience Metric That Can Tank Your Rankings
Google’s Core Web Vitals are a set of real-world speed and stability metrics that directly impact search rankings. They measure three things:
- LCP (Largest Contentful Paint): How fast the main content loads. Target: under 2.5 seconds.
- FID (First Input Delay) / INP (Interaction to Next Paint): How responsive the page is to user clicks. Target: under 100ms for FID, under 200ms for INP.
- CLS (Cumulative Layout Shift): How much the page layout shifts unexpectedly. Target: under 0.1.
Common pitfalls:
- Unoptimized images (serving 2000px-wide photos on a mobile screen).
- Render-blocking JavaScript (scripts that delay the page from displaying).
- Third-party scripts (analytics, ads, chatbots) that load before the main content.
- Compress images using WebP format and lazy-load below-the-fold images.
- Defer non-critical JavaScript and inline critical CSS.
- Set explicit width and height on all images and embeds to prevent layout shifts.
| Metric | Good | Needs Improvement | Poor |
|---|---|---|---|
| LCP | ≤ 2.5s | 2.5s – 4.0s | > 4.0s |
| FID | ≤ 100ms | 100ms – 300ms | > 300ms |
| CLS | ≤ 0.1 | 0.1 – 0.25 | > 0.25 |
What an agency should do: Run a Core Web Vitals audit using Lighthouse and CrUX data, identify the specific scripts or assets causing delays, and provide a prioritized fix list. Beware of agencies that promise “instant speed boost” without a detailed technical breakdown—real improvements require server-level changes, image optimization, and often a CDN migration.
3. XML Sitemaps & robots.txt: Your Site’s Welcome Mat
Your XML sitemap tells Google which pages exist and how often they change. Your robots.txt file tells Google which pages it should not crawl. Both are easy to set up, but even easier to mess up.

Common mistakes:
- Sitemap includes `noindex` pages (waste of crawl budget).
- Sitemap excludes important pages (they may never get indexed).
- robots.txt blocks CSS or JavaScript files (Google can’t render the page properly).
- robots.txt accidentally blocks the entire site (yes, this happens more often than you’d think).
- Submit your sitemap in Google Search Console and check for errors (e.g., URLs returning 404, blocked by robots.txt).
- Open your robots.txt file and verify that it allows access to your site’s CSS, JS, and image files.
- Ensure your sitemap only contains canonical URLs—no paginated pages, no filter pages, no session IDs.
- Use a dynamic sitemap generator that automatically excludes `noindex` pages.
- Add a `Disallow:` directive only for sensitive or non-public directories (e.g., `/admin/`, `/staging/`).
- Test your robots.txt using the “robots.txt Tester” in Google Search Console.
4. Canonical Tags & Duplicate Content: Stop Competing Against Yourself
Duplicate content doesn’t trigger a penalty, but it does dilute your ranking signals. When Google finds two pages with nearly identical content, it has to decide which one to show in search results. Sometimes it picks the wrong one. That’s where canonical tags come in.
A canonical tag (`rel="canonical"`) tells Google: “This URL is the master version. Ignore the others.”
Where duplicate content typically lives:
- HTTP vs. HTTPS versions of the same page.
- `www` vs. non-`www` versions.
- Trailing slash vs. no trailing slash.
- URL parameters (e.g., `?sort=price`, `?color=red`).
- Printer-friendly versions or AMP pages.
- Use a crawler (like Screaming Frog or Sitebulb) to identify pages with identical or near-identical content.
- Check that every page has a self-referencing canonical tag (points to itself).
- Verify that canonical tags are consistent across all versions of a page.
- Set a preferred domain in Google Search Console (www or non-www).
- Use 301 redirects to consolidate duplicate URLs into one canonical version.
- For parameter-based duplicates, use Google Search Console’s URL Parameters tool to tell Google how to handle them.
- Canonical tags pointing to 404 pages (wasted signal).
- Cross-domain canonical tags (e.g., pointing to a syndicated version on another site) without proper attribution.
- Hreflang tags conflicting with canonical tags for multilingual sites.
5. On-Page Optimization & Keyword Research: Where Content Meets Strategy
On-page optimization is the practice of aligning your page content with search intent and technical best practices. It starts with keyword research—not just finding high-volume terms, but understanding intent mapping. Are users looking for information (informational), a specific website (navigational), or ready to buy (transactional)? A page optimized for the wrong intent will never rank well, no matter how many backlinks it has.
The on-page checklist:
- Title tag: Unique, includes primary keyword, under 60 characters.
- Meta description: Compelling, includes primary keyword and a call-to-action, under 160 characters.
- Header structure: One H1 (matches the page title), logical H2/H3 hierarchy.
- Keyword placement: Primary keyword in the first 100 words, naturally throughout the content, and in at least one H2.
- Internal linking: Link to related pages using descriptive anchor text.
- Image optimization: Descriptive file names, alt text, compressed format.
Common mistakes:
- Targeting only high-volume keywords (high competition, low conversion).
- Ignoring long-tail keywords (lower volume, higher intent, easier to rank).
- Writing for search engines instead of humans (keyword stuffing, unnatural phrasing).

6. Link Building & Backlink Profile: Quality Over Quantity
Link building remains one of the strongest ranking signals, but the era of “any link is a good link” is long dead. A link from a spammy directory can hurt your backlink profile, while quality links from relevant sites help it. Metrics like Domain Authority and Trust Flow give you a snapshot of your profile’s health, but they’re not ranking factors themselves—they’re correlational indicators.
What to check:
- Backlink growth: Are you gaining links naturally, or is there a sudden spike (possible negative SEO attack)?
- Link quality: Are links coming from relevant, authoritative sites in your niche?
- Anchor text distribution: Too many exact-match anchors (e.g., “best SEO agency”) can trigger a spam filter.
- Toxic links: Links from link farms, PBNs, or sites with low Trust Flow.
- Use a tool like Ahrefs or Majestic to audit your backlink profile monthly.
- Disavow toxic links using Google’s Disavow Tool—but only after confirming they’re truly harmful (not just low-quality).
- Build links through genuine outreach: guest posts on industry blogs, resource page inclusion, broken link building, or digital PR.
| Metric | Healthy | Risky |
|---|---|---|
| Domain Authority | Steady growth | Sudden drop or stagnation |
| Trust Flow | Balanced ratio with Citation Flow | High CF, low TF (unnatural ratio) |
| Anchor text | Branded + generic + partial match | High proportion of exact-match commercial anchors |
| Linking domains | Gradual increase from relevant sites | Spike from unrelated, low-DA sites |
Risks of black-hat link building:
- Google’s manual actions can remove your site from search results entirely.
- Recovery requires a lengthy reconsideration request and often a full link audit.
- Even if you don’t get penalized immediately, low-quality links offer diminishing returns.
7. Putting It All Together: Your Technical SEO Checklist
Here’s a consolidated checklist you can use to run your own audit or brief an agency:
- Crawl budget: Check crawl stats in GSC. Are important pages being crawled? Is your server responding quickly?
- Core Web Vitals: Run Lighthouse and CrUX reports. Are LCP, FID/INP, and CLS in the green?
- XML sitemap: Is it submitted? Does it exclude `noindex` pages? Are there errors?
- robots.txt: Is it blocking essential resources? Does it allow access to CSS and JS?
- Canonical tags: Does every page have a self-referencing canonical? Are there cross-domain conflicts?
- Duplicate content: Use a crawler to find near-identical pages. Consolidate or `noindex`.
- On-page optimization: Check title tags, meta descriptions, headers, and keyword placement.
- Backlink profile: Audit for toxic links, anchor text distribution, and growth pattern.
If you’re working with an agency, use this checklist as your briefing document. Ask for clear deliverables, timelines, and reporting cadences. A good agency will welcome the scrutiny—because they know that a healthy technical foundation is what makes every other SEO effort pay off.
For more on how to structure your SEO strategy, check out our guide on helpful content updates and learn how to align technical fixes with content that actually serves your audience.

Reader Comments (0)