Technical SEO is the practice of optimizing website infrastructure so search engines can crawl, index, and render pages effectively. It covers site architecture, crawlability, page speed, mobile optimization, security, structured data, and JavaScript rendering.
What is Technical SEO?
Technical SEO encompasses all optimization activities that affect how search engines access, crawl, interpret, and index your website. Unlike on-page optimization which focuses on content and HTML elements, technical SEO targets website infrastructure.
Core technical SEO components:
- Crawlability - Ensuring search engines can access your pages
- Indexability - Ensuring crawled pages get added to the search index
- Site speed - Page loading performance and Core Web Vitals
- Mobile optimization - Responsive design and mobile usability
- Security - HTTPS implementation and secure connections
- Site architecture - URL structure, internal linking, navigation
- Structured data - Schema markup for rich results
- JavaScript rendering - Ensuring JS-dependent content is accessible to crawlers
Crawlability and Indexing
Crawlability determines whether search engines can discover and access your pages. Indexability determines whether those pages get added to the search index.
How Search Engine Crawling Works
Search engines use crawlers (bots) to discover and read website content:
- Crawler discovers URL (from sitemap, links, or direct submission)
- Crawler requests the page from server
- Server returns HTML response
- Crawler parses content and follows links
- Content gets processed for indexing
You can analyze your server log files to understand exactly how Googlebot crawls your site. Log file analysis reveals crawl frequency, which pages bots visit most, and where crawl budget gets wasted on low-value URLs.
Orphan Pages
Orphan pages are pages with no internal links pointing to them. Search engines struggle to discover these pages because crawlers follow links to find content. Even if an orphan page sits in your XML sitemap, the lack of internal links signals low importance.
Run Screaming Frog to find orphan pages by comparing crawled URLs against your sitemap. Then add contextual internal links from relevant pages to bring orphans back into your site’s link graph.
Crawl Budget Optimization
Crawl budget is the number of pages search engines crawl on your site within a given timeframe.
| Factor | Impact on Crawl Budget |
|---|---|
| Site speed | Faster sites get crawled more |
| Server errors | 5xx errors waste crawl budget |
| Duplicate content | Crawlers waste time on duplicates |
| Redirect chains | Multiple redirects slow crawling |
| Sitemap accuracy | Outdated sitemaps misdirect crawlers |
Crawl budget optimization strategies:
- Remove or noindex low-value pages
- Fix crawl errors in Google Search Console
- Reduce redirect chains (run Screaming Frog to find them)
- Update XML sitemap regularly
- Block non-essential pages in robots.txt
robots.txt Configuration
The robots.txt file tells crawlers which pages to access or avoid.
User-agent: *
Allow: /
Disallow: /admin/
Disallow: /api/
Disallow: /search?
Sitemap: https://example.com/sitemap.xml
robots.txt best practices:
- Place at domain root (example.com/robots.txt)
- Use for crawl guidance, not security
- Don’t block CSS/JS needed for rendering
- Include sitemap reference
- Test with Google’s robots.txt Tester
XML Sitemap
XML sitemaps help search engines discover pages and understand site structure.
Sitemap requirements:
- Maximum 50,000 URLs per sitemap
- Maximum 50MB uncompressed file size
- Include only canonical, indexable pages
- Update lastmod dates accurately
- Submit to Google Search Console
Core Web Vitals
Core Web Vitals are Google’s metrics for measuring user experience. They became a ranking factor in 2021.
Three Core Web Vitals Metrics
| Metric | Measures | Good Score |
|---|---|---|
| LCP (Largest Contentful Paint) | Loading performance | < 2.5 seconds |
| INP (Interaction to Next Paint) | Interactivity responsiveness | < 200 milliseconds |
| CLS (Cumulative Layout Shift) | Visual stability | < 0.1 |
Use PageSpeed Insights for Core Web Vitals diagnostics on individual pages. For site-wide data, check the Core Web Vitals report in Google Search Console. GTmetrix provides waterfall charts that help pinpoint exactly which resources cause slowdowns.
LCP Optimization
LCP measures how long the largest content element takes to load.
Common LCP issues:
- Slow server response time
- Render-blocking JavaScript/CSS
- Large, unoptimized images
- Client-side rendering delays
LCP solutions:
- Use CDN for faster delivery
- Preload critical resources
- Optimize and compress images
- Implement server-side rendering
INP Optimization
INP measures responsiveness to user interactions (replaced FID in 2024).
INP optimization strategies:
- Minimize JavaScript execution time
- Break up long tasks into smaller chunks
- Use web workers for heavy processing
- Optimize event handlers
- Profile bottlenecks with Chrome DevTools Performance panel
CLS Optimization
CLS measures unexpected layout shifts during page load.
Prevent CLS issues:
- Set explicit dimensions for images/videos
- Reserve space for ad units
- Avoid inserting content above existing content
- Use CSS transform for animations
Page Speed Optimization
Page speed affects both rankings and user experience. Slow pages lose visitors fast.
Speed Optimization Checklist
Server optimization:
- Enable GZIP/Brotli compression
- Use HTTP/2 or HTTP/3
- Implement server-side caching
- Choose hosting near target audience
For sites targeting Malaysian audiences, consider hosting providers with data centers in Singapore or Kuala Lumpur. A CDN with edge nodes in Southeast Asia (such as Cloudflare’s KUL PoP) reduces latency significantly compared to serving from US or European origins.
Resource optimization:
- Compress and resize images
- Use modern formats (WebP, AVIF)
- Minify CSS and JavaScript
- Remove unused code
Delivery optimization:
- Use Content Delivery Network (CDN)
- Implement browser caching
- Preload critical resources
- Lazy load below-fold content
Mobile-First Indexing
Google primarily uses the mobile version of content for indexing and ranking.
Mobile Optimization Requirements
| Aspect | Requirement |
|---|---|
| Responsive design | Content adapts to screen size |
| Same content | Mobile has same content as desktop |
| Touch-friendly | Buttons/links easily tappable |
| Readable text | No horizontal scrolling needed |
| Fast loading | Optimized for mobile networks |
Mobile usability checklist:
- Use responsive design (not separate mobile site)
- Ensure text readable without zooming
- Size tap targets appropriately (48px minimum)
- Avoid horizontal scrolling
- Test with Google Mobile-Friendly Test
HTTPS and Security
HTTPS is a confirmed ranking factor. It protects data transmission between users and your server.
HTTPS implementation:
- Obtain SSL/TLS certificate (Let’s Encrypt provides free certificates that auto-renew, making HTTPS accessible for any site)
- Install certificate on server
- Redirect HTTP to HTTPS
- Update internal links to HTTPS
- Update sitemap and canonical tags
Security headers to implement:
- Strict-Transport-Security (HSTS)
- X-Content-Type-Options
- X-Frame-Options
- Content-Security-Policy
Site Architecture
Site architecture affects how search engines understand content relationships and how link equity flows through your site.
URL Structure Best Practices
Semantic URL hierarchy:
example.com/ (Root)
example.com/seo/ (Category/Seed)
example.com/seo/technical-seo/ (Subcategory)
example.com/seo/technical-seo/core-web-vitals/ (Topic)
URL rules:
- Use lowercase letters
- Separate words with hyphens
- Keep URLs short but descriptive
- Include primary keyword
- Avoid parameters when possible
Breadcrumb Navigation
Breadcrumbs show users and search engines where a page sits within your site hierarchy. They reduce pogo-sticking, improve crawl efficiency, and can appear directly in Google search results.
Keep navigation depth shallow. Most pages should be reachable within three clicks from the homepage. Deep pages (four or more levels) get crawled less frequently and accumulate less link equity.
Implement BreadcrumbList schema markup alongside visible breadcrumbs so Google can display the path in SERPs:
{
"@context": "https://schema.org",
"@type": "BreadcrumbList",
"itemListElement": [
{ "@type": "ListItem", "position": 1, "name": "Home", "item": "https://example.com/" },
{ "@type": "ListItem", "position": 2, "name": "SEO", "item": "https://example.com/seo/" },
{ "@type": "ListItem", "position": 3, "name": "Technical SEO" }
]
}
Internal Linking for Technical SEO
Internal links distribute page authority and establish topical relationships.
Internal linking strategy:
- Link from high-authority pages to important pages
- Use descriptive anchor text
- Create logical content hierarchies
- Update old content with links to new pages
Pagination Handling
Google deprecated rel="prev" and rel="next" in 2019. That means search engines no longer use those tags to understand paginated sequences. You still have options:
- View all page - Combine paginated content onto a single page when practical. Google prefers this.
- Load more / infinite scroll - Works well for users, but ensure each item has a unique crawlable URL. Use the History API so the URL updates as users scroll.
- Standard pagination - Use self-referencing canonical tags on each page. Don’t noindex paginated pages; they carry link equity.
For large product catalogs or article archives, combine pagination with a robust internal linking structure so crawlers reach deep pages efficiently.
Hreflang Implementation
If your site targets multiple languages or regions, hreflang tags tell search engines which version to show each audience. This prevents duplicate content issues across language variants.
Implementation methods:
- HTML
<link>tags in the<head>section - HTTP headers (useful for PDFs and non-HTML files)
- XML sitemap hreflang annotations
Common hreflang mistakes:
- Missing return tags (if page A references page B, page B must reference page A)
- Using incorrect language/region codes
- Pointing hreflang to non-canonical URLs
- Forgetting the
x-defaulttag for fallback
For a site targeting both English (en-MY) and Malay (ms-MY) audiences in Malaysia:
<link rel="alternate" hreflang="en-MY" href="https://example.com/en/page/" />
<link rel="alternate" hreflang="ms-MY" href="https://example.com/ms/page/" />
<link rel="alternate" hreflang="x-default" href="https://example.com/en/page/" />
Validate hreflang tags using Google Search Console’s International Targeting report or Screaming Frog’s hreflang validation feature.
Structured Data (Schema Markup)
Structured data helps search engines understand content meaning and can generate rich results.
Common Schema Types
| Schema Type | Use Case |
|---|---|
| Article | Blog posts, news articles |
| Organization | Company information |
| LocalBusiness | Local business details |
| Product | E-commerce products |
| FAQ | Frequently asked questions |
| HowTo | Step-by-step instructions |
| BreadcrumbList | Navigation path |
Schema Implementation
Use JSON-LD format (Google’s preferred method):
{
"@context": "https://schema.org",
"@type": "Article",
"headline": "Technical SEO Guide",
"author": {
"@type": "Organization",
"name": "Editorial Team"
},
"datePublished": "2024-02-15",
"publisher": {
"@type": "Organization",
"name": "Semantic.my"
}
}
JavaScript Rendering and SEO
Google renders JavaScript, but it does so in two phases. First, it crawls and indexes the raw HTML. Later, it sends the page to a rendering queue (Web Rendering Service) for JavaScript execution. This second phase can be delayed by hours or even days.
Why this matters for SEO:
- Content that depends on JS may not get indexed immediately
- Googlebot has a limited rendering budget, similar to crawl budget
- Some JavaScript frameworks produce empty HTML shells that search engines cannot parse without rendering
Best practices for JS-heavy sites:
- Use server-side rendering (SSR) or static site generation (SSG) for SEO-critical pages
- Provide meaningful HTML content before JavaScript executes
- Avoid loading primary content via lazy-loaded JS triggered only by scroll
- Test how Google sees your page using the URL Inspection tool in Search Console (click “Test Live URL” then “View Tested Page”)
CMS-Specific Technical SEO Considerations
WordPress: WordPress handles many technical SEO basics out of the box, but it still needs attention. Use a plugin like Yoast SEO or Rank Math to manage meta tags, sitemaps, and schema. Keep plugins lean because each one adds JS and CSS that slow the site down. Choose a well-coded theme, enable caching (WP Super Cache or WP Rocket), and use a CDN.
Shopify:
Shopify manages hosting, SSL, and mobile responsiveness automatically. However, it has limitations: you cannot fully edit robots.txt (though Shopify added partial control in 2021), URL structures follow a rigid pattern (/collections/, /products/), and duplicate content from tag pages needs careful canonical management. Use Shopify’s built-in sitemap and supplement with structured data for products and reviews.
Technical SEO Audit Checklist
Crawlability
- No critical pages blocked in robots.txt
- XML sitemap submitted and up-to-date
- No excessive redirect chains
- Crawl errors fixed in Search Console
- Orphan pages identified and linked
Indexability
- Important pages are indexable
- Canonical tags properly implemented
- No accidental noindex tags
- Duplicate content resolved
Performance
- Core Web Vitals passing (LCP, INP, CLS)
- Page load under 3 seconds
- Images optimized and compressed
Mobile
- Mobile-friendly test passing
- Content parity with desktop
- Touch targets appropriately sized
Security
- HTTPS enabled site-wide
- No mixed content warnings
- Security headers implemented
- SSL certificate valid
Structure
- Clean URL structure
- Proper heading hierarchy
- Schema markup implemented
- Internal linking optimized
- Hreflang tags validated (multilingual sites)
- Breadcrumbs with schema markup
Common Technical SEO Mistakes
- Blocking important resources - CSS/JS blocked in robots.txt prevents proper rendering
- Missing canonical tags - Causes duplicate content issues across paginated and filtered pages
- Slow server response - TTFB over 600ms hurts all other metrics
- Unoptimized images - Large images are often the single biggest drag on page speed
- Broken internal links - Waste crawl budget and frustrate users. Run Screaming Frog to find broken links and redirect chains.
- Missing mobile optimization - Critical under mobile-first indexing
- No HTTPS - Security and ranking impact; no excuse with free options like Let’s Encrypt
- Client-side rendering without fallback - JS-only content risks delayed or missed indexing
Technical SEO forms the foundation for all other SEO work. Content optimization and link building cannot deliver results if search engines cannot crawl, index, and render your pages properly. Start with a technical audit to identify issues, then address problems in order of impact. Regular monitoring through Google Search Console and PageSpeed Insights keeps your site healthy over time.
For sites targeting local customers, technical fundamentals like mobile optimization and page speed are especially critical - local search visibility depends on them since most local queries happen on mobile devices. And while technical SEO ensures your pages are discoverable, building external authority through backlinks and brand signals determines how competitively those pages rank.
Focus on Core Web Vitals, mobile optimization, and clean site architecture. These fundamentals support both search engine crawling and user experience, driving sustainable organic growth.
Baca dalam Bahasa Malaysia: SEO Teknikal | Panduan Pemula SEO