Content Analysis
Automated content quality metrics — word count, readability, keyword extraction, and duplicate detection for every crawled page.
Content Analysis extracts quality metrics from every page the Edge Crawler processes. Word count, readability scores, keyword density, and content hashing are all computed automatically during crawling — no extra credits required.
Metrics Extracted
The following metrics are computed and stored for every crawled page:
| Metric | Description |
|---|---|
| Word Count | Total words in the main content area |
| Reading Time | Estimated reading time based on word count |
| Readability | Flesch-Kincaid grade level score |
| Content Hash | Unique hash of the page content for duplicate detection |
| Image Count | Number of images found on the page |
| Internal Links | Count of links pointing to other pages on the same domain |
| External Links | Count of links pointing to other domains |
| Schema Types | JSON-LD schema types detected (Product, Article, etc.) |
View these metrics on the Content tab of any URL Detail page.
Content Quality
Thin Content Detection
Pages with fewer than 100 words of main content are flagged as thin content. Thin pages can hurt your domain's overall SEO performance — search engines may consider them low-value or fail to rank them.
The Site-Wide SEO Dashboard shows a count of all thin content pages across your domain for easy triage.
Readability
The Flesch-Kincaid grade level indicates the education level needed to understand the content:
| Grade Level | Interpretation |
|---|---|
| 5–8 | Easy to read — suitable for a broad audience |
| 9–12 | Moderate — appropriate for most web content |
| 13+ | Difficult — consider simplifying for wider reach |
Lower readability scores generally correlate with better engagement, especially for consumer-facing content.
Keyword Extraction
SerpWise extracts top keywords from each page using TF-IDF (Term Frequency–Inverse Document Frequency) analysis across your domain. This surfaces the most distinctive terms on each page relative to your other content.
- Per-page keywords — the most prominent terms on a specific page
- Domain-wide analysis — keywords are weighted against your full page index, highlighting terms that are uniquely important to each page rather than common site-wide terms
Use keyword data to verify pages are targeting intended topics and to spot content gaps.
Duplicate Detection
Each page's content is hashed during crawling. When multiple pages share the same content hash, they're flagged as potential duplicates.
The Content tab on the URL Detail page shows:
- The content hash for the current page
- Any other pages on the same domain with a matching hash
Duplicate content can dilute search rankings. Use canonical tags or redirects to consolidate duplicate pages.
How Data Is Collected
Content metrics are extracted automatically during every crawl — there's no separate process to enable or configure. When the Edge Crawler indexes a page, the audit engine parses the HTML, extracts content, and computes all metrics in a single pass.
Content analysis is included with every crawl at no additional credit cost. Metrics are updated each time a page is re-crawled.
Fix with AI
One-click fixes for SEO audit issues and AI readiness recommendations — automatically generate rules, meta tags, and structured data without manual configuration.
Per-URL Analytics
Deep traffic analytics for individual URLs — page views, bot activity, response times, referrers, and rule trigger data.