CrawlProof
View raw .md

AEO Audit for kuubi.fi

Target: https://www.kuubi.fi/
Score: 43 / 100
Generated: 2026-05-13T06:55:43.036Z
Pages crawled: 9
Findings: 20 pass · 58 warn · 3 fail · 0 unknown


1. Crawl Summary

2. Data Found

Data PointFound?SourceNotes
PricingNo
Customer logosNo
Social proofNo
Recent launchesNo
Blog post activityNo
New hiresNoOften only on a /blog/team or LinkedIn page
Headline copyYesHomepageCreativity loves technology
PositioningNo
Executive teamNo
Product/service descriptionsYesHomepageFrom meta description
Case studies or testimonialsNo
Contact/demo/signup pathsYesNavigation links

3. Homepage Audit

  • ⚠️ Missing canonical link Add <link rel="canonical" href="https://your-domain"> to prevent dup-content confusion.
  • ⚠️ Twitter Card: missing card Add twitter:card, twitter:title, twitter:description, twitter:image for richer previews in social and AI agent surfaces.
  • Alt text coverage: 20% 8/40 images have alt text.
  • Homepage fetched successfully HTTP 200 · 242349 bytes · 849ms
  • Page load time: 0.85s Fast — well within AI crawler budgets.
  • declared
  • Single H1 Creativity loves technology
  • <title> present (53 chars)
  • Meta description present (120 chars)
  • Open Graph tags complete
  • Critical content is server-rendered Raw and rendered text are within 1% of each other.
  • Content volume: 1028 words Substantive content — AI models have enough to summarize and recommend.
  • Heading structure: 34 (h1:1, h2:5, h3:28) Multiple headings help AI chunk and outline your page.
  • Internal links: 44 44 internal + 8 external links help crawlers navigate.

4. Schema / Structured Data Audit

  • No JSON-LD structured data found Add JSON-LD blocks (Organization, SoftwareApplication, FAQPage, BreadcrumbList) so AI answer engines can ingest your data without guessing.

5. robots.txt and sitemap.xml Audit

  • ⚠️ robots.txt does not reference a Sitemap Add Sitemap: https://yoursite.com/sitemap.xml to robots.txt.
  • robots.txt present 73 chars
  • sitemap.xml present (111 URLs)

6. LLM / AI Crawler Accessibility

  • ⚠️ llms.txt missing Add /llms.txt — a concise, link-rich summary that helps LLMs orient on your site.
  • ⚠️ GPTBot not explicitly addressed No User-agent: GPTBot block in robots.txt. We recommend explicit Allow rules so crawlers don't fall back to defaults.
  • ⚠️ ClaudeBot not explicitly addressed No User-agent: ClaudeBot block in robots.txt. We recommend explicit Allow rules so crawlers don't fall back to defaults.
  • ⚠️ PerplexityBot not explicitly addressed No User-agent: PerplexityBot block in robots.txt. We recommend explicit Allow rules so crawlers don't fall back to defaults.
  • ⚠️ Google-Extended not explicitly addressed No User-agent: Google-Extended block in robots.txt. We recommend explicit Allow rules so crawlers don't fall back to defaults.
  • ⚠️ OAI-SearchBot not explicitly addressed No User-agent: OAI-SearchBot block in robots.txt. We recommend explicit Allow rules so crawlers don't fall back to defaults.
  • ⚠️ Applebot-Extended not explicitly addressed No User-agent: Applebot-Extended block in robots.txt. We recommend explicit Allow rules so crawlers don't fall back to defaults.
  • ⚠️ CCBot not explicitly addressed No User-agent: CCBot block in robots.txt. We recommend explicit Allow rules so crawlers don't fall back to defaults.
  • ⚠️ skill.md missing Add /skill.md describing what your site lets agents do — speeds up agent task routing.
  • ⚠️ /.well-known/security.txt missing Publish a /.well-known/security.txt with at least a Contact: line. Crawlers and security researchers expect it; AI systems use it as a trust signal.

7. Positioning Clarity

  • ⚠️ No pricing/plans link found AI summaries commonly include pricing. Add a /pricing page even if pricing is custom.
  • ⚠️ Value-prop language not detected Pages with phrases like 'we help X', 'platform for Y', 'built for Z' are easier for LLMs to summarize.
  • About/Team path discoverable
  • H1 communicates value Creativity loves technology
  • Contact / signup path discoverable

8. Missing or Hard-to-Find Information

  • 9 data point(s) could not be found from public pages · Pricing · Customer logos · Social proof · Recent launches · Blog post activity · New hires · Positioning · Executive team · Case studies or testimonials
  • ⚠️ Add JSON-LD structured data Start with Organization on the root layout and SoftwareApplication or Product on /pricing. Add FAQPage on any FAQ section.

  • ⚠️ Add /llms.txt A short Markdown-flavored summary at the root. Include your H1, value prop, top 5–10 links, and pricing summary.

  • ⚠️ Add a /pricing page Even contact-us pricing benefits from a /pricing page that LLMs can link to in answers.

  • ⚠️ Add a canonical link Prevents dup-content drift and tells AI crawlers which URL is authoritative.

    <link rel="canonical" href="https://yoursite.com/" />
    
  • ⚠️ Add Twitter Card meta tags Used by social platforms and AI agents for richer previews.

    <meta name="twitter:card" content="summary_large_image" />
    <meta name="twitter:title" content="Your Page Title" />
    <meta name="twitter:description" content="50–160 char description." />
    <meta name="twitter:image" content="https://yoursite.com/og-image.jpg" />
    
  • ⚠️ Add alt text to all meaningful images Decorative-only images can use empty alt='', but logos, screenshots, and product images need descriptive alt.

  • ⚠️ Allow GPTBot in robots.txt Add an explicit User-agent: GPTBot Allow: / block so this AI crawler can read your site.

  • ⚠️ Allow ClaudeBot in robots.txt Add an explicit User-agent: ClaudeBot Allow: / block so this AI crawler can read your site.

  • ⚠️ Allow PerplexityBot in robots.txt Add an explicit User-agent: PerplexityBot Allow: / block so this AI crawler can read your site.

  • ⚠️ Allow Google-Extended in robots.txt Add an explicit User-agent: Google-Extended Allow: / block so this AI crawler can read your site.

  • ⚠️ Allow OAI-SearchBot in robots.txt Add an explicit User-agent: OAI-SearchBot Allow: / block so this AI crawler can read your site.

  • ⚠️ Allow Applebot-Extended in robots.txt Add an explicit User-agent: Applebot-Extended Allow: / block so this AI crawler can read your site.

  • ⚠️ Allow CCBot in robots.txt Add an explicit User-agent: CCBot Allow: / block so this AI crawler can read your site.

  • ⚠️ Add /skill.md Describe what an agent can do with your site (e.g., 'Search docs', 'Look up pricing'). Useful for agentic flows.

  • ⚠️ Publish /.well-known/security.txt A security contact builds trust with crawlers and researchers. Minimal example:

    Contact: mailto:security@yourdomain.com
    Expires: 2027-01-01T00:00:00.000Z
    Preferred-Languages: en
    
  • ⚠️ State your audience explicitly Use phrases like 'Built for B2B SaaS marketing teams' on the homepage and About page.

  • ⚠️ Reference your sitemap in robots.txt Add Sitemap: https://yoursite.com/sitemap.xml so crawlers don't have to guess.

10. Priority To-Do List

  • P1 — Add JSON-LD structured data Start with Organization on the root layout and SoftwareApplication or Product on /pricing. Add FAQPage on any FAQ section.

  • P2 — Add /llms.txt A short Markdown-flavored summary at the root. Include your H1, value prop, top 5–10 links, and pricing summary.

  • P2 — Add a /pricing page Even contact-us pricing benefits from a /pricing page that LLMs can link to in answers.

  • P3 — Add a canonical link Prevents dup-content drift and tells AI crawlers which URL is authoritative.

    ```html
    <link rel="canonical" href="https://yoursite.com/" />
    ```
    
  • P3 — Add Twitter Card meta tags Used by social platforms and AI agents for richer previews.

    ```html
    <meta name="twitter:card" content="summary_large_image" />
    <meta name="twitter:title" content="Your Page Title" />
    <meta name="twitter:description" content="50–160 char description." />
    <meta name="twitter:image" content="https://yoursite.com/og-image.jpg" />
    ```
    
  • P3 — Add alt text to all meaningful images Decorative-only images can use empty alt='', but logos, screenshots, and product images need descriptive alt.

  • P3 — Allow GPTBot in robots.txt Add an explicit User-agent: GPTBot Allow: / block so this AI crawler can read your site.

  • P3 — Allow ClaudeBot in robots.txt Add an explicit User-agent: ClaudeBot Allow: / block so this AI crawler can read your site.

  • P3 — Allow PerplexityBot in robots.txt Add an explicit User-agent: PerplexityBot Allow: / block so this AI crawler can read your site.

  • P3 — Allow Google-Extended in robots.txt Add an explicit User-agent: Google-Extended Allow: / block so this AI crawler can read your site.

  • P3 — Allow OAI-SearchBot in robots.txt Add an explicit User-agent: OAI-SearchBot Allow: / block so this AI crawler can read your site.

  • P3 — Allow Applebot-Extended in robots.txt Add an explicit User-agent: Applebot-Extended Allow: / block so this AI crawler can read your site.

  • P3 — Allow CCBot in robots.txt Add an explicit User-agent: CCBot Allow: / block so this AI crawler can read your site.

  • P3 — Add /skill.md Describe what an agent can do with your site (e.g., 'Search docs', 'Look up pricing'). Useful for agentic flows.

  • P3 — Publish /.well-known/security.txt A security contact builds trust with crawlers and researchers. Minimal example:

    ```
    Contact: mailto:security@yourdomain.com
    Expires: 2027-01-01T00:00:00.000Z
    Preferred-Languages: en
    ```
    
  • P3 — State your audience explicitly Use phrases like 'Built for B2B SaaS marketing teams' on the homepage and About page.

  • P3 — Reference your sitemap in robots.txt Add Sitemap: https://yoursite.com/sitemap.xml so crawlers don't have to guess.


Report by CrawlProof. Reusable after every major website change.