# robots.txt for AntonCV # Generated: 2025-11-30T22:20:14.358Z # Allow all legitimate crawlers User-agent: * Allow: / # Block specific paths if needed (uncomment as needed) # Disallow: /api/ # Disallow: /admin/ # Disallow: /_astro/ # Sitemap location Sitemap: https://antoncv.com/sitemap.xml # Common crawlers - explicit rules User-agent: Googlebot Allow: / User-agent: Googlebot-Image Allow: / User-agent: Bingbot Allow: / User-agent: Slurp Allow: / User-agent: DuckDuckBot Allow: / User-agent: Baiduspider Allow: / User-agent: YandexBot Allow: / # Allow SEO tools User-agent: AhrefsBot Allow: / User-agent: SemrushBot Allow: / # Allow AI assistants and answer engines User-agent: ChatGPT-User Allow: / User-agent: anthropic-ai Allow: / User-agent: Claude-Web Allow: / User-agent: PerplexityBot Allow: / User-agent: Perplexity-Bot Allow: / User-agent: Grok-bot Allow: / User-agent: Applebot-Extended Allow: / # Block AI scrapers and training bots # These bots collect data for AI model training User-agent: GPTBot Disallow: / User-agent: CCBot Disallow: / User-agent: cohere-ai Disallow: / User-agent: Omgilibot Disallow: / User-agent: FacebookBot Disallow: / User-agent: Diffbot Disallow: / User-agent: Bytespider Disallow: / User-agent: ImagesiftBot Disallow: / # Block bad bots and scrapers User-agent: MJ12bot Disallow: / User-agent: DotBot Disallow: / User-agent: BLEXBot Disallow: / User-agent: PetalBot Disallow: / # Crawl delay for polite bots (in seconds) # Uncomment if you want to slow down aggressive crawlers # Crawl-delay: 10