Reach security professionals who buy.

850K+ monthly readers 72% have budget authority
Advertise on SecurityCareers.help β†’

If you run web properties for your organization β€” whether a corporate blog, a product marketing site, or a security awareness portal β€” you’re operating in a threat environment that looks nothing like it did three years ago. You’re not just defending against human attackers anymore. You’re defending against automated bot farms draining ad revenue, AI scrapers harvesting your content without attribution, and a CDN ecosystem that can be your best friend or your worst bottleneck depending on how you configure it.

This guide is written for CISOs, CCOs, and security architects who need to understand the full stack β€” and explain it clearly to the marketing teams who live and die by traffic numbers, ad RPMs, and search rankings.

We learned most of this the hard way.


The Bot Farm Problem Is a Revenue Problem (Marketing Will Care About This)

In early 2026, the CISO Marketplace network β€” which runs 11 cybersecurity content sites β€” got hit by a large-scale bot campaign originating from Lanzhou, China and several Singapore-based ASNs. The bots weren’t trying to breach anything. They were inflating page view metrics, which triggered invalid traffic detection at our ad network partner. The result: immediate removal from the ad monetization program.

This is the attack vector marketing doesn’t see coming. Nobody filed a CVE. No data was exfiltrated. But the business impact was immediate and measurable β€” ad revenue went to zero while we investigated.

We documented the full technical response, including Cloudflare block rules and ASN filtering, in our sister site: We Got Hit by the Mysterious Lanzhou Bots β€” Here’s Everything You Need to Fight Back

The lesson for security architects: bot mitigation is not just a DDoS problem. It’s an analytics integrity problem, an ad fraud problem, and increasingly an AI training data problem.


The Architecture: Where Traffic Gets Clean or Dirty

Understanding where in the stack traffic is validated is the foundation of any web security architecture conversation. Here’s how a well-configured modern stack works:

INTERNET
    β”‚
    β”œβ”€β”€ Human visitors (clean)
    β”œβ”€β”€ Bot farms (dirty β€” inflate metrics, drain ad revenue)
    β”œβ”€β”€ AI scrapers (gray area β€” harvest content for LLM training)
    └── Legitimate crawlers (Google, Bing, etc.)
    β”‚
    β–Ό
β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”
β”‚       CLOUDFLARE EDGE (WAF)        β”‚
β”‚  β€’ Bot detection (ML-based scoring) β”‚
β”‚  β€’ Geo-blocking / ASN blocking      β”‚
β”‚  β€’ Rate limiting                    β”‚
β”‚  β€’ Challenge pages (JS/CAPTCHA)     β”‚
β”‚  β€’ DDoS absorption                  β”‚
β”‚  β†’ Bad actors blocked/challenged    β”‚
β”‚  β†’ Clean traffic passes through     β”‚
β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜
    β”‚
    β–Ό
β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”
β”‚       CDN EDGE CACHE                β”‚
β”‚  β€’ Static assets served from edge   β”‚
β”‚  β€’ Cache-hit ratio: target 80%+     β”‚
β”‚  β€’ Cache-miss β†’ origin fetch        β”‚
β”‚  β€’ Dramatically reduces origin load β”‚
β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜
    β”‚
    β–Ό
β”Œβ”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”
β”‚       ORIGIN SERVER                 β”‚
β”‚  β€’ Only sees clean cache-miss trafficβ”‚
β”‚  β€’ Static site = minimal attack     β”‚
β”‚    surface (no DB, no server logic) β”‚
β””β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”€β”˜

When you deploy static Astro sites on Cloudflare Pages with Cloudflare proxy enabled, the cache-hit ratio on a well-optimized site runs 85–90%. That means 85–90% of all requests never reach your origin β€” they’re served from Cloudflare’s edge in milliseconds, and the origin server is effectively invisible to the internet.

For the CISO explaining this to a VP of Marketing: β€œCloudflare is the bouncer. The CDN cache is the coatroom. The origin server is the VIP room β€” almost nobody gets there.”


Why Hosting Provider Matters: The Netlify-Cloudflare Proxy Gap

Here’s a subtlety that burned us and that security architects should be aware of: not all hosting providers play nicely with Cloudflare proxy.

When your domain is Cloudflare-proxied (orange cloud in DNS), traffic flows:

Visitor β†’ Cloudflare edge β†’ Hosting provider origin

Some hosting providers β€” Netlify in particular β€” have their own Cloudflare accounts and their own edge network. When Cloudflare (your account) tries to proxy to Netlify’s origin, you’re actually routing through two separate Cloudflare accounts with conflicting configurations. This creates:

  • SSL certificate validation failures
  • Incorrect origin IP exposure in some edge cases
  • Double-CDN inefficiency (caching at two layers with different TTL configs)
  • Build/deploy count consumption at the hosting level even when Cloudflare blocks the traffic before it should trigger a redeploy

The fix: host on Cloudflare Pages natively. Traffic never leaves Cloudflare’s network. Your origin is a Cloudflare Pages worker. The WAF, CDN cache, and origin are all in the same infrastructure. There’s no proxy-to-proxy conflict.

After migrating 11 sites from Netlify to Cloudflare Pages in March 2026, cache hit rates immediately improved and invalid traffic from China/Singapore stopped hitting origin entirely β€” it was absorbed at the WAF layer before touching the builds.


The AI Scraper Problem Is Different From Bot Farms

Bot farms want your traffic metrics. AI scrapers want your content.

In 2025–2026, every major AI lab (OpenAI, Anthropic, Google, Meta, Mistral) and dozens of smaller players operate web crawlers that harvest content at scale for LLM training. They look like legitimate search crawlers but they’re not providing backlinks, referral traffic, or attribution in return. They consume bandwidth and origin compute β€” and they may publish your content verbatim in AI responses, eliminating the reader’s need to visit your site.

The bandwidth math matters for marketing: If an AI scraper hits 10,000 pages on your site at 50KB each, that’s 500MB of egress. At scale across thousands of sites, this is a meaningful cost β€” and on origin-based hosting, it shows up in your bill.

Mitigating AI scrapers:

  1. robots.txt β€” The traditional approach. Most reputable AI crawlers (GPTBot, ClaudeBot, PerplexityBot) respect robots.txt. Add:

    User-agent: GPTBot
    Disallow: /
    
    User-agent: ClaudeBot
    Disallow: /
    
    User-agent: PerplexityBot
    Disallow: /
    

    Be selective β€” you may want some AI engines to index your content for GEO purposes (see below).

  2. ai.txt β€” An emerging standard (inspired by robots.txt) that explicitly defines which AI systems may use your content and under what terms. Not yet universally adopted, but worth implementing now.

  3. Cloudflare AI Audit β€” Cloudflare’s bot management dashboard now categorizes AI crawlers separately from traditional bots. You can block specific AI ASNs at the WAF level without touching robots.txt.

  4. Rate limiting at the WAF layer β€” Even if you allow AI crawlers, rate limiting their request frequency to 1 request/second protects your origin while still permitting indexing.


GEO: Teaching AI Agents to Find and Cite You

Here’s the architectural consideration nobody’s talking to marketing about yet: Generative Engine Optimization (GEO).

When a security professional asks ChatGPT, Perplexity, or Claude β€œwhat’s the best resource for understanding NIST CSF 2.0?” β€” the AI doesn’t use PageRank. It uses what was in its training data and what it can retrieve in real-time via search integration. If your site isn’t structured for AI comprehension, it doesn’t matter how well you rank on Google.

Structural changes that improve GEO:

1. llms.txt β€” The New robots.txt for AI

Place a file at yourdomain.com/llms.txt that tells AI systems what your site is, what it covers, and your key content. Example:

# Security Careers Help
> Career guidance for cybersecurity professionals β€” job search, certifications, salary negotiation, career progression from analyst to CISO.

## Core Content
- [Cybersecurity Career Paths](/career-paths/) β€” From SOC analyst to CISO
- [Certification Guide](/certifications/) β€” CISSP, CISM, CEH, CompTIA
- [Salary Data](/salary/) β€” By role, region, and experience
- [Interview Prep](/interviews/) β€” Technical and behavioral

## About
Published by CISO Marketplace. Audience: security professionals at all career stages.

AI agents that process llms.txt can accurately answer β€œwhat does securitycareers.help cover?” rather than guessing from sparse training data.

2. Schema Markup for Entity Clarity

AI systems use structured data to understand who you are. Add Organization and WebSite schema to your site’s <head>:

{
  "@context": "https://schema.org",
  "@type": "Organization",
  "name": "Security Careers Help",
  "url": "https://securitycareers.help",
  "description": "Career guidance for cybersecurity professionals",
  "sameAs": [
    "https://www.linkedin.com/company/...",
    "https://twitter.com/..."
  ]
}

The sameAs array creates entity disambiguation β€” it tells AI systems that your website, LinkedIn page, and Twitter account are all the same entity.

3. FAQ and Structured Long-Form Content

AI models extract information from well-structured content. Articles with clear H2/H3 hierarchies, explicit FAQ sections (<details>/<summary> or FAQ schema), and definitive statements perform better in AI-generated responses than loosely structured prose.

4. Internal Linking as Knowledge Graph Edges

From an AI comprehension standpoint, your internal link structure is your site’s knowledge graph. Pages that link to each other are understood as related concepts. For security career sites, this means: every article about β€œCISSP certification” should link to your β€œCISO career path” article, your β€œsalary guide,” and your β€œstudy resources” page. This creates a coherent topic cluster that AI systems can traverse and represent accurately.


The CISO Playbook: Explaining This to Marketing

Here’s the one-page translation for your CMO or VP of Marketing:

What Cloudflare does for your site:

  • Absorbs bot traffic before it touches your servers (protects ad revenue from invalid traffic)
  • Serves 85%+ of pages from cache (faster load times = better SEO = more conversions)
  • Blocks country/ASN-level attacks (China/Singapore bot farms never reach your analytics)
  • Provides unified traffic analytics that separate human vs. bot vs. crawler traffic

What AI scrapers mean for your content strategy:

  • Your content may appear in ChatGPT/Perplexity answers without a backlink or visit β€” this is traffic you’re β€œgiving away”
  • You can block AI scrapers entirely, or selectively allow the ones that drive referral traffic (Perplexity does send traffic)
  • Optimizing for AI-generated answers (GEO) is now as important as traditional SEO

What static site architecture means for security:

  • No database = no SQL injection surface
  • No server-side code = no RCE surface
  • Everything served from CDN edge = no origin IP exposure
  • Compromise scenario is limited to content manipulation during build, not runtime

What this means for budget:

  • Cloudflare Pages: free tier handles most content sites entirely
  • Eliminated Netlify build credits (~$50–200/month depending on tier and build frequency)
  • Egress costs effectively zero (Cloudflare Pages doesn’t charge for bandwidth)
  • Bot-driven invalid traffic eliminated = ad revenue protection

Architecture Checklist for Security-Minded Web Teams

Before your next site launch or infrastructure review, validate these:

  • All domains Cloudflare-proxied (orange cloud, not grey) β€” confirms WAF is active
  • Bot Fight Mode enabled in Cloudflare dashboard
  • Country/ASN block rules configured for known bad actors (adjust based on your analytics)
  • Rate limiting rules β€” cap requests per IP per minute at a threshold your real users won’t hit
  • Cache-hit ratio β‰₯ 80% in Cloudflare analytics β€” if lower, review cache TTL settings
  • robots.txt updated for AI crawlers β€” decide allow/deny per crawler
  • llms.txt implemented and accessible
  • Organization schema in <head> with sameAs entity links
  • No sensitive infrastructure on the same domain as public content (separate admin, API, and content domains)
  • Cloudflare AI Audit dashboard reviewed monthly β€” identify new AI crawlers as they emerge

Conclusion

The web security architecture conversation has expanded significantly in 2026. It’s no longer just about blocking attackers β€” it’s about protecting ad revenue from bot fraud, managing the AI scraper economy, and structuring your content so AI agents understand and accurately represent who you are.

For security professionals advising marketing teams: the good news is that most of this is Cloudflare configuration and content structure β€” not expensive tooling. The bad news is it requires ongoing attention as the bot landscape and AI crawling ecosystem evolve.

The organizations that get ahead of this will have cleaner analytics, better ad monetization, lower hosting costs, and stronger AI search presence. That’s a compelling business case that translates directly from the CISO’s office to the boardroom.


Related reading: If you want to see the raw technical battle against a bot farm in real time, read We Got Hit by the Mysterious Lanzhou Bots on Breached.Company β€” that’s where this architecture was stress-tested.