Bots now outnumber humans on the web, causing problems like competitive price scraping and unauthorized AI training data collection. A solution combines bot detection services with edge-based WebAssembly functions that act as rewriting proxies. The proxy intercepts bot traffic, uses CSS selectors to identify content sections, and dynamically replaces or redacts sensitive information like pricing. For AI bots, the system can use LLMs to generate summaries instead of exposing original content, allowing sites to maintain search authority while protecting intellectual property. The redacted content is cached at the edge to reduce upstream load.
Table of contents
A Sampling of The Bot ProblemsIntercepting Bots at the EdgeA Rewriting ProxyAI versus AIConclusionSort: