Poisoning Well
Poisoning Well 31st March 2025 One of the many pressing issues with Large Language Models (LLMs) is they are trained on content that isn’t theirs to consume. Since most of what they consume is on the open web, it’s difficult for authors to withhold consent without also depriving legitimate agents (AKA humans or “meat bags”) of information. Some well-meaning but naive developers have implored authors to instate robots.txt rules, intended to block LLM-associated crawlers. User-agent: GPTBot D