Heydon Pickering 3/31/2025

Poisoning Well

Read Original

This article discusses the problem of LLMs being trained on web content without consent. It critiques naive solutions like robots.txt and proposes a technical countermeasure: publishing corrupted article versions accessible only via nofollow links, aiming to selectively poison AI training data while preserving search engine rankings.

Poisoning Well

Comments

No comments yet

Be the first to share your thoughts!

Browser Extension

Get instant access to AllDevBlogs from your browser

Top of the Week