Microsoft deletes blog telling users to train AI on pirated Harry Potter books

Ars Technica
by Ashley Belanger
February 20, 2026
AI-Generated Deep Dive Summary
Microsoft has removed a blog post that critics argued encouraged developers to use pirated copies of *Harry Potter* books to train AI models, following significant backlash in online forums like Hacker News. The now-deleted article, authored by Pooja Kamath, a senior product manager at Microsoft with over a decade of experience at the company, aimed to showcase how to integrate generative AI into applications using Azure SQL DB, LangChain, and large language models (LLMs). In an effort to make the example relatable, Kamath suggested using a popular dataset like *Harry Potter* books to demonstrate the tool's capabilities. The blog post, which was published in November 2024, sparked controversy because it appeared to endorse the use of pirated material for AI training. Critics argued that this not only encouraged illegal activity but also set a concerning precedent for how tech companies approach data sourcing and ethical considerations in AI development. Microsoft's decision to remove the post came after facing criticism on social media platforms, where users questioned the company's promotion of such practices. This incident highlights the growing tension around ethical AI development and corporate responsibility. Tech enthusiasts and developers are increasingly scrutinizing how major companies like Microsoft acquire and use data, particularly when it comes to training AI models. The controversy underscores the importance of transparency and accountability in the tech industry, especially as AI becomes more integrated into everyday applications. As companies continue to push the boundaries of generative AI, they must navigate complex ethical landscapes to maintain trust with their audiences and stakeholders. The backlash also reflects broader concerns about the potential misuse of AI technology. By promoting tools that could be misused for unethical purposes, tech firms risk damaging their reputation and alienating users who prioritize ethical practices.
Verticals
techscience
Originally published on Ars Technica on 2/20/2026