Anthropic Accuses 3 Chinese Companies of Harvesting Its Data


Anthropic Accuses 3 Chinese Companies of Harvesting Its Data
Anthropic’s terms forbid anyone to harvest data for distillation. Separately, over not sharing its technology with Pentagon, the company’s CEO Dario Amodei (inset) has made clear his ethical concerns about unchecked govt use of AI

SAN FRANCISCO: The San Francisco artificial intelligence startup Anthropic has accused three Chinese companies of improperly harvesting large amounts of data from its AI technologies in an effort to accelerate the development of their own systems. Anthropic said in a blog post Monday that DeepSeek, Moonshot and MiniMax — three prominent Chinese startups — used about 24,000 fraudulent accounts to generate more than 16 million conversations with its Claude chatbot that could be used to teach skills to their own chatbots. Using data from one AI system to train another — a process called distillation — is common in AI work. But Anthropic’s terms of service forbid anyone to surreptitiously harvest data for distillation and do not allow its technologies to be used in China. OpenAI, Anthropic’s primary rival, has also accused Chinese companies of lifting large amounts of data from its chatbot, ChatGPT, for similar proposes. In a memo sent to the House Select Committee on China last week, OpenAI said that DeepSeek and other Chinese startups were using new and “obfuscated” distillation methods as part of their “ongoing efforts to free-ride” on technologies developed by OpenAI and other U.S. companies. Like OpenAI, Anthropic said the practice was a national security risk, adding that it could allow China to build AI technologies to create bioweapons or tools for mass surveillance. The startup has guardrails on its technologies designed to prevent them from being used in those ways, but the guardrails can be stripped away during distillation. Anthropic called on government officials and other AI companies to help prevent Chinese companies from distilling American models. “These campaigns are growing in intensity and sophistication,” Anthropic said in its post. “The window to act is narrow, and the threat extends beyond any single company or region. Addressing it will require rapid, coordinated action among industry players, policymakers and the global AI community.” DeepSeek, Moonshot and MiniMax did not immediately respond to requests for comment. Anthropic published its post amid a tussle with the Defense Department over the Pentagon’s use of its technologies. The Pentagon has approved Anthropic’s technologies for use with classified tasks, but it is threatening to sever ties with the startup because Anthropic does not want its technologies used in situations involving autonomous weapons or domestic surveillance. Last year, DeepSeek spooked Silicon Valley tech companies and sent the U.S. financial markets into a tailspin after releasing AI technologies that matched the performance of anything else on the market. Until then, the prevailing wisdom in Silicon Valley had been that the most powerful systems could not be built without billions of dollars in specialized computer chips. But DeepSeek said it had created its technologies using far fewer resources. Like U.S. companies, DeepSeek, Moonshot and MiniMax build their AI technologies using computer code and data corralled from across the internet. AI companies around the world lean heavily on a practice called open sourcing, which means they freely share the code that underpins their technologies and reuse code shared by others. They see this as a way of accelerating technological development. AI companies also need enormous amounts of online data to train their AI systems. The leading systems learn their skills by analyzing just about all of the text on the internet. Distillation is often used to train new systems. This is often allowed by open source technologies. But if a company takes data from proprietary technology, the practice may be legally problematic. Anthropic, which is now valued at $380 billion, is facing multiple lawsuits accusing it of illegally using copyrighted internet data to train its systems. In September as part of a landmark legal settlement, Anthropic agreed to pay $1.5 billion to a group of authors and publishers after a judge ruled it had illegally downloaded and stored millions of copyrighted books. It was the largest payout in the history of U.S. copyright cases. OpenAI and other AI companies face similar suits, including a lawsuit brought by The New York Times against OpenAI and its partner Microsoft. That suit contends that millions of articles published by the Times were used to train automated chatbots that now compete with the news outlet as a source of reliable information. Both OpenAI and Microsoft deny the claims. This article originally appeared in The New York Times.

  • Related Posts

    US moves closer to tougher H-1B wage norms as proposal clears federal review

    The US has moved a step closer to revising wage rules for foreign workers after a proposed regulation affecting the H-1B visa and employment-based green card process cleared federal review…

    Continue reading
    State of the Union 2026: When and where to watch Donald Trump’s speech to Congress — details

    Donald Trump was set to deliver the State of the Union in Washington, his second major address to Congress this term and the last before the 2026 midterms. It was…

    Continue reading

    Leave a Reply

    Your email address will not be published. Required fields are marked *