AI training data fight escalates as Studio Ghibli and Japanese publishers challenge OpenAI
Studio Ghibli and a coalition of major Japanese publishers, represented by the Content Overseas Distribution Association, demand that OpenAI stop training its models on their works. They argue that generative tools are learning and replicating protected styles without consent, and Japanese officials signal they are ready to enforce local copyright rules if complaints go unanswered.
Japan’s creative heavyweights are drawing a clear boundary. Studio Ghibli, known for films like Spirited Away and My Neighbor Totoro, has joined publishers including Bandai Namco and Square Enix to press OpenAI to halt the use of their content in model training. Their message, delivered through CODA in a formal letter, puts fresh global pressure on how tech companies source data for AI.
Why are Japanese creators pushing back now?
The trigger is simple to explain and hard to ignore. Popular generative systems can produce images and video that feel unmistakably “anime” or even “Ghibli-like.” That might delight fans. For rights holders, it looks a lot like unlicensed style replication at industrial scale. Millions of users are testing prompts that reference famous studios and characters, which blurs the line between homage and unauthorized copying. You can see how filmmakers are growing increasingly concerned as these results spread.
In its October 28 statement, CODA said tools such as the video generator often called Sora 2 can output visuals that closely resemble Japanese anime, games and manga. Their view is that such fidelity is unlikely without extensive exposure to those works during training. Studio Ghibli’s leadership has echoed the point, saying creators expect global tech firms to respect their rights, as reported by TechCrunch.

What is OpenAI’s position and how does the law apply?
OpenAI has discussed opt-out frameworks with some rights holders in the past. CODA argues that opt-out is not enough under Japanese law, which generally expects prior permission when copyrighted works are used beyond limited exceptions. This is a key difference from the United States, where fair use is broader and still being tested in court for AI training.
Japan’s legal backdrop is nuanced. The Copyright Act includes a data analysis exception that has been interpreted to allow text and data mining, but policymakers have been reassessing how far that should extend to generative AI. Government bodies have issued guidance for AI businesses and signaled that enforcement will follow if creators’ rights are ignored. Officials have publicly called anime and manga cultural treasures and warned that ongoing complaints will be taken seriously, as noted by Gamespot and other outlets.
Get the latest tech updates and insights directly in your inbox.
What does this clash mean for AI and copyright worldwide?
This is not a niche fight. It is a test of how the world treats the data that powers AI. Courts in multiple countries are still wrestling with whether training on copyrighted works without a license can be fair, transformative use, or whether it crosses into infringement. High profile cases in the United States, including actions from media companies and artists, highlight how unsettled the rules remain.
Legal scholars in Japan and abroad note that the question is bigger than images. Voice cloning, music generation and text models all raise similar issues. As Professor Hiroshi Hasegawa of Keio University has argued, the AI era is forcing a fresh look at how we define copying and derivation for machine learning, a point reflected in coverage by Storyboard18.
Recommended tech
AI features are quickly moving into everyday laptops. The TechBull recommends the latest Copilot+ PCs, like the Lenovo IdeaPad Slim 3X AI Laptop, which bring on-device AI acceleration to creative and productivity workflows.
What happens next for OpenAI and Japan’s creative sector?
CODA wants two things. First, a stop to training on members’ works without permission. Second, a good faith process to address claims about model outputs that mirror protected styles or characters. With ministries indicating they are monitoring the situation, the stakes are clear. If the parties cannot find a compromise, lawsuits, regulatory action or both could reshape how datasets are sourced and how generative outputs are controlled.

For now, the standoff holds. Japanese creators have drawn a line that is as much about cultural stewardship as it is about economics in a world where AI is changing our digital world. However this dispute resolves, it will influence the playbook for AI training and licensing far beyond Japan.
FAQs
What exactly did CODA request from OpenAI?
CODA asked OpenAI to stop using members’ works for training without permission and to respond promptly and sincerely to future rights claims related to model outputs that resemble protected content.
Does Japanese copyright law allow training on copyrighted data without a license?
Japan has a data analysis exception that has been interpreted to cover text and data mining. Rights holders argue that large scale training for generative models goes beyond that spirit and requires permission. Policymakers are reassessing how the exception applies to today’s AI.
How is this different from the United States?
The U.S. relies on fair use, which is flexible and decided case by case. Several lawsuits are testing how fair use applies to AI training. Japan’s approach has been more conservative, emphasizing authorization when copyrighted material is used beyond narrow exceptions.
What models or tools are in the spotlight?
Video and image generators are front and center, including tools often cited by creators for producing anime-like styles. Rights holders say these outputs suggest the models were trained on protected works without consent.
Could this lead to a licensing framework?
It might. Many in the industry expect more dataset transparency, opt-in licensing, and compensation schemes for creators, especially as cross-border legal risks grow for model providers.





[…] “This partnership provides greater recognition to the creators behind the music,” Poraz explained in an official WhoSampled blog statement. It’s a step toward giving credit where it’s due, which is crucial as AI’s role in content creation continues to be a hot topic, raising questions about data usage and artist rights. […]