Microsoft’s GitHub Copilot sued over “software piracy on an unprecedented scale”
The lawsuit stated that this is the first class-action case in the US challenging the training and output of AI systems


Microsoft’s GitHub Copilot is being sued in a class action lawsuit that claims the artificial intelligence product is committing software piracy on an unprecedented scale.
The case was launched on 3 November by Matthew Butterick, a designer and programmer, along with the Joseph Saveri Law Firm to investigate GitHub Copilot. The team has filed a class action lawsuit in the San Francisco federal court on behalf of potentially millions of GitHub users.
The lawsuit seeks to challenge the legality of GitHub Copilot, as well as OpenAI Codex which powers the AI tool, and has been filed against GitHub, its owner Microsoft, and OpenAI.
GitHub and OpenAI launched Copilot in June 2021, an AI-based product that aims to help software coders by providing or filling in blocks of code using smart suggestions. It charges users $10 per month or $100 a year for its service.
“By training their AI systems on public GitHub repositories (though based on their public statements, possibly much more), we contend that the defendants have violated the legal rights of a vast number of creators who posted code or other work under certain open-source licences on GitHub,” said Butterick.
These licences include a set of 11 popular open source licences that all require attribution of the author’s name and copyright. This includes the MIT licence, the GNU General Public Licence, and the Apache licence.
The case claimed that Copilot violates and removes these licences offered by thousands, possibly millions, of software developers, and is therefore committing software piracy on an unprecedented scale.
Get the ITPro daily newsletter
Sign up today and you will receive a free copy of our Future Focus 2025 report - the leading guidance on AI, cybersecurity and other IT challenges as per 700+ senior executives
RELATED RESOURCE
Big payoffs from big bets in AI-powered automation
Automation disruptors realise 1.5 x higher revenue growth
Copilot, which is entirely run on Microsoft Azure, often simply reproduces code that can be traced back to open-source repositories or licensees, according to the lawsuit. The code never contains attributions to the underlying authors, which is in violation of the licences.
“It is not fair, permitted, or justified. On the contrary, Copilot’s goal is to replace a huge swath of open source by taking it and keeping it inside a GitHub-controlled paywall. It violates the licences that open-source programmers chose and monetises their code despite GitHub’s pledge never to do so,” detailed the class-action complaint.
Moreover, the case stated that the defendants have also violated GitHub’s own terms of service and privacy policies, the DMCA code 1202 which forbids the removal of copyright-management information, and the California Consumer Privacy Act.
“As far as we know, this is the first class-action case in the US challenging the training and output of AI systems,” said Butterick. “It will not be the last. AI systems are not exempt from the law. Those who create and operate these systems must remain accountable. If companies like Microsoft, GitHub, and OpenAI choose to disregard the law, they should not expect that we the public will sit still.
“AI needs to be fair and ethical for everyone. If it’s not, then it can never achieve its vaunted aims of elevating humanity. It will just become another way for the privileged few to profit from the work of the many,” he added.
When asked for comment, GitHub highlighted that it had announced on 1 November that it’s set to bring in new features to the Copilot platform in 2023.
Whenever the tool suggests a code fragment, it’s hoping to provide developers with an inventory of similar code found in GitHub public repositories as well as the ability to organise the inventory by filters like the commit date, repository licence, and more.
IT Pro has contacted Microsoft and OpenAI for further comment.
In October 2022, developer Tim Davis, professor of computer science at Texas A&M University, wrote on Twitter that GitHub Copilot had emitted large chunks of his copyrighted code, with no attribution to him.
Davis added that he could probably reproduce his entire sparse matrix libraries from simple prompts, aiming to underline the similarity between his work and what the AI tool produced.
“The code in question is different from the example given. Similar, but different. If you can find a way to automatically identify one as being derivative of the other, patent it,” responded Alex Graverly on Twitter, creator of GitHub Copilot.
This comes at a time when Microsoft is looking at developing Copilot technology for use in similar programmes for other job categories, like office work, cyber security, or video game design, according to a Bloomberg report.
Microsoft's chief technology officer revealed that the tech giant will build some of the tools itself, while others will be provided by its customers, partners, and rivals.
Examples of what the technology could do include helping video game creators make dialogue for non-playable characters, while the tech giant’s cyber security teams are investigating how the tool can help combat hackers.
GitHub did admit that in some cases Copilot can produce copied code, with the current version of the tool aiming to prevent suggestions that match existing code in public repositories.
Zach Marzouk is a former ITPro, CloudPro, and ChannelPro staff writer, covering topics like security, privacy, worker rights, and startups, primarily in the Asia Pacific and the US regions. Zach joined ITPro in 2017 where he was introduced to the world of B2B technology as a junior staff writer, before he returned to Argentina in 2018, working in communications and as a copywriter. In 2021, he made his way back to ITPro as a staff writer during the pandemic, before joining the world of freelance in 2022.
-
Bigger salaries, more burnout: Is the CISO role in crisis?
In-depth CISOs are more stressed than ever before – but why is this and what can be done?
By Kate O'Flaherty Published
-
Cheap cyber crime kits can be bought on the dark web for less than $25
News Research from NordVPN shows phishing kits are now widely available on the dark web and via messaging apps like Telegram, and are often selling for less than $25.
By Emma Woollacott Published
-
2024 was the year where AI finally started returning on investment
Opinion It's taken a while, but enterprises are finally beginning to see the benefits of AI
By Ross Kelly Last updated
-
Google CEO says more than 25% of the company's code is now AI-generated – is this the future of software development?
News Google’s AI-generated code focus is a sign of the times in software development
By George Fitzmaurice Last updated
-
Apple is working on an AI coding tool similar to GitHub Copilot - and it could be a game changer for its generative AI ambitions
News The AI coding tool from Apple will function in a similar way to GitHub Copilot, and would mark a major foray into the generative AI race for the tech giant
By Ross Kelly Published
-
Meta’s Code Llama AI coding tool just got a big performance boost
News New versions of generative AI coding tool arrives, supporting Python, C++, Java, PHP and more
By Steve Ranger Published
-
GitHub Copilot Enterprise promises to bring back “the joy of coding”
News The GitHub Copilot Enterprise tier marks the latest AI-powered product roll-out from the firm
By Solomon Klappholz Published
-
The best AI tools for business to try today
In-depth Explore our curated list of top AI tools for business, showcasing how generative AI is transforming workflows across industries—from design to data analysis—offering accessible, powerful solutions to enhance productivity and creativity.
By Barry Collins Last updated
-
Meta just released its answer to GitHub Copilot, and it’s free
News Meta's specialized programming LLM can be run on a single GPU
By Rory Bathgate Published
-
ChatGPT gives wrong answers to programming questions more than 50% of the time
News Some developers may be placing too much faith in generative AI, experts warn
By Ross Kelly Published