Subscribe to Ex Machina to receive new posts directly to your inbox.
There are many who claim that the harms AI can cause can only be addressed by a new legislation specifically designed to address them. That said existing laws have, more often than not, been framed in terms that are broad enough to deal with these harms regardless of the technology that caused them.
This is a link-enhanced version of an article that first appeared in the Mint. You can read the original here.
In an earlier article in this column, I spoke of a lecture that Justice Easterbrook once delivered on the subject of property in cyberspace. His talk was titled ‘Cyberspace and the Law of the Horse,’ which was his way of highlighting the foolishness of coming up with fresh laws to regulate new technologies when general principles of law already addressed the concerns these new technologies raised.
He argued that there had been a number of cases where horses were bought and sold; where redress was sought for injuries caused when people were either kicked by a horse or had fallen off one of them. None of this, he said, moved us to enact a special “Law of the Horse” to deal with all these “new” harms. The general law of transfer of property, of torts, and the regulation of commercial transactions were more than enough to provide a legal framework within which these harms could be addressed.
This is advice that many of those who are writing about artificial intelligence (AI) regulation would do well to heed. Much of what is being written about AI and the harms it could cause point to the need for fresh regulation that writers feel will address the harms AI can cause. Some jurisdictions, like the EU, have already acted on these suggestions by enacting new AI legislation. Given the coercive power of the Brussels Effect, many other countries are following in its footsteps and it is more likely than not that Indian regulators will be compelled to consider going down the EU’s path.
But do we really need a new law to specify how the harms caused by AI need to be addressed? To answer this question, I took a closer look at some of the harms attributed to AI to see if they were so special that they needed a new regulation, or whether existing law was sufficient to address the harms they caused.
Let’s take, for instance, fake news—the concern that AI can be used to generate content (either audio or video) that can make it seem as if individuals had said or done something they, in fact, did not do. Voice AI has reached a point where it is capable of mimicking a real person’s speech in both intonation and mannerism. Video AI can generate footage that makes it appear as if they are doing things they never did. This is impersonation, and, regardless of the technology used, is not only a violation under the Information Technology Act of 2000, it is also a crime under the Bharatiya Nyaya Sanhita, 2023. Where it causes harm to a person’s reputation, it also amounts to defamation.
Then there is the concern that AI might be used to incite communal tension—that someone might use it to create the impression that certain persons are purposefully offending the sensibilities of those of another caste or community with a view to incite conflict. The use of electronic communications to promote disharmony and feelings of enmity between religious, racial or regional groups is a crime under Section 196 of the Bharatiya Nyaya Sanhita. It matters little what technology was used, and there is no need to enact a special law just to address this concern.
Pretty much any harm that you think is exacerbated by AI—be it misleading advertising, election interference, forgery or bias—is covered by existing provisions of law. In almost every instance, since these laws have been drafted so broadly as to cover a wide range of circumstances, it will make little difference if the harm is committed using a technology that did not exist at the time when the law relevant to the particular case was enacted.
This is why I do not believe there is any need to enact a brand-new AI Act to address the so-called harms of AI. What we might need to do is educate our regulators about AI and the fact that it makes it possible to commit crimes in ways that were not previously possible. We also need to train them on new forensic techniques that will help them investigate AI crimes, so that they can better detect whether or not AI has been used. Wherever possible, we should teach them how they can use AI itself to do this detection work. Chances are, this will reduce the time and cost of investigating AI harms.
While we may not need to enact a new law to address the harms caused by AI, there may be a need for us to amend existing regulations to maximize the benefits we can extract from this new technology. Since most existing legal frameworks were conceptualized before Generative AI, some of the language used in these statutes may, if interpreted strictly, not permit some of what AI does.
Take, for example, intellectual property law. Copying a literary work without the permission of its author is technically a violation of copyright. However, despite having ingested literary works as part of their training data-sets, AI models almost never output verbatim the actual excerpts of those works. And since copyright law only protects the specific expression of ideas in original works of authorship—not the underlying ideas in a work—prompt responses that reference the ideas in a literary work do not violate the spirit of the law.
Many countries have recognized this distinction and introduced specific exemptions to exclude “training” from the strict applicability of the law. This is referred to as the text and data mining (TDM) exemption, and jurisdictions like Japan and Europe have incorporated it into their law. We should review our own laws to make sure existing legal provisions do not accidentally come in the way of our AI ambitions.
Collect this post as an NFT.
Over 2.6k subscribers
Great analysis again. Totally agree that the laws and regulations around AI balance the risk v/s benefit, and on a topic by topic basis. What makes sense in the West probably won[t make sense for us
We're back with the 25th edition of Paragraph Picks, highlighting a few hand-selected pieces from the past week or so.
@tch writes about how smart wallets can achieve mainstream adoption through better opportunities for financial growth, robust social recovery mechanisms for access restoration, and advanced protection features to ensure privacy and security in the evolving blockchain ecosystem. "Safe, secure, and smart wallets will make people richer and their lives easier and more fulfilling." https://paragraph.xyz/@0x2e70f8a381efd5142474e8ce61fa3d05d3ce0576/musings-on-the-future-of-actually-smart-wallets
@macbudkowski points out that rather than attempting to onboard the next billion users to Ethereum through a single mainstream app, the focus should be on building hundreds of well-targeted apps that solve specific problems for smaller groups of users, creating incremental growth and paving the way for broader adoption. "What Ethereum needs is 100s of apps targeting 100k-10 million users." https://kanfa.macbudkowski.com/onboarding-next-billion-users-ethereum
@papa writes about how the rise of pseudonymous tools and zero-knowledge (ZK) applications on decentralized networks like Farcaster demonstrates the potential for private, verifiable communication and collaboration, blending individual privacy with trust in digital identities. "Pseudonymity has allowed individuals to share ideas & truths unburdened by fear of retaliation." https://paragraph.xyz/@papajams.eth/farcasters-zk-anons
@tomu writes about how Farcaster is pioneering programmable social networks, integrating onchain actions into social interactions through AI agents and open graphs, creating seamless, collaborative ecosystems where engagement drives economic activity. "Agents turning social posts into onchain actions are creating a new way to communicate that wasn’t possible before." https://tomu.xyz/programmable-social
Nice piece :) I have long wondered where the web3 (fun!) games are. Seems like a lot of potential for mechanics and particularly cross-party metagames i.e. you level up INT in an RPG by doing winning a Catan game or whatever
Truly enjoyed this!!