Navigating Open and Closed-Source Generative AI in Cyberspace
Introduction
India stands at a critical juncture in the domain of
artificial intelligence (AI), particularly in the realm of generative AI
(GenAI). As both the government and private sector seek to harness the
potential of AI, a key issue emerges: how to balance open-source and
closed-source GenAI models within the country’s legal and regulatory framework.
This challenge poses significant implications for India's burgeoning tech
industry, national security, and cyber governance, especially as the country
aspires to become a global leader in AI.
The Rise of
Generative AI
Generative AI refers to AI systems capable of creating text,
images, music, and even complex software code. Models like OpenAI's GPT,
Google's Bard, and others are part of a larger technological revolution,
offering businesses, governments, and individuals unprecedented capabilities.
These models can facilitate content creation, automate tasks, and even support
decision-making across sectors.
However, the debate surrounding open-source versus
closed-source models has gained traction. Open-source models, such as Meta’s
LLaMA (Large Language Model Meta AI), offer greater transparency and allow
developers to modify and improve the AI, while closed-source models, such as
OpenAI’s GPT-4, prioritize control, intellectual property, and proprietary
innovation.
In India, this debate is not just technical but legal and
policy-driven, bringing up questions about cybersecurity, data privacy, and
ethical AI deployment.
The Open vs
Closed-Source Debate in India
Open-Source GenAI:
Open-source AI systems are lauded for their transparency,
which promotes innovation by enabling developers to modify and adapt the code.
In India's thriving tech ecosystem, this model aligns well with the culture of
innovation and low-cost solutions. Open-source GenAI could empower small
businesses and startups by offering access to cutting-edge AI without the steep
costs associated with proprietary systems.
However, open-source models present certain risks. With
greater accessibility comes the potential for misuse. Malicious actors could
modify open-source models to create deepfakes, disinformation campaigns, or
cyber-attacks. India, already grappling with misinformation issues in its vast
digital landscape, may find open-source GenAI to be a double-edged sword.
Closed-Source GenAI:
Closed-source models, on the other hand, offer better
control and security. Companies like OpenAI keep their models proprietary to
safeguard against misuse and protect intellectual property. For India,
closed-source GenAI could be an attractive option for sectors such as defense,
finance, and critical infrastructure, where data security is paramount.
However, the closed-source nature of these models could
limit the growth of domestic AI capabilities. Relying on proprietary
technologies from global corporations could lead to dependency on foreign
entities, stifling India's ambitions of becoming a self-reliant AI powerhouse.
Furthermore, closed models lack the transparency that open-source ones offer,
making it harder for Indian regulators to assess compliance with data
protection laws.
The Legal and
Regulatory Conundrum
India's legal framework for AI, though still evolving, faces
unique challenges in regulating both open and closed-source GenAI models.
1. Data Privacy: India is in the process of implementing the
Digital Personal Data Protection Act (DPDPA) of 2023, which emphasizes user
consent and data protection. The deployment of GenAI models—whether open or
closed—requires massive amounts of data for training. If data used for these
models includes personal information, it raises significant privacy concerns.
Open-source models, in particular, could be more difficult to regulate in terms
of data handling and compliance with privacy laws due to their decentralized nature.
2. Cybersecurity: Open-source GenAI models could become
tools for cybercriminals if misused, raising concerns for India's cybersecurity
apparatus. The proliferation of deepfakes, automated misinformation, and
AI-driven cyber-attacks could overwhelm existing legal frameworks.
Closed-source models offer better control, but India must ensure that they do
not operate as opaque systems immune to legal scrutiny.
3. Intellectual Property (IP) and Copyright: Generative AI
raises complex questions regarding IP and copyright laws. For instance, if an
AI-generated piece of art or music infringes on existing copyrighted works, who
is held responsible—the developer of the open-source model, the end user, or
the model itself? India’s current IP laws are not yet equipped to handle such
intricate issues, adding another layer of complexity.
4. AI Ethics and Accountability: Ensuring ethical AI use is
a global concern, and India is no exception. Open-source models, with their
decentralized development, could struggle to adhere to ethical guidelines, such
as bias mitigation and responsible AI usage. Closed-source models, while more
controlled, may still operate under opaque algorithms, making accountability
difficult.
India’s Policy
Response and the Way Forward
The Indian government has recognized the need for a balanced
approach to AI regulation. The National Strategy for AI (NSAI) focuses on
making India a global AI hub while ensuring responsible AI deployment. However,
striking a balance between open and closed-source models will require nuanced
policymaking.
1. Regulating Open-Source GenAI: India could consider a
regulatory framework that encourages innovation while setting boundaries for
the misuse of open-source AI. Mandatory audits, traceability of modifications,
and ethical guidelines could help mitigate risks without stifling innovation.
2. Supporting Closed-Source GenAI with Safeguards: While
closed-source models offer security, India must ensure that these models are
compliant with domestic laws and do not operate in a "black box."
Transparent AI practices, regulatory oversight, and ethical standards could
help ensure accountability.
3. Public-Private Partnerships (PPP): Given the rapid
evolution of AI, India’s policymakers could collaborate with industry
stakeholders to co-create frameworks that balance innovation and regulation.
PPPs could also facilitate the development of homegrown AI models, reducing
dependency on foreign technologies.
4. International Cooperation: AI transcends borders, and
India’s regulatory framework must align with global best practices. Cooperation
with international organizations on AI standards, ethics, and cybersecurity
could help India navigate the challenges posed by both open and closed-source
models.
Conclusion
India’s journey towards AI dominance hinges on its ability
to balance the advantages and challenges posed by open and closed-source GenAI
models. The legal regime governing cyberspace will play a critical role in
determining how these technologies can be deployed safely, ethically, and
innovatively. As the country looks to shape its future in the AI-driven world,
a comprehensive and forward-looking policy framework will be essential in
maintaining this delicate balance.
India must ensure that the pursuit of AI excellence does not
come at the cost of cybersecurity, privacy, or ethical standards, making the
balancing act between open and closed-source GenAI a defining challenge for its
legal regime in cyberspace.
Kapeesh Law Chambers
Lawyers Chamber Block-II,
Delhi High Court, New Delhi-110003
admin@kapeeshlawchambers.com
www.kapeeshlawchambers.com