Bahra University, Shimla Hills,
Waknaghat, Dist. Solan, H.P.
Research Paper
Legal Frameworks For Emerging AI Technologies: Challenges And
Opportunities In The 21st Century
[Link] Ranaut *, Sahil Badhan *
*
Research Supervisor, Professor, School of Law, Bahra University, Shimla Hills, H.P.
*
Research Scholar, LL.M 2nd Semester, School of Law, Bahra University, Shimla Hills, H.P.
Abstract:
The rapid proliferation of Artificial Intelligence (AI) technologies in recent years has brought about
remarkable advancements across diverse sectors—ranging from healthcare and finance to
governance and law enforcement. While AI offers transformative capabilities in data processing,
prediction, and automation, it simultaneously introduces profound legal, ethical, and societal
complexities. Traditional legal structures, crafted in a pre-digital era, are now under considerable
strain to adapt to the intricacies of machine learning, autonomous systems, and algorithmic
decision-making. This paper examines the evolving legal frameworks that aim to govern emerging
AI technologies, focusing on the challenges they pose to democratic accountability, privacy,
liability, and fairness. It also identifies potential opportunities that AI law can harness to support
innovation and uphold human rights. Through a detailed doctrinal and comparative analysis, the
research explores global approaches to AI governance, highlighting the gaps and proposing future
strategies for responsive regulation. The paper further elaborates on how law must evolve to
balance technological innovation with ethical and constitutional safeguards. In doing so, it calls
for a legal paradigm that is dynamic, anticipatory, and deeply rooted in human-centric values.
Keywords: Artificial Intelligence, AI Regulation, Legal Frameworks, Algorithmic Bias, Liability,
Ethics, Emerging Technologies, Governance, Future of Law.
1. Introduction
The integration of Artificial Intelligence into modern society represents one of the most profound
technological shifts of the 21st century 1. From automating routine tasks to predicting complex
human behaviour, AI is no longer confined to science fiction—it is an active participant in real-
world decision-making processes. Governments employ AI for predictive governance, businesses
use it for consumer profiling, and health systems rely on it for diagnostics. However, these
capabilities, while promising, raise fundamental legal questions concerning accountability,
fairness, transparency, and the safeguarding of fundamental rights 2. AI does not operate within a
moral vacuum; its decisions can influence lives in ways traditionally reserved for human judgment.
For instance, when a self-driving vehicle is involved in an accident, who bears legal
responsibility—the manufacturer, the software developer, or the end user? Similarly, when
algorithms deny loan applications or determine bail eligibility based on biased data, the
implications for equality and justice are severe. These dilemmas expose the legal system’s
limitations in regulating increasingly autonomous and opaque technologies3.
The primary concern that frames this research is the lack of a cohesive, flexible, and anticipatory
legal framework capable of addressing the multifaceted implications of AI technologies. As
jurisdictions struggle to catch up with technological advances, the legal vacuum risks enabling
harm, amplifying inequality, and undermining democratic values. This research seeks to identify
how legal systems can evolve to fill this void—ensuring that AI development remains aligned with
ethical norms, legal accountability, and public interest4.
2. Existing Legal Frameworks Impacting AI in India
2.1 Information Technology Act, 2000
1
United Nations Educational, Scientific and Cultural Organization (UNESCO), Recommendation on the Ethics of
Artificial Intelligence (2021) [Link] accessed 20 April 2025.
2
Data Security Council of India (DSCI), AI in India: A Strategic Overview (2020) 11.
3
Indian Ministry of Electronics and IT (MeitY), Responsible AI Strategy for India: Part 1 (2021) 6.
4
Usha Ramanathan, The Right to Privacy Judgment and its Ramifications (2018) 4 Indian Journal of Constitutional
Law 34.
Originally enacted to govern electronic transactions, the IT Act is India's backbone for regulating
cyberspace. Though it lacks explicit AI references, it indirectly governs AI systems through the
following provisions:
Section 43A: Provides compensation for failure to protect sensitive personal data. AI models using
user data without sufficient safeguards could be held accountable under this section.
Section 66: Addresses hacking and data theft, relevant where AI systems are used to infiltrate data
systems.
Shortcomings: The Act fails to address issues unique to AI such as autonomous decision-making,
algorithmic transparency, and ethical accountability.
2.2 Digital Personal Data Protection Act, 2023
This recent legislation aims to safeguard personal data. Key provisions with AI implications
include:
Consent Framework: AI platforms must obtain express consent before using personal data.
Purpose Limitation: Data must only be processed for specified purposes. AI algorithms operating
on indiscriminate datasets could violate this.
Significant Data Fiduciaries: Entities handling large volumes of data (like AI firms) face higher
compliance.
Limitations: Lacks specific mechanisms to address automated decision-making or the right to
explanation, as seen in GDPR.
2.3 Constitution of India
Fundamental rights under the Constitution are increasingly intersecting with AI technologies:
Article 14 (Right to Equality): Biased AI decisions (e.g., in policing or credit scoring) can lead to
discrimination.
Article 21 (Right to Life and Personal Liberty): Post Puttaswamy judgment, the right to privacy
has become central, and AI surveillance without oversight challenges this.
2.4 Consumer Protection Act, 2019
This Act holds relevance in AI-based service delivery:
Misleading Advertisements and Defective Services: AI-generated content or recommendations
that mislead users can attract liability.
Autonomous Services: Businesses using chatbots or automated systems must ensure they comply
with consumer rights.
2.5 Bharatiya Nyaya Sanhita (BNS) Act, 2023
While BNS does not explicitly address artificial intelligence, several sections are adaptable to AI-
related crimes:
Section 69 – Impersonation: Criminalizes pretending to be someone else. Eg: AI-generated voice
cloning used in financial fraud
Section 316 – Cheating: Penalizes deception for dishonest gain. Eg: AI chatbots deceiving people
into scams
Section 354 – Obscenity: Regulates circulation of obscene material. Eg: Deepfake pornography
shared online
Section 356 – Defamation: Penalizes harm to reputation. Eg: AI-generated fake news targeting
public figures
Section 336 – Criminal Intimidation: Covers threats to harm. Eg: AI-powered deepfakes used to
blackmail victims While BNS does not explicitly address artificial intelligence, several sections
are adaptable to AI-related crimes:
3 AI-Related Offenses and Sector-Wise Impact on Indian Society
Ye Sector AI Real Life Concerns Relevan Estimat Notable
ar Application Offences & Societal t Legal ed % of Reports/Inci
s Impact Provisio Populat dents
(2019- ns ion
2025) Affecte
d
201 Healthca Diagnostics, AI-based Limited CPA 2.5% WHO AI in
9 re surgery app trials led to 2019, Healthcare
assistance misdiagnose diagnostic IPC Sec India Report
d rural errors in 304A, (2019)5
patient in rural AI now
Maharashtra
5
WHO, Artificial Intelligence in Healthcare in India: Opportunities and Challenges (2019)
[Link]
deploymen BNS Sec
ts 106
202 Agricultu Crop AI service High costs IT Act, 10% NITI Aayog
0 re monitoring charged and Agricult AI for All –
pest marginal technical ural Agri Report6
detection farmers for barriers for Produce
inaccurate marginal Act
pesticide farmers
forecasts
202 Educatio Automated EdTech Language RTE Act, 5% NCERT
1 n tutors, startup’s exclusion IT Act Digital
grading grading tool in grading Education
systems marked non- apps Policy
English affected Review7
medium state board
students students
unfairly
202 Law Predictive Hyderabad Human Article 3% NHRC
2 Enforce policing, police AI rights 14, IPC Advisory on
ment surveillance flagged violations Sec 153 AI
tools tribal youth surfaced now Surveillance
as potential due to BNS Sec (2022)8
threat opaque 192
facial
recognition
use
6
NITI Aayog, National Strategy for AI in Agriculture (2020) [Link]
7
NCERT, Policy Review on Digital Education Tools in India (2021) [Link]
8
National Human Rights Commission (NHRC), Advisory on Use of AI Surveillance by Police Forces (2022)
[Link]
202 Finance Credit Loan app Biases in SEBI 6% RBI Sandbox
3 scoring, loan denied FinTech Act, RBI Report on
approval credit to loan Guidelin FinTech Bias
applicant algorithms es (2023)9
based on excluding
flawed AI low-
profiling income
groups
202 E- Recommend Customer Profiling IT Act 7% TRAI Report
4 commerc ation received without Sec 43A, on AI in
e engines, harmful consent Consum Digital
customer product due and er Economy
service bots to targeted Protectio (2024)10
misleading misinforma n Act
AI tion issues
recommend
ation
202 Public Welfare AI tool Digital IT Act, 4% MeitY
5 Governa allocation, rejected exclusion Digital Consultation
nce document welfare due to India Act on AI
processing application language (Draft) Inclusion
of visually and literacy (2025)11
impaired barriers in
woman rural AI use
9
Reserve Bank of India, Report of the Working Group on Digital Lending Including Lending Through Online
Platforms and Mobile Apps (2023) [Link]
10
Telecom Regulatory Authority of India (TRAI), AI and the Digital Economy: Regulation and Responsibility
(2024) [Link]
11
Ministry of Electronics and Information Technology (MeitY), Draft Digital India Act and Inclusive AI
Consultation Report (2025) [Link]
4. Challenges Associated with Artificial Intelligence in the 21st Century
The expansion of AI technologies, while revolutionary, brings with it a set of pressing challenges
that legal and regulatory systems worldwide are struggling to keep pace with. These challenges
include ethical dilemmas, privacy risks, labour displacement, and geopolitical security concerns:
4.1 Ethical and Legal Gaps: AI systems are now participating in decision-making processes that
bear significant moral weight—such as granting parole, assessing job candidates, or diagnosing
medical conditions. However, most legal systems are still unequipped to establish responsibility
when AI makes flawed or biased decisions. The so-called “black box” problem—where the
decision-making process of an AI system is opaque even to its developers—undermines core
principles of transparency, accountability, and justice. Existing liability laws, rooted in human
intent and control, are inadequate to address harms caused by autonomous systems operating
independently
4.2 Data Privacy and Surveillance Risks: AI's functionality depends heavily on massive volumes
of data, much of which is personal, sensitive, or collected without meaningful consent. This raises
grave concerns about the erosion of digital privacy and the potential for surveillance by both
corporations and governments. In regions where data protection laws are weak or underdeveloped,
such as India, individuals face heightened vulnerability. Comparatively, the European Union’s
General Data Protection Regulation (GDPR) offers a robust model for privacy rights and
accountability, but similar standards are not yet globally enforced.
4.3 Employment and Economic Displacement: As AI continues to automate routine and
repetitive tasks, there is a legitimate fear that millions of jobs, especially in the manufacturing and
service sectors, may become obsolete 12. While some reports predict that AI will also create new
job roles, the transition may not be equitable. Without proper social safety nets, retraining
programs, and legal protections for gig and platform workers, this shift could exacerbate existing
inequalities and lead to widespread socio-economic instability.
4.4 AI and Global Security Risks: The deployment of AI in military applications, particularly in
the form of autonomous weapons systems, introduces a new set of ethical and security threats.
These systems can independently select and engage targets, raising serious concerns about
compliance with international humanitarian law and the ethics of warfare. The absence of a global
12
Jack Balkin, The Three Laws of Robotics in the Age of Big Data (2017) 78 Ohio State Law Journal 1217.
treaty or regulatory framework to govern such technology could potentially fuel an AI arms race,
endangering international peace and security. 13
5. Transformative Opportunities Offered by AI in Modern Society
Despite its challenges, AI holds immense promise to reshape society in constructive and
transformative ways across multiple domains:
5.1 Revolutionising Healthcare: AI is driving remarkable innovations in healthcare, from early
disease detection through AI-enhanced imaging tools to virtual assistants that support patient
interactions. AI systems can analyse vast amounts of medical data to identify patterns, predict
outbreaks, and customise treatment plans, thereby improving both the quality and accessibility of
healthcare services—especially in underserved or rural areas.
5.2 Enhancing Public Governance: Governments worldwide are increasingly adopting AI to
improve service delivery, detect fraud, and make data-driven decisions. In India, AI has been
successfully integrated into public grievance redressal systems, smart traffic management, and
welfare programme distribution14. These tools not only improve efficiency and transparency but
also help reduce administrative bottlenecks and corruption.
5.3 Redefining Education and Skills Development: AI technologies are transforming education
through personalised learning platforms that adapt to individual students’ needs. Virtual tutors and
automated grading systems reduce teacher workload and enable scalable learning environments.
These innovations are vital in promoting inclusive education and in preparing populations for
future job markets impacted by automation 15.
5.4 Promoting Environmental Sustainability: AI plays a crucial role in tackling climate change
and promoting ecological balance. From weather forecasting and disaster prediction to smart
energy grids and precision agriculture, AI technologies enable smarter resource management16.
They help reduce emissions, monitor biodiversity, and enhance food security through data-driven
farming methods.
13
AlgorithmWatch, Automated Decision-Making Systems in the EU Public Sector (2020) 6.
14
IEEE Global Initiative, Ethically Aligned Design: A Vision for Prioritizing Human Well-being with Autonomous
and Intelligent Systems (First Edition, 2019) 89.
15
Nikolas Guggenberger, Essential Facilities in the Digital Economy (2020) 24 Stanford Technology Law Review
109.
16
Malgorzata Kurkowska, Artificial Intelligence and Public Law: Procedural Safeguards in Automated Decision-
Making (2021) 47 European Law Review 113.
5.5 Stimulating Economic Growth and Innovation: AI is reshaping the global economy by
enhancing business efficiency, automating complex processes, and enabling data-driven
innovation. Companies use AI to forecast market trends, optimise supply chains, and tailor services
to customer preferences. Start-ups and established firms alike benefit from AI's capacity to drive
competitive advantage and economic scalability.
6. Future Outcomes
The future trajectory of AI legal regulation will likely shape the broader contours of global
governance, innovation, and civil liberties. Some foreseeable outcomes include:
Creation of AI-specific Courts or Tribunals: To handle complex liability and rights-based issues
arising from autonomous systems.
Mandatory AI Impact Assessments: Before deployment, especially for public sector AI, systems
may be subject to human rights and fairness audits.
Dynamic and Hybrid Legal Instruments: Law may adopt hybrid models—combining hard law,
soft law, and industry-led codes of conduct—to remain flexible and responsive.
Embedding Legal Tech in Justice Systems: AI may not only be regulated but also used to improve
legal access and efficiency—through tools like automated drafting, e-discovery, and virtual
hearings.
The evolution of these frameworks will determine whether AI is harnessed for empowerment or
becomes a tool of control and inequality.
Conclusion
Artificial intelligence represents one of the most transformative technological advancements of the
21st century, offering unprecedented opportunities for social and economic development.
However, its rapid growth also poses complex legal, ethical, and regulatory challenges that demand
urgent and thoughtful [Link]’s current legal framework, while evolving, is insufficiently
equipped to address the multifaceted implications of AI. The absence of dedicated AI legislation,
gaps in data protection, unclear liability regimes, and fragmented regulatory oversight create
vulnerabilities that risk undermining fundamental rights and public trust. Drawing on comparative
experiences from the European Union’s rights-centric approach and the United States’ innovation-
driven model, India has the unique opportunity to craft a hybrid framework that integrates
constitutional safeguards with pragmatic, sector-specific regulation. Such a framework would
promote responsible AI innovation while protecting individual dignity, privacy, and equality.
To realize this vision, India must enact comprehensive AI laws, strengthen data protection,
establish specialized regulatory bodies, and foster multi-stakeholder engagement. Additionally,
capacity building and alignment with international standards will be crucial to navigating the
global AI landscape effectively. Ultimately, regulating AI in India is not merely a technical or legal
challenge but a democratic imperative to ensure that technology serves as a tool for inclusive
growth, justice, and human flourishing. The journey ahead demands collaborative efforts grounded
in India’s constitutional ethos and socio-economic realities to shape AI governance that is both
visionary and grounded in justice.