
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 11 Issue: 12 | Dec 2024 www.irjet.net p-ISSN: 2395-0072
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 11 Issue: 12 | Dec 2024 www.irjet.net p-ISSN: 2395-0072
Priyanka Sajnani1
1School of Computer Science Engineering and Application, D Y Patil International University Akurdi pune 411035
Abstract - Artificial Intelligence (AI) continues to revolutionise industries, presenting immense opportunities while simultaneously introducing intricate challenges of accountability, transparency, andethics. This reviewexplores themultifacetedlandscapeofAIaccountabilitybysynthesising insights from diverse domains, including legal liability, intellectual property, ethical considerations, and regulatory frameworks. Keythemes includethe complexities ofascribing accountability in AI-driven systems, especially in scenarios involving criminal liability and decision-making failures, and the proposal of electronic legal personhood for autonomous systems. The interplay between human biases, such as selfservingattributioningenerativeAIuse,andthesociotechnical nature of AI governance is examined to reveal gaps in current accountabilityframeworks. Further, this paper delvesintothe regulatory implications of transparency, particularly under frameworks like the General Data Protection Regulation (GDPR), and the emerging need for relational transparency that extends beyond information provision to meaningful communication. Ethicalconsiderationsinsectorslikefinancial services and intellectual property law are critically analysed, emphasising the need for clear guidelines to navigate algorithmic bias, fairness, and trustworthiness. Through this synthesis, the review proposes a novel framework integrating legal, ethical, and societal dimensions to redefine AI accountability. This approach underscores the importance of proactive governance mechanisms, such as enhanced transparency measures, context-sensitive accountability architectures, and industry-wide ethical standards, to foster responsible AI innovation.
Key Words: AI accountability, legal liability, transparency, ethical AI, intellectual property, GDPR, criminal responsibility, governance mechanisms, algorithmic fairness.
Artificial intelligence (AI) has become an integral part of modern life, driving innovation and reshaping industries ranging from healthcare and education to business and governance.Amongthemosttransformativeadvancements withinAIisthedevelopmentofGenerativeAI(GAI),which leveragesdeeplearningandmachinelearningalgorithmsto produce novel content, including text, images, and audio, basedonuserinput.ToolslikeChatGPT,GitHubCopilot,and DALL-E have garnered significant attention due to their potentialtoenhanceproductivity,creativity,andefficiency
in tasks suchas writing, coding, and designing. This rapid adoptionofGAItoolshighlightstheirtransformativerolein improvingproblem-solvinganddecision-makingcapabilities intheworkplace.
However, the integration of GAI systems into human activitiesbringsforthchallengesrelatedtoethics,reliability, and accountability. Instances such as biased outputs in recommendation letters or the spread of misinformation underscore the risks associated with GAI misuse or overreliance. These issues highlight the importance of ensuringaccountabilityinAIsystemstomitigatepotential harms.AccountabilityinAI encompassestheobligationof actors involved in the design, deployment, or use of such systemstoexplaintheiractions,justifydecisions,andtake responsibilityforadverseoutcomes.Withoutwell-defined accountabilityframeworks,addressingtherootcausesofAI failures becomes challenging, thereby increasing the likelihoodofrepetitionandescalatingrisks.
Akeyobstaclein establishingaccountabilitywithinGAIis the "many hands" problem, wherein responsibility is diffusedamongmultiplestakeholders,suchasdevelopers, organisations,andusers.Thiscomplexityisexacerbatedby the opaque nature of GAI systems, which often lack transparency in decision-making processes. For example, whenGAI-generatedoutputsresultinerrorsordamages,it becomesdifficulttopinpointresponsibility.Thisdiffusionof accountability can lead to scenarios where individuals underestimatetheirroleintheproperuseofthesesystems, fostering overreliance and potentially detrimental consequences,suchasautomationbiasordatabreaches.
Understanding how accountability is perceived and distributed among stakeholders is critical for fostering ethical and sustainable AI usage. Previous research has predominantlyfocusedontheaccountabilityofsingleactors, suchasdevelopersororganisations,andhasexploredhow accountability influences trust, adoption, and behaviour. However, less attention has been given to multi-actor accountability frameworks, particularly in the context of Generative AI. Examining how employees attribute accountabilitytodifferentactors,suchasthemselves,theAI system, or its creators, offers valuable insights into addressingthe"manyhands"problem.Byexploringthese dynamics, we can advance frameworks that promote equitable responsibility-sharing and minimise bias in human-AIcollaborations.
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 11 Issue: 12 | Dec 2024 www.irjet.net p-ISSN: 2395-0072
Thispaperaimstocontributetothisgrowingdiscourseby investigating the perception of accountability among employees when interacting with GAI systems. Drawing uponattributiontheoryandconceptslikeself-servingbias, thisstudyseekstounravelthefactorsshapingaccountability attributionsandtheirimplicationsfortheethicalintegration ofGAIinprofessionalandorganisationalcontexts.
Generative AI and Its Role in Transformative Work
Generative AI (GAI) has emerged as a powerful subset of artificialintelligence,capableofcreatinghuman-likecontent suchastext,music,andvisualswhilesynthesizingdatafor analytical insights. Unlike traditional AI, GAI leverages advancedpatternrecognitiontogenerateuniqueandoriginal outputs, demonstrating transformative potential across diversedomains(Dasborough,2023).TheadoptionofGAIin professional environments has been associated with increased efficiency, accessibility, and cost-effectiveness, driving significant advancements in digital evolution and businessprocesses(Goldsteinetal.,2023).Forinstance,in HumanResourceManagement,GAIenhancestaskssuchas jobdescriptioncreationandintelligentsearch,empowering professionalstoinnovateandshifttheirfocustowardhighvalue,creativework(Saengrithetal.,2022).
Attheindividuallevel,GAIfostersoperationalexcellenceby augmenting decision-making and problem-solving capabilities,markingaparadigmshiftinworkplacedynamics. Its integration not only optimizes repetitive tasks but also redefines roles, enabling employees to engage in more strategicandmeaningfulactivities(Chenetal.,2023).This dualcapacitytogenerateandanalyzecontentmakesGAIan indispensabletoolinmodernorganizationaltransformation.
Domain Key Tool Benefits
Human Resources (HR)
ChatGPT
Education DuolingoAI
ContentCreation DALL-E
Automated resume screening
Personalizedlearning pathways
Efficient visual contentgeneration
The integration of GAI introduces challenges in accountability, particularly in addressing ethical concerns and ambiguities surrounding responsible actors. Accountability in AI systems is conceptualized as the obligation of involved parties developers, users, and decision-makers toexplain,justify,andtakeresponsibility foroutcomesinfluencedbyAI,asdefinedbyBovens(2007). However,achievingaccountabilitybecomesconvoluteddue
to the "problem of many hands," where multiple actors contribute to the system, creating difficulty in attributing responsibility(Cooperetal.,2022).
Thisambiguityisparticularlyevidentindisagreementsover whoshouldbeheldaccountableforAIoutcomes.Whilesome arguethatalgorithmdevelopersbearprimaryresponsibility (Diakopoulos, 2016), others contend that designers and managersalsoshareaccountabilityforsocietalimpacts(Shin & Park, 2019). The legal framework adds another layer of complexity, as algorithms lack legal personhood, limiting their accountability under current laws (Deibel, 2021). Despite these challenges, studies in human-computer interactionsuggestthatpeopleoftenperceiveAIasasocial actor, applying expectations and accountability norms typicallyreservedforhumans(Moon,2003).
Thismultifaceteddebatehighlightstheurgencyofclarifying accountabilitymechanismsforGAIsystems.Asorganizations increasingly rely on these technologies, the need for transparent frameworks that delineate responsibilities amongstakeholdersisparamount.
Challenge Description Example
Diffused Responsibility
Shared accountability across multiple actors makes pinpointing liabilitydifficult
Autonomousvehicle accidents
Lack of Legal Clarity Absence of specific lawsforAIsystems UnregulatedAItools inhealthcare
EthicalConcerns Potential for bias and unethicaldecisions Algorithmic discrimination in hiring
AccountabilityremainsapivotalfocusinAIgovernance, emphasizingtransparency,fairness,andadherencetoethical standards.Sinclair(1995)andTheodorou&Dignum(2020) highlighttheneedforstructuredaccountabilityframeworks tonavigateAI'ssociotechnicalcomplexity.Currentpolicies, especiallywithintheEuropeanUnion,advocateforrobust accountability mechanisms that include compliance, reporting,oversight,andenforcementtomitigatepotential risks.ThesemechanismsaimtoenhancetrustandensureAI systems align with societal values. However, challenges persist in defining accountability across diverse contexts, illustratinggapsinharmonizingitsconceptualizationacross sociotechnicalandlegalframeworks.
The rapid evolution of AI technologies has exposed inadequaciesincurrentlegalsystems,particularlyregarding liability for AI's autonomous actions. Baldwin (2019) and Sukhodolovetal.(2020)explorescenariossuchasaccidents
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 11 Issue: 12 | Dec 2024 www.irjet.net p-ISSN: 2395-0072
caused by self-driving cars, where culpability remains unclear. The legal debate extends to whether AI entities couldbegrantedlegalpersonhoodorsubjectedtomodified criminalliabilityprinciples.Existingframeworksoftenfailto account for the distributed responsibilities among developers, users, and operators, highlighting a pressing need for updated regulations. These gaps raise significant concerns about the effectiveness of current laws in addressingthenuancedchallengesposedbyAI,particularly initspotentialtoautonomouslycommitharmfulactions.
EthicalgovernanceinAIintertwinesdeeplywithitslegal dimensions. Researchers, including Al-Qusi (2018), emphasizepreservinghumanrightsandprioritizingsocietal welfareasfoundationaltoAIsystemdevelopment.TheIEEE SA(2016)furtheradvocatesembeddingethical principles into AI design, ensuring safety and justice remain uncompromised.ThediscussionextendstoAI'spotentialto operateautonomously,necessitatingaproactiveapproachto regulationthatincorporatesbothtechnologicalinnovation and ethical accountability. A sociotechnical perspective, integrating diverse stakeholders' input, is essential for crafting governance mechanisms that are equitable and effective.
While existing studies provide valuable insights, significant gaps remain in integrating accountability frameworkswithevolvingAIcapabilities.Currentresearch often overlooks the practical implementation of these frameworks across diverse sociotechnical systems. Additionally, limited consensus exists on the scope and nature of AI's legal personality, leaving ambiguities in its application. Future research must address these inconsistencies by proposing adaptive, multidisciplinary solutionsthatalignlegal,ethical,andtechnicaldimensions.
This review underscores the critical interplay between accountability,legaladaptability,andethicalstandardsinAI governance.Thesubsequentdiscussionexploresstrategies forbridgingidentifiedgaps,focusingondevelopingcohesive frameworksthatbalanceinnovationwithsocietalvalues.
Framework Key Features
Strengths
Limitations
Responsible AIDesign Ensures ethical principles in AI development Proactive risk mitigation Implementati onchallenges
Explainable AI(XAI) Focus on transparency and interpretability Enhances usertrust Limited applicabilityfor complex models
Legal Personhood Model
Treats AI as a legalentity Clarifies liability Contradicts traditional legal systems
Themethodologyforthisresearchreviewpaperinvolveda systematic approach to collecting and analyzing relevant literatureonthethemesofaccountability,legalframeworks, and ethical implications in Artificial Intelligence (AI). The primaryfocuswastoidentifystudiesaddressinggovernance mechanisms, liability issues, and the ethical design of AI systems.Athematicorganizationwasemployedtosynthesize insights and highlight interconnections between the identifiedthemes.
Thefollowingdatabaseswereextensivelyused:
Databases: IEEEXplore,SpringerLink,ScienceDirect,JSTOR, andGoogleScholar.
Search Keywords: "AI accountability frameworks," "legal implications of AI," "ethical AI governance," "AI liability," "sociotechnical systems in AI," and "autonomous systems regulations."
Selection Criteria:
Peer-reviewedarticlespublishedbetween2015and2024.
Studies focusing on European AI governance and broader globalperspectives.
Articlesaddressingpracticalcasestudies,suchasself-driving carsandautonomousAIsystems.
Literature emphasizing multidisciplinary approaches involvinglegal,ethical,andtechnicaldimensions.
Thissystematicapproachwaschosentoensuretheinclusion of high-quality, relevant studies that comprehensively address the multifaceted nature of AI governance. Peerreviewedsourceswereprioritizedtoupholdacademicrigor. Thecombinationofthematicandchronologicalorganization allowed for a clear analysis of evolving discussions and emerging gaps. By integrating diverse perspectives, this methodology ensured a robust foundation for identifying inconsistencies and proposing actionable solutions in the subsequentdiscussion.
This study advances our understanding of how accountability attributions function in the context of GenerativeAI(GAI)systems,particularlywhendeployedin professionalenvironments.Rootedinattributiontheoryand theself-servingbias,ourfindingsprovidenuancedinsights intohowemployeesnavigateaccountabilityinsuccessand failurescenarios,sheddinglightontheevolvingdynamicsof human-AIcollaboration.
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 11 Issue: 12 | Dec 2024 www.irjet.net p-ISSN: 2395-0072
Ourresultsrevealanunexpecteddeparturefromtraditional self-serving bias. Contrary to established research, employeesattributedaccountabilityprimarilytothemselves, regardlessofwhethertheoutcomeofusingGAIwassuccess orfailure.Whileself-attributionduringsuccessalignswith the desire to preserve self-worth, the attribution of accountability during failures, often linked to self-critical reflection, indicates a deeper sense of professional responsibility. This finding underscores the unique implications of GAI in fostering heightened personal accountability among users, a behavior less commonly observedwithnon-GAIsystems.
Interestingly,participantsdemonstratedanacuteawareness of shared accountability, recognizing the interconnected roles of developers, organizations, and the AI systems themselves. These findings support the "many hands" problem,emphasizingthecollectiveresponsibilityinherent inGAIsystemdevelopmentanduse.Notably,higherlevelsof sharedaccountabilitywere observedinsuccessscenarios, reflecting users' efforts to balance personal achievements with a collaborative acknowledgment of external contributions.
Thisstudycontributestoaccountabilityliteratureinseveral ways.First,itprovidesempiricalevidencethatGAIsystems elicit accountability dynamics distinct from other AI technologies. Unlike traditional AI,GAI’s opaque decisionmakingprocessesamplifyusers'perceptionsofcontroland accountability, suggesting the need for more tailored frameworks to address the unique characteristics of GAI systems.
Second,whileexistingstudiesoftenexamineaccountability in isolation, focusing on individual actors, this research highlightstheinterplaybetweenmultiplestakeholders.By demonstrating that employees simultaneously attribute accountabilitytothemselves,developers,andorganizations, this study enriches our understanding of multi-actor accountabilityattribution.Italsoprovidesafoundationfor future research on balancing personal and collective accountabilityinAIgovernance.
Lastly, this study offers a novel perspective by examining accountabilityinbothsuccessandfailurecontexts,revealing distinct attribution patterns. This dual-outcome approach provides a more comprehensive understanding of accountability behaviors, enabling future researchers to explore how different outcomes influence long-term accountabilityperceptions.
Our findings have several implications for practice. For employees, the tendency to self-attribute accountability highlightstheimportanceof fosteringcritical engagement with GAI systems. Organizations should encourage employees to view GAI as a tool for augmentation rather thanautomation,promotinginformeddecision-makingand reducingover-relianceonAIoutputs.
For developers, the shared attribution of accountability underscores the need to design systems that align with users'expectationsandfacilitatetransparency.Developers should proactively address concerns about GAI's opaque nature,ensuringusersfeelsupportedandinformedintheir interactionswiththesesystems.
Organizations must also recognize their role in shaping accountability perceptions. By providing training, establishing clear guidelines, and fostering a culture of sharedresponsibility,organizationscanenhancetrustinGAI systems and improve their integration into workplace practices.
Thisstudyisnotwithoutlimitations.First,therelianceona singleexperimentalcontextlimitsthegeneralizabilityofour findings.Futurestudiesshouldreplicatethisresearchacross diverseindustries,culturalsettings,andGAIapplicationsto validate our conclusions. Additionally, while our methodology focused on employees' self-reports, incorporating observational or longitudinal approaches couldprovidericherinsightsintoaccountabilitydynamics overtime.
Second,thecomplexityofaccountabilityattributionsinGAI systemswarrantsfurtherexploration.Futureresearchcould investigatehowfactorssuchasorganizationalculture,task complexity, and user expertise influence attribution patterns. Incorporating moderators like self-efficacy or cognitive load could also deepen our understanding of accountabilitybehaviorsindifferentcontexts.
Finally,theroleofGAIinreshapingteamdynamicsremains anopenquestion.Futurestudiescouldexplorehowshared accountability among team members affects collaboration and decision-making, particularly in high-stakes environments.Addressingthesequestionswillbecrucialfor developingholisticaccountabilityframeworksthatsupport ethicalandeffectiveGAIintegration.
[1]L.McDonald,“AIsystemsandliability:Anassessmentof theapplicabilityofstrictliability&acaseforlimitedlegal personhoodforAI,” UniversityofStAndrewsLawJournal,vol. 3,no.1,2023,pp.5–21,doi:10.15664/stalj.v3i1.2645.
International Research Journal of Engineering and Technology (IRJET) e-ISSN: 2395-0056
Volume: 11 Issue: 12 | Dec 2024 www.irjet.net p-ISSN: 2395-0072
[2] N. S. Uzougbo, C. G. Ikegwu, and A. O. Adewusi, “Legal accountabilityandethical considerationsofAIinfinancial services,” GSC Advanced Research and Reviews,vol.19,no.2, 2024,pp.130–142,doi:10.30574/gscarr.2024.19.2.0171.
[3]G.Du,S.Lins,I.Blohm,andA.Sunyaev,“Myfault,notAI’s fault.Self-servingbiasimpactsemployees’attributionofAI accountability,” ICIS 2024 Proceedings,no.27,2024.
[4] G. N. Reeke Jr. and G. M. Edelman, “Real brains and artificialintelligence,” Daedalus,vol.117,1988,pp.143–173, [Online]. Available: https://www.jstor.org/stable/20025142.
[5] H. Felzmann, E. F. Villaronga, C. Lutz, and A. TamoLarrieux, “Transparency you can trust: Transparency requirementsforartificialintelligencebetweenlegalnorms and contextual concerns,” Manuscript submitted for publication,2024.
[6]C.Novelli,M.Taddeo,andL.Floridi,“Accountability in artificialintelligence:Whatitisandhowitworks,” Journalof Legal Studies, University of Bologna,vol.27,2023,pp.401–418.
[7]Y.EllameyandA.Elwakad,“Thecriminalresponsibilityof artificial intelligence systems: A prospective analytical study,” Corporate Law & Governance Review, vol. 5, no. 1, 2023,pp.92–100,doi:10.22495/clgrv5i1p8.
[8] J.-C. Bélisle-Pipon, E. Monteferrante, M.-C. Roy, and V. Couture, “Artificial intelligence ethics has a black box problem,” AI & Society,vol.38,no.6,2022,pp.1507–1522.
[9] M. Busuioc, “Accountable artificial intelligence,” Public Administration Review, vol. 81, no. 5, 2021, pp. 825–836, doi:10.1111/puar.13437.
[10]N.Diakopoulos,“Accountabilityinalgorithmicdecision making,” Communications of theACM,vol.59,no.2,2016,pp. 56–62.
[11]K.E.Martin,“Ethicalimplicationsandaccountabilityof algorithms,” Journal of Business Ethics,vol.160,no.4,2018, pp.833–850.
[12]B.D.Mittelstadt,P.Allo,M.Taddeo,S.Wachter,andL. Floridi,“Theethicsofalgorithms:Mappingthedebate,” Big Data & Society, vol. 3, no. 2, 2016, pp. 1–21, doi:10.1177/2053951716679679.