Human-Centric AI (HCAI) designs or develops artificial intelligence (AI) systems that incorporate human needs, values, and collaborative relationships with AI. HCAI differs from traditional AI, which works somewhat autonomously or with minimum human involvement, with the human component as its defining feature and the basis for adding interpretability. The presented paradigm is then applied to various areas, including healthcare, finance, autonomous systems, and cybersecurity, because it helps enable collaborative work between humans and AI to optimize efficiency, safety, and trust. To successfully collaborate with humans, such systems must operate in complex issues and follow established guidelines, ethical considerations, and regulatory frameworks. Using scholarly and industrial sources to understand the field, this paper covers key aspects of HAI, including the best practices for interaction of the human-AI, factors for trust building, regulatory compliance, business scalability, and emerging challenges.
References
[1]
Ozmen Garibay, O., Winslow, B., Andolina, S., Antona, M., Bodenschatz, A., Coursaris, C., etal. (2023) Six Human-Centered Artificial Intelligence Grand Challenges. InternationalJournalofHuman-ComputerInteraction, 39, 391-437. https://doi.org/10.1080/10447318.2022.2153320
[2]
Cheong, B.C. (2024) Transparency and Accountability in AI Systems: Safeguarding Wellbeing in the Age of Algorithmic Decision-Making. FrontiersinHumanDynamics, 6, Article 1421273. https://doi.org/10.3389/fhumd.2024.1421273
[3]
Singhal, A., Neveditsin, N., Tanveer, H. and Mago, V. (2024) Toward Fairness, Accountability, Transparency, and Ethics in AI for Social Media and Health Care: Scoping Review. JMIRMedicalInformatics, 12, e50048. https://doi.org/10.2196/50048
[4]
Abbasi, S. and Rahmani, A.M. (2023) Artificial Intelligence and Software Modeling Approaches in Autonomous Vehicles for Safety Management: A Systematic Review. Information, 14, Article 555. https://doi.org/10.3390/info14100555
[5]
Afroogh, S., Akbari, A., Malone, E., Kargar, M. and Alambeigi, H. (2024) Trust in AI: Progress, Challenges, and Future Directions. HumanitiesandSocialSciencesCommunications, 11, Article No. 1568. https://doi.org/10.1057/s41599-024-04044-8
[6]
McKinsey & Company (2023) The State of AI in 2023: Generative AI’s Breakout Year.
The European Commission (2025) The Commission Publishes Guidelines on the Definition of AI Systems to Facilitate the Application of the First AI Act’s Rules. Shaping Europe’s Digital Future. https://digital-strategy.ec.europa.eu/en/library/commission-publishes-guidelines-ai-system-definition-facilitate-first-ai-acts-rules-application
[9]
Ryan, M. and Stahl, B.C. (2020) Artificial Intelligence Ethics Guidelines for Developers and Users: Clarifying Their Content and Normative Implications. JournalofInformation, CommunicationandEthicsinSociety, 19, 61-86. https://doi.org/10.1108/jices-12-2019-0138
[10]
Singh, V. (2024) AI Has Arrived and Is Vital to Business Scalability. IEEE Computer Society. https://www.computer.org/publications/tech-news/trends/ai-for-business-scalability
[11]
Amershi, S., Weld, D., Vorvoreanu, M., Fourney, A., Nushi, B., Collisson, P., etal. (2019) Guidelines for Human-AI Interaction. Proceedingsofthe 2019 CHIConferenceonHumanFactorsinComputingSystems, Glasgow, 4-9 May 2019, 1-13. https://doi.org/10.1145/3290605.3300233
[12]
Hancock, P.A., Billings, D.R., Schaefer, K.E., Chen, J.Y., De Visser, E.J. and Parasuraman, R. (2020) A Meta-Analysis of Factors Affecting Trust in Human-Robot Interaction. HumanFactors, 62, 134-149.