Skip navigation

A dynamic contextual responsibility framework for evaluating large language models in socio-technical contexts

A dynamic contextual responsibility framework for evaluating large language models in socio-technical contexts

Ibitoye, Ayodeji Olusegun ORCID logoORCID: https://orcid.org/0000-0002-5631-8507, Johnson, Deepika Roselind, Sorinolu, Babafemi Gabriel, Orji, Rita and Abikoye, Oluwakemi Christiana (2026) A dynamic contextual responsibility framework for evaluating large language models in socio-technical contexts. AI and Ethics, 6 (191). ISSN 2730-5961 (Online) (doi:10.1007/s43681-026-01072-9)

[thumbnail of Open Access Article]
Preview
PDF (Open Access Article)
52669 IBITOYE_A_Dynamic_Contextual_Responsibility_Framework_For_Evaluating_Large_Language_Models_(OA)_2026.pdf - Published Version
Available under License Creative Commons Attribution.

Download (1MB) | Preview

Abstract

Current Responsible AI metrics, including truthfulness, bias, and toxicity scores, often reduce responsibility in large language models (LLMs) to static technical proxies, obscuring the contextual, ethical, and temporal dynamics through which accountability is produced in real-world settings. This study introduces Dynamic Contextual Responsibility (DCR), a conceptual and operational framework that defines responsibility as a dynamic, context-conditioned, and socio-technical relation shaped by system behaviour, governance arrangements, and institutional norms. DCR integrates five dimensions, ethical foundations, contextual grounding, behavioural properties, governance mechanisms, and temporal dynamics, into a unified and interpretable construct. To illustrate its operational implications, the framework is examined through multi-model, multi-context, and multi-temporal evaluations using established benchmarks such as TruthfulQA, FEVER, and HotpotQA. The analysis shows that approximately 22% of outputs classified as responsible under static metrics are reclassified once contextual and temporal factors are considered, revealing latent ethical and governance risks. By foregrounding context, governance, and temporal change, DCR advances Responsible AI evaluation toward more dynamic, transparent, and plural forms of accountability, with direct relevance for emerging regulatory regimes, including the EU AI Act and the NIST AI Risk Management Framework.

Item Type: Article
Uncontrolled Keywords: responsible AI, contextual responsibility, dynamic accountability, algorithmic governance, LLM risk calibration, socio-technical evaluation
Subjects: Q Science > Q Science (General)
Q Science > QA Mathematics
Q Science > QA Mathematics > QA75 Electronic computers. Computer science
Faculty / School / Research Centre / Research Group: Faculty of Engineering & Science
Faculty of Engineering & Science > School of Computing & Mathematical Sciences (CMS)
Last Modified: 18 Mar 2026 10:25
URI: https://gala.gre.ac.uk/id/eprint/52669

Actions (login required)

View Item View Item

Downloads

Downloads per month over past year

View more statistics