A dynamic contextual responsibility framework for evaluating large language models in socio-technical contexts
Ibitoye, Ayodeji Olusegun ORCID: https://orcid.org/0000-0002-5631-8507, Johnson, Deepika Roselind, Sorinolu, Babafemi Gabriel, Orji, Rita and Abikoye, Oluwakemi Christiana
(2026)
A dynamic contextual responsibility framework for evaluating large language models in socio-technical contexts.
AI and Ethics, 6 (191).
ISSN 2730-5961 (Online)
(doi:10.1007/s43681-026-01072-9)
Preview |
PDF (Open Access Article)
52669 IBITOYE_A_Dynamic_Contextual_Responsibility_Framework_For_Evaluating_Large_Language_Models_(OA)_2026.pdf - Published Version Available under License Creative Commons Attribution. Download (1MB) | Preview |
Abstract
Current Responsible AI metrics, including truthfulness, bias, and toxicity scores, often reduce responsibility in large language models (LLMs) to static technical proxies, obscuring the contextual, ethical, and temporal dynamics through which accountability is produced in real-world settings. This study introduces Dynamic Contextual Responsibility (DCR), a conceptual and operational framework that defines responsibility as a dynamic, context-conditioned, and socio-technical relation shaped by system behaviour, governance arrangements, and institutional norms. DCR integrates five dimensions, ethical foundations, contextual grounding, behavioural properties, governance mechanisms, and temporal dynamics, into a unified and interpretable construct. To illustrate its operational implications, the framework is examined through multi-model, multi-context, and multi-temporal evaluations using established benchmarks such as TruthfulQA, FEVER, and HotpotQA. The analysis shows that approximately 22% of outputs classified as responsible under static metrics are reclassified once contextual and temporal factors are considered, revealing latent ethical and governance risks. By foregrounding context, governance, and temporal change, DCR advances Responsible AI evaluation toward more dynamic, transparent, and plural forms of accountability, with direct relevance for emerging regulatory regimes, including the EU AI Act and the NIST AI Risk Management Framework.
| Item Type: | Article |
|---|---|
| Uncontrolled Keywords: | responsible AI, contextual responsibility, dynamic accountability, algorithmic governance, LLM risk calibration, socio-technical evaluation |
| Subjects: | Q Science > Q Science (General) Q Science > QA Mathematics Q Science > QA Mathematics > QA75 Electronic computers. Computer science |
| Faculty / School / Research Centre / Research Group: | Faculty of Engineering & Science Faculty of Engineering & Science > School of Computing & Mathematical Sciences (CMS) |
| Last Modified: | 18 Mar 2026 10:25 |
| URI: | https://gala.gre.ac.uk/id/eprint/52669 |
Actions (login required)
![]() |
View Item |
Downloads
Downloads per month over past year
Tools
Tools