In high stakes organizational contexts like healthcare, artificial intelligence (AI) systems are increasingly being designed to augment complex coordination tasks. This paper investigates how the ethical stakes of such systems are shaped by their epistemic framings: what aspects of work they represent, and what they exclude. Drawing on an embedded study of AI development for operating room (OR) scheduling at a Canadian hospital, we compare scheduling-as-imagined in the AI design process: rule-bound, predictable, and surgeon-centric, with scheduling-as-practiced as a fluid, patient-facing coordination process involving ethical discretion. We show how early representational decisions narrowed what the AI could support, resulting in epistemic foreclosure: the premature exclusion of key ethical dimensions from system design. Our findings surface the moral consequences of abstraction and call for a more situated approach to designing healthcare process-specialized artificial intelligence systems.
翻译:暂无翻译