Dados Bibliográficos

AUTOR(ES) Florian Jaton
AFILIAÇÃO(ÕES) STS Lab, University of Lausanne, Lausanne, Switzerland
ANO 2021
TIPO Artigo
PERIÓDICO Big Data & Society
ISSN 2053-9517
E-ISSN 2053-9517
DOI 10.1177/20539517211013569
CITAÇÕES 9
ADICIONADO EM 2025-08-18

Resumo

This theoretical paper considers the morality of machine learning algorithms and systems in the light of the biases that ground their correctness. It begins by presenting biases not as a priori negative entities but as contingent external referents—often gathered in benchmarked repositories called ground-truth datasets—that define what needs to be learned and allow for performance measures. I then argue that ground-truth datasets and their concomitant practices—that fundamentally involve establishing biases to enable learning procedures—can be described by their respective morality, here defined as the more or less accounted experience of hesitation when faced with what pragmatist philosopher William James called 'genuine options'—that is, choices to be made in the heat of the moment that engage different possible futures. I then stress three constitutive dimensions of this pragmatist morality, as far as ground-truthing practices are concerned: (I) the definition of the problem to be solved (problematization), (II) the identification of the data to be collected and set up (databasing), and (III) the qualification of the targets to be learned (labeling). I finally suggest that this three-dimensional conceptual space can be used to map machine learning algorithmic projects in terms of the morality of their respective and constitutive ground-truthing practices. Such techno-moral graphs may, in turn, serve as equipment for greater governance of machine learning algorithms and systems.

Ferramentas