1. Algorithmic Decision-Making Based on Machine Learning from Big Data: Can Transparency Restore Accountability?
- Author
-
de Laat, Paul B.
- Subjects
Opacity ,Computer science ,As is ,Big data ,02 engineering and technology ,Transparency ,0603 philosophy, ethics and religion ,Machine learning ,computer.software_genre ,Computer security ,Competitive advantage ,History and Philosophy of Science ,0202 electrical engineering, electronic engineering, information engineering ,Interpretability ,Accountability ,Ex-ante ,business.industry ,06 humanities and the arts ,Transparency (behavior) ,Algorithm ,Philosophy ,020201 artificial intelligence & image processing ,060301 applied ethics ,Artificial intelligence ,business ,computer ,Research Article - Abstract
Decision-making assisted by algorithms developed by machine learning is increasingly determining our lives. Unfortunately, full opacity about the process is the norm. Would transparency contribute to restoring accountability for such systems as is often maintained? Several objections to full transparency are examined: the loss of privacy when datasets become public, the perverse effects of disclosure of the very algorithms themselves ("gaming the system" in particular), the potential loss of companies' competitive edge, and the limited gains in answerability to be expected since sophisticated algorithms usually are inherently opaque. It is concluded that, at least presently, full transparency for oversight bodies alone is the only feasible option; extending it to the public at large is normally not advisable. Moreover, it is argued that algorithmic decisions preferably should become more understandable; to that effect, the models of machine learning to be employed should either be interpreted ex post or be interpretable by design ex ante.
- Published
- 2017