2023
DOI: 10.1007/s42803-023-00075-w
|View full text |Cite
|
Sign up to set email alerts
|

On reading and interpreting black box deep neural networks

James E. Dobson

Abstract: The deep neural networks used in computer vision and in recent large language models are widely recognized as black boxes, a term that describes their complicated architectures and opaque decision-making mechanisms. This essay outlines several different strategies through which humanist researchers and critics of machine learning might better understand and interpret the class of deep learning methods known as Transformers. These strategies expose different aspects of what might be “learned” as Transformers ar… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...

Citation Types

0
3
0

Year Published

2024
2024
2025
2025

Publication Types

Select...
5

Relationship

0
5

Authors

Journals

citations
Cited by 10 publications
(3 citation statements)
references
References 26 publications
0
3
0
Order By: Relevance
“…Already in 2017, Doshi-Velez and Kim had put forward the goal of building a "rigorous science of interpretable machine learning" (Doshi-Velez & Kim, 2017). Since then, countless papers and projects have contributed to an ever-growing pool of approaches, aiming at text and image models alike (see (Molnar, 2020), or the contribution by James Dobson in this issue (Dobson, 2023) for an overview of the most popular techniques). At the same time, the question of exposing a model's particular perspective on the world, that is, describing its biases and limitations, is, at its core, a humanities question.…”
mentioning
confidence: 99%
See 2 more Smart Citations
“…Already in 2017, Doshi-Velez and Kim had put forward the goal of building a "rigorous science of interpretable machine learning" (Doshi-Velez & Kim, 2017). Since then, countless papers and projects have contributed to an ever-growing pool of approaches, aiming at text and image models alike (see (Molnar, 2020), or the contribution by James Dobson in this issue (Dobson, 2023) for an overview of the most popular techniques). At the same time, the question of exposing a model's particular perspective on the world, that is, describing its biases and limitations, is, at its core, a humanities question.…”
mentioning
confidence: 99%
“…The contributions to this special issue, together, make a significant leap towards the methodological and epistemological reflection of such methods, a reflection that so far has been missing in the digital humanities context. James Dobson, in "Reading and Interpreting Black Box Deep Neural Networks", gives an in-depth introduction to interpretability and explainability methods from computer science research, arguing for a "research program informed by tool criticism in which the use of computational tools is conceived of as a metainterpretive act" (Dobson, 2023). Tobias Blanke, Tommaso Venturini, and Kari De Pryck demonstrate how political concepts like "leadership", which are difficult to operationalize, can be better approximated through XAI techniques (Blanke et al, 2023).…”
mentioning
confidence: 99%
See 1 more Smart Citation