Computer vision technologies are being developed and deployed in a variety of socially consequential domains including policing, employment, healthcare and more. In this sense, computer vision has ceased to be a purely academic endeavor, but rather one that impacts people around the globe on a daily basis. Yet, despite the rapid development of new algorithmic methods, advancing state-of-the-art numbers of standard benchmarks, and, more generally, the narrative of progress that surround public discourses about computer vision, the reality of how computer vision tools are impacting people paints a far darker picture.
In practice, computer vision systems are being weaponized against already over-surveilled and over-policed communities (Garvie et al. 2016; Joseph and Lipp, 2018; Levy, 2018), perpetuating discriminatory employment practices (Ajunwa, 2020), and more broadly, reinforcing systems of domination through the patterns of inclusion and exclusion operative in standard datasets, research practices, and institutional structures within the field (West, 2019; Miceli et al., 2020; Prabhu and Birhane, 2020). In short, the benefits and risks of computer vision technologies are unevenly distributed across society, with the harms falling disproportionately on marginalized communities.
In response, a critical public discourse surrounding the use of computer-vision based technologies has also been mounting (Wakabayashi and Shane, 2018; Frenkel, 2018). For example, the use of facial recognition technologies by policing agencies has been heavily critiqued (Stark, 2019; Garvie, 2019) and, in response, companies such as Microsoft, Amazon, and IBM have pulled or paused their facial recognition software services (Allyn, 2020a; Allyn, 2020b). Buolamwini and Gebru (2018) showed that commercial gender classification systems have high disparities in error rates by skin-type and gender, Hamidi et al., (2018) discusses the harms caused by the mere existence of automatic gender recognition systems, and Prabhu and Birhane (2020) demonstrate shockingly racist and sexist labels in popular computer vision datasets--resulting in the removal of datasets such as Tiny Images (Torralba et al., 2008). Policy makers and other legislators have cited some of these seminal works in their calls to investigate unregulated usage of computer vision systems (Garvie, 2019).
Outside the computer vision community, we observe the broader field of AI beginning to grapple with the impacts of the the technologies that are being developed and deployed in practice: entire interdisciplinary conferences have emerged to study societal impacts of machine learning tools (e.g. FAccT*, AIES); prominent NLP conferences have integrated ethics tracks into their submission processes and regularly host ethics-oriented workshops and tutorials; interdisciplinary, ethics-oriented, and critical workshops have been introduced into major machine learning conferences (e.g. NeurIPS, ICML, ICLR). Meanwhile, despite growing concerns about the development and use of computer vision technologies, the computer vision research community has been slower to engage with these issues.
The computer vision community has not been unchanged however. Indeed, we observe an increased focus on fairness-oriented methods of model and dataset development. However, much of this work is constrained by a purely technical understanding of fairness -- an understanding that has come to mean parity of model performance across sociodemographic groups -- that offers a narrow way of understanding how computer vision technologies intersect with systems of oppression that structure their development and use in the real world (Selbst et al., 2019; Gangadharan, 2020). In contrast to this approach, we believe it is essential to approach computer vision technologies from a sociotechnical lens, examining how marginalized communities are excluded from their development and impacted by their deployment.
Our workshop will also center the perspectives and stories of communities who have been harmed by computer vision technologies and the dominant logics operative within this field. We believe it is important to host these conversations from within the CVPR venue so that researchers and practitioners within the computer vision field can engage with these perspectives and understand the lived realities of marginalized communities impacted by the outputs of the field. In doing so, we hope to shift the focus away from singular technical understandings of fairness and towards justice, equity, and accountability. We believe this is a critical moment for computer vision practitioners and for the field as a whole to come together and reimagine what this field might look like. We have great faith in the CVPR community and believe our workshop will foster the difficult conversations and meaningful reflection upon the state of the field that is essential to begin constructing a different mode of operating. In doing so, our workshop will be a critical step towards ensuring this field progresses in an equitable, just, and accountable manner.