About

The COVID-19 pandemic has gifted video conferencing companies such as Zoom with a vast amount of biometric data to be rendered knowable, translatable and ready for economic exchange – such as faces, voices, and chat scripts. It is however as yet unclear what the explosion in video conferencing software means for the exploitation and monetisation of potentially valuable data for NLP, facial recognition, ML and AI training and other drivers within the digital economy which require urgent critical scrutiny. Frequently we are offered ‘solutions’ to issues around data privacy and security that are based on a loose form of trust, with debate frequently reduced to the prevention of external malicious actors gaining access. This has led to a focus on ‘end-to-end’ encryption – but this still leaves individuals with limited access, control, and verifiability to challenge terms and conditions within. COVID-19 has, however, left us with little choice but to increase the volume of interactions we have in online spaces such as the video-call ‘room’, and with limited agency in how our personal data might be being stored and exploited. This leaves us with significant ethical, privacy, and political concerns.

Zoom Obscura is a project that aims to give agency to the users of newly ubiquitous video conferencing technologies such as Zoom, while still allowing them to participate in online spaces and debates, enabling us to negotiate our own presence and our own value in these new spaces. We aim to do so by bringing together artists, academics, hackers, designers and creative technologists to develop critical interventions that make the problematic workings of these technologies legible to wider audiences while empowering users to experiment with, and control how their personal data (visual, audible, text input) manifest in online spaces. Playing on metaphorical (Kofman 1999) and material concepts of the Camera Obscura, with its inverted images and use of light and shadow, Zoom Obscura addresses these issues by harnessing the critical power of art, design and technology; blending and bringing into tension skills and genres to produce a range of interventions which give users the power to take back some of the agential power from platforms such as Zoom.

Looking to the future, the project asks how can we contest / resist the inevitability of a future structured around video calls, conferences and seminars? Can we push back against the normalisation of the practices we have so quickly and readily adopted in the COVID-19 state of exception? How can we regain control of how our images and words manifest in these spaces? How can we move beyond encryption as a solution to privacy /security problems? Encryption might be a technical fix, and a politically popular (although controversial) narrative, but it doesn’t solve the ethical problems that sit beneath, through, and around its implementation. This project seeks to explore a data ethics beyond encryption and technological solutionism.