In AI We Trust. Helga Nowotny

Чтение книги онлайн.

Читать онлайн книгу In AI We Trust - Helga Nowotny страница 9

In AI We Trust - Helga  Nowotny

Скачать книгу

with access to an unprecedented and still growing amount of data, these developments have extended the power of predictions and their applicability across an enormous range of natural and social phenomena. Scientific predictions are no longer confined to science.

      Much of their successful spread and eager adoption is due to the fact that the power of predictive algorithms is performative. An algorithm has the capability to make happen what it predicts when human behaviour follows the prediction. Performativity means that what is enacted, pronounced or performed can affect action, as shown in the pioneering work on the performativity of speech acts and non-verbal communication by J. L. Austin, Judith Butler and others. Another well-known social phenomenon is captured in the Thomas theorem – ‘If men define situations as real, they are real in their consequences’ – dating back to 1928 and later reformulated by Robert K. Merton in terms of self-fulfilling prophecy. The time has come to acknowledge what sociologists have long since known and apply it also to predictive algorithms.

      The propensity of people to orient themselves in relation to what others do, especially in unexpected or threatening circumstances, enhances the power of predictive algorithms. It magnifies the illusion of being in control. But if the instrument gains the upper hand over understanding we lose the capacity for critical thinking. We end up trusting the automatic pilot while flying blindly in the fog. There are, however, situations in which it is crucial to deactivate the automatic pilot and exercise our own judgement as to what to do.

      At the same time, distrust of AI creeps in and the concerns grow. Some of them, like the fears about surveillance or the future of work, are well known and widely discussed. Others are not so obvious. When self-fulfilling prophecies begin to proliferate, we risk returning to a deterministic worldview in which the future appears as predetermined and hence closed. The space vital to imagining what could be otherwise begins to shrink. The motivation as well as the ability to stretch the boundaries of imagination is curtailed. To rely purely on the efficiency of prediction obscures the need for understanding why and how. The risk is that everything we treasure about our culture and values will atrophy.

      Moreover, in a world governed by predictive analytics there is neither a place nor any longer the need for accountability. When political power becomes unaccountable to those over whom it is exercised, we risk the destruction of liberal democracy. Accountability rests on a basic understanding of cause and effect. In a democracy, this is framed in legal terms and is an integral part of democratically legitimated institutions. If this is no longer guaranteed, surveillance becomes ubiquitous. Big data gets even bigger and data is acquired without understanding or explanation. We become part of a fine-tuned and interconnected predictive system that is dynamically closed upon itself. The human ability to teach to others what we know and have experienced begins to resemble that of a machine that can teach itself and invent the rules. Machines have neither empathy nor a sense of responsibility. Only humans can be held accountable and only humans have the freedom to take on responsibility.

      Obviously, my journey does not end there. ‘Life can only be understood backwards, but it must be lived forward.’ This quotation from Søren Kierkegaard awaits an interpretation in relation to our movements between online and offline worlds, between the virtual self, the imagined self and the ‘real’ self. How does one live forward under these conditions, given their opportunities and constraints? The quotation implies a disjunction between Life as an abstraction that transcends the personal, and living as the conscious experience that fills every moment of our existence. With the stupendous knowledge we now have about Life in all its diversity, forms and levels, about its origins in the deep past and its continued evolution, is not now the moment to bring this knowledge to bear on how to live forward? The human species has overtaken biological evolution whose product we still are. Science and technology have enabled us to move forward at accelerating speed along the pathways of a cultural evolution that we are increasingly able to shape.

      And yet, here we are, facing a global sustainability crisis with many dire consequences and mounting geopolitical tensions. As I write, we are in the grip of a pandemic, with others to follow if the natural habitats of animals that carry zoonotic viruses capable of spreading to humans continue to be eroded. The deficiencies of our institutions, created in previous centuries and designed to meet challenges different from our own, stare us in the face. The spectre of social unrest and polarized societies has returned, when what is needed is greater social coherence, equality and social justice if we are to escape our current predicament.

      Even the most sophisticated neural networks modelled on a simplified version of the brain can only detect regularities and identify patterns based on data that comes from the past. No causal reasoning is involved, nor does an AI pretend that it is. How can we live forward if we fail to understand Life as it has evolved in the past? Some computer scientists, such as Judea Pearl and others, deplore the absence of any search for cause–effect relationships. ‘Real intelligence’, they argue, involves causal understanding. If AI is to reach such a stage it must be able to reason in a counterfactual way. It is not sufficient merely to fit a curve along an indicated timeline. The past must be opened up in order to understand a sentence like ‘what would have happened if …’. Human agency consists in what we do, but understanding what we did in the past in order to make predictions about the future must always involve the counterfactual that we could have acted differently. In transferring human agency to an AI we must ensure that it has the capacity to ‘know’ this distinction that is basic to human reasoning and understanding (Pearl and Mackenzie 2018).

      The power of algorithms to churn out practical and measurable predictions that are useful in our daily lives – whether in the management of health systems, in automated financial

Скачать книгу