Machine Habitus. Massimo Airoldi
Чтение книги онлайн.
Читать онлайн книгу Machine Habitus - Massimo Airoldi страница 7
![Machine Habitus - Massimo Airoldi Machine Habitus - Massimo Airoldi](/cover_pre1045657.jpg)
In 1989, Lenat noted in the pages of the journal Machine Learning that ‘human-scale learning demands a human-scale amount of knowledge’ (1989: 255), which was not yet available to AI researchers at the time. An impressive advancement of machine learning methods occurred two decades later, thanks to a ‘fundamental socio-technological transformation of the relationship between humans and machines’, consisting in the capturing of human cognitive abilities through the digital accumulation of data (Mühlhoff 2020: 1868). This paradigmatic change has made the ubiquitous automation of social and cultural tasks suddenly possible on an unprecedented scale. What matters here sociologically is ‘not what happens in the machine’s artificial brain, but what the machine tells its users and the consequences of this’ (Esposito 2017: 250). According to Esposito, thanks to the novel cultural and communicative capabilities developed by ‘parasitically’ taking advantage of human-generated online data, algorithms have substantially turned into ‘social agents’.
Recent accomplishments in AI research – such as AlphaGo, the deep learning system that achieved a historic win against the world champion of the board game Go in 2016 (Chen 2016; Broussard 2018), or GPT-3, a powerful algorithmic model released in 2020, capable of autonomously writing poems, computer code and even philosophical texts (Weinberg 2020; Askell 2020) – indicate that the ongoing shift toward the increasingly active and autonomous participation of algorithmic systems in the social world is likely to continue into the near future. But let’s have a look at the past first.
Algorithms and their applications, from Euclid to AlphaGo
The term ‘algorithm’ is believed to derive from the French bastardization of the name of the ninth-century Persian mathematician al-Khwārizmī, the author of the oldest known work of algebra. Being originally employed in medieval Western Europe to indicate the novel calculation methods alternative to those based on Roman numerals, in more recent times the term has come to mean ‘any process of systematic calculation […] that could be carried out automatically’ (Chabert 1999: 2). As Chabert remarks in his book A History of the Algorithm: ‘algorithms have been around since the beginning of time and existed well before a special word had been coined to describe them’ (1999: 1). Euclid’s algorithm for determining the greatest common divisor of two integers, known since the fourth century BCE, is one of the earliest examples.
More generally, algorithms can be intended as computational recipes, that is, step-by-step instructions for transforming input data into a desired output (Gillespie 2014). According to Gillespie (2016: 19), algorithms are essentially operationalized procedures that must be distinguished from both their underlying ‘model’ – the ‘formalization of a problem and its goal, articulated in computational terms’ – and their final context of application, such as the technical infrastructure of a social media platform like Facebook, where sets of algorithms are used to allocate personalized content and ads in users’ feeds. Using a gastronomic metaphor, the step-by-step procedure for cooking an apple pie is the algorithm, the cookbook recipe works as the model, and the kitchen represents the application context. However, in current public and academic discourse, these different components and meanings tend to be conflated, and the term algorithm is broadly employed as a synecdoche for a ‘complex socio-technical assemblage’ (Gillespie 2016: 22).
‘Algorithm’ is thus a slippery umbrella term, which may refer to different things (Seaver 2017). There are many kinds of computational recipes, which vary based on their realms of application as well as on the specific ‘algorithmic techniques’ employed to order information and process data (Rieder 2020). A single task, such as classifying texts by topic, may concern domains as diverse as email ‘spam’ filtering, online content moderation, product recommendation, behavioural targeting, credit scoring, financial trading and more – all of which involve a plethora of possible input data and outputs. Furthermore, text classification tasks can be executed in several – yet all ‘algorithmic’ – ways: by hand, with pen and paper only; through rule-following software applying models predefined by human programmers (e.g. counting topic-related word occurrences within texts); or via ‘intelligent’ machine learning systems that are not explicitly programmed a priori. These latter can be either supervised – i.e. requiring a preliminary training process based on data examples, as in the case of naive Bayes text classifiers (Rieder 2017) – or unsupervised, that is, machine learning techniques working without pre-assigned outputs, like Latent Dirichlet Allocation in the field of topic modeling (Bechmann and Bowker 2019).
This book does not aim to offer heavily technical definitions, nor an introduction to algorithm design and AI technologies; the reader can easily find such notions elsewhere.2 Throughout the text, I will frequently make use of the generic terms ‘algorithm’ and ‘machine’ to broadly indicate automated systems producing outputs based on the computational elaboration of input data. However, in order to highlight the sociological relevance of the quali-quantitative transition from Euclid’s calculations to today’s seemingly ‘intelligent’ artificial agents like GPT-3 and AlphaGo, some preliminary conceptual distinctions are needed. It is apparent, in fact, that the everyday socio-cultural implications of algebraic formulas solved for centuries by hand or via mechanical calculators are not even close in magnitude to those of the algorithms currently governing information networks.
Below I briefly outline the history of algorithms and their applications – from ancient algebra to rule-following models running on digital computers, and beyond to platform-based machine learning systems. This socio-technical evolution can be roughly broken into three main eras, visually summarized in Figure 1 at the end of this section. Without pretending to be exhaustive, the proposed periodization focuses especially on the emergence of ‘public relevance algorithms’ (Gillespie 2014: 168), that is, automated systems dealing with the social matter of human knowledge, experience and practice.
Analogue Era (–1945)
Taking analogue to mean ‘not-digital’ (Sterne 2016), this first historical phase ranges in principle from the invention and manual application of algorithms by ancient mathematicians to the realization of the first digital computers right after the Second World War. Within this period, algorithms were applied either by human-supervised mechanical devices or by humans themselves (Pasquinelli 2017). In fact, up until the early twentieth century, the word ‘computer’ indicated a person employed to make calculations by hand. Mechanical computers started to be conceptualized at the beginning of the nineteenth century, following Leibniz’s early intuitions about the mechanization of calculus (Chabert 1999), as well as a rising demand for faster and more reliable calculations from companies and governments (Wilson 2018; Campbell-Kelly et al. 2013). Aiming to automatize the compilation of tables for navigation at sea, particularly strategic for the British Empire, in the 1820s the mathematician Charles Babbage designed the first mechanical computer, the Difference Engine, which was then followed by the more ambitious Analytical Engine – ideally capable of performing ‘any calculation that a human could specify for it’ (Campbell-Kelly et al. 2013: 8). Babbage’s proto-computers were pioneering scientific projects that remained largely on paper, but more concrete applications of simpler electro-mechanical ‘algorithm machines’ (Gillespie 2014) came to light by the end of the century. In 1890, Hollerith’s electric tabulating system was successfully employed to process US census data, paving the way for the foundation of IBM. Thanks to the punched-card machines designed by Hollerith, information on over 62 million American citizens was processed within ‘only’ two and a half years, compared with the seven years taken by the previous census, with an estimated saving of 5 million dollars (Campbell-Kelly et al. 2013: 17–18). The mass production of desk calculators and business accounting machines brought algorithms closer to ordinary people’s everyday routines. Still, information was computationally transformed and elaborated solely through analogue means (e.g. punched cards, paper tapes) and under human supervision.