- From: Paola Di Maio <paola.dimaio@gmail.com>
- Date: Sat, 1 Aug 2020 10:31:22 +0800
- To: W3C AIKR CG <public-aikr@w3.org>
- Message-ID: <CAMXe=SrX2tdx7FHXcxCLU54=xD4MVeSQP=3Ys7nX76G_0eEaBg@mail.gmail.com>
One of the triggers for my interest in AI KR has been a certain divergence I observed between reality and the labels, These observations then spun the interest into bias, system integrity, cognitive dissonance, schizophrenia as briefly mentioned in various threads. KR is key to solve this divergences, although it comes with challenges. I received this article via a tweet (credit to FP) and managed to access the article below, which I think is relevant albeit behind paywall I paste the concluding remarks The return of the social: Algorithmic identity in an age of symbolic demise Dan M. Kotliar <https://journals.sagepub.com/doi/abs/10.1177/1461444820912535?casa_token=SPnD5M0aiAYAAAAA%3ARant3XSJzRQm-9enJOgz0P-HCjaYPE4TMpqz7J8IdFZhDFMP5gKZujH_3NkMlgt5ou9VgQjcGU8UPg&journalCode=nmsa> <https://orcid.org/0000-0001-7028-1678> First Published July 22, 2020 Research Article <https://crossmark.crossref.org/dialog?doi=10.1177%2F1461444820912535&domain=journals.sagepub.com&uri_scheme=https%3A&cm_version=v2.0> https://doi.org/10.1177/1461444820912535 The return of the social: Algorithmic identity in an age of symbolic demise Dan M. Kotliar Stanford University, USA and The Hebrew University of Jerusalem, Israel Abstract This article explores the socio-algorithmic construction of identity categories based on an ethnographic study of the Israeli data analytics industry. While algorithmic categorization has been described as a post-textual phenomenon that leaves language, social theory, and social expertise behind, this article focuses on the return of the social—the process through which the symbolic means resurface to turn algorithmically produced clusters into identity categories. I show that such categories stem not only from algorithms’ structure or their data, but from the social contexts from which they arise, and from the values assigned to them by various individuals. I accordingly argue that algorithmic identities stem from epistemic amalgams—complex blends of algorithmic outputs and human expertise, messy data flows, and diverse inter-personal factors. Finally, I show that this process of amalgamation arbitrarily conjoins quantitative clusters with qualitative labels, and I discuss the implausibility of seeing named algorithmic categories as explainable ..........article behind paywall Conclusion Data-mining algorithms can indeed create unnamed, incomprehensible categories, and in that, they may disregard language, social theory, and social expertise. But such obscure algorithmic outputs often get re-socialized. The old epistemic devices, the traditional symbolic means, get re-introduced in response to people’s values and needs. That is, while these categories are automatically created by algorithms (according to the internal algorithmic structure, and the types of data they run on), the categories’ names, and in effect, their perceived meanings, are often added after the fact, in light of various social factors. Hence, the creation of identity categories entails at least two stages: an algorithmic, calculative one—which includes the creation of opaque, quantified, clusters; and a qualitative, communicative one—which re-introduces language, social theory, and social expertise into the equation. Identity categories may stem from an algorithmic analysis of some datafied part of society, but their names often originate from, and are similarly designed as, something different—as an attempt to answer to the economic, occupational, or even 1162 new media & society 22(7) emotional needs of the people who buy and use such systems. In other words, algorithmic identities are more than “measurable types” (Cheney-Lippold, 2017: 47), and algorithmic categorization involves more than the calculative version of reality. The identities that these algorithms create are “qualculated” (Cochoy, 2008), as much as they are calculated. They stem from automated mathematical analyses, but they also echo the inter-organizational relationships, intra-organizational tensions, and the idiosyncratic values of the people who develop, buy, and use such algorithms. Thus, the creation of algorithmic identity categories does not signal a complete epistemic turn, nor an absolute symbolic demise. Instead, these categories are created from what I term epistemic amalgams—complex blends of algorithmic outputs and human values, messy data flows, and traditional social labels. The “social” returns, but it is not a re-emergence of the traditional disciplinary powers, but their reinvention, their complex amalgamation with other epistemes. Hence, a user may be described as an “introvert,” but this description is not based on his or her psychological evaluation. Instead, these labels are primarily aimed at evoking certain responses in certain cultural actors. And thus, even if users can see the categories they have been put into, and even if these categories seem familiar, users cannot know how companies name their categories, or how many likes, to which kinds of cars, turn them into “introverts.” Language returns, but the gap between the lingual signifier and its algorithmically formed signified is wider than ever before. Therefore, while the return of the social may seem like a ray of light into algorithmic black boxes, and while named categories may be presented as enhancing the explainability of algorithmic outputs, it is, in fact, an epistemic mirage, a simulacrum of sorts (Baudrillard, 1994). The qualitative stages behind these categories are just as opaque, and just as black boxed, as the calculative ones. Nevertheless, the amalgamated structure of these categories, and the fact that they remain opaque, does not make them meaningless, nor powerless. Thornton (2018), following Kaplan (2014), recently described the current era as an era of “linguistic capitalism”—one in which words get commodified and auctioned by online advertising services, such as Google AdWords. In this case, however, it is not a “commodification of words” (Kaplan, 2014: 62), but a commodification through words—a monetization made through re-symbolization. That is, while the commodification of language strips language from its social context, and deprives it of much of its symbolic value, the process I describe here does the exact opposite—it turns the algorithmic clusters into social categories, thus bringing them back into the social domain. As Dourish and Gómez Cruz (2018) recently argued, data need to be narrated in order to work. In this case, identity categories get commodified only after they get re-socialized, only after a subjective, symbolic value gets re-attached to them. Thus, while data can indeed form new, digital, types of orality (Papacharissi, 2015), it seems like human language is still an essential component in contemporary capitalism, in turning symbolic value into economic value— an epistemic bridge between quantification and monetization. Similarly, while language, theory, and expertise may be superfluous to some algorithmic systems, they are still essential components for the functioning of algorithmic powers, and for their gradual instilment into society. This work contributes to our understanding of algorithms and their power, as well as to our conceptualization of identity in an algorithmic age. As algorithmic profiling enters almost every social realm, it is crucial that we better our understanding of the Kotliar 1163 socio-cultural construction of such profiles and the complex epistemic structures behind such categories. Moreover, the account presented here contributes to the growing literature on algorithms and culture (Christin, 2018; Dourish, 2016; Ribak, 2019; Seaver, 2017), by illustrating how people’s values and needs, and their specific social position, affect the ways people are algorithmically described. Finally, while the research on algorithmic identity has mostly dealt with the datafication of reality, this article calls for a scholarly focus on the re-socialization of datafied life. Seeing algorithmization as a multi-stage process, and particularly focusing on its qualitative stages, could help identify the meeting points between algorithms and society, and the exact circumstances in which algorithmic outputs are culturally shaped
Received on Saturday, 1 August 2020 02:32:22 UTC