Abstract
This article situates contemporary critical media literacy into a postdigital context. It examines recent advances in data literacy, with an accent to Big Data literacy and data bias, and expands them with insights from critical algorithm studies and the critical posthumanist perspective to education. The article briefly outlines differences between older software technologies and artificial intelligence (AI), and introduces associated concepts such as machine learning, neural networks, deep learning, and AI bias. Finally, it explores the complex interplay between Big Data and AI and teases out three urgent challenges for postdigital critical media literacy. (1) Critical media literacy needs to reinvent existing theories and practices for the postdigital context. (2) Reinvented theories and practices need to find a new balance between the technological aspects of data and AI literacy with the political aspects of data and AI literacy, and learn how to deal with non-predictability. (3) Critical media literacy needs to embrace the posthumanist challenge; we also need to start thinking what makes AIs literate and develop ways of raising literate thinking machines. In our postdigital age, critical media literacy has a crucial role in conceptualisation, development, and understanding of new forms of intelligence we would like to live with in the future.
1 Introduction
This morning, during my regular coffee and newspaper ritual, I read an interesting article: ‘Amazon scraps secret AI recruiting tool that showed bias against women’ (Dastin 2018). In one of its development centres, Amazon’s team of experts has designed an Artificial Intelligence (AI) software ‘to review job applicants’ resumes with the aim of mechanizing the search for top talent’ and ‘taught’ the software using 10-year archive of recruitment at Amazon. After it started working, at the surprise of researchers, the software showed strong bias against women. To resolve the problem, ‘Amazon edited the programs to make them neutral to these particular terms. But that was no guarantee that the machines would not devise other ways of sorting candidates that could prove discriminatory’. Therefore, they scraped the AI completely (Dastin 2018). This borderline sensationalist story, which seemed to amuse to death a group of freshers at the next table, carries a deep message. These days our society is increasingly reliant on collecting large amounts of data (the so-called Big Data) and processing this data using automated systems of various hues and shapes. However, probably for the first time in history, these systems (which are often called AIs), function in ways which cannot be predicted even by their designers and makers. After they designed and ‘taught’ their AI, Amazon’s engineers could not predict its later behaviour or even avoid (perhaps a different type) of bias. Amazon AI’s bias against women is not a technical glitch, or even an error in design – AI’s ‘independent mind’ is a feature built in the very essence of its workings.
A television set may reproduce many programs, but the choice of program is firmly in the hands of the viewer; during commercial breaks, everyone watches the same advertisements. Streaming services such as YouTube use recommendation systems to direct us towards watching certain content and offer personalized advertisements, yet we can always choose to watch something else. Critical media literacy is important, because it helps us navigate and produce these media. However, once we know what we would like to do, we are technically in full control. As viewers, we can change TV program, delete our browsing history, open a fresh YouTube account, etc., and escape the long arm of recommender systems. As producers, we can make a video, upload it to YouTube and make sure that it will show exactly what we wanted. However, Amazon’s AI has gone one significant step further – shaped by people, ‘taught’ by data about people, it makes ‘own’ choices and decisions. If YouTube videos worked similarly to Amazon’s AI recruitment tool, they would self-edit after we upload them and use recommender systems to offer themselves to audiences which they find appropriate. When similar systems are put in place, say, for determining our credit scores or health insurance, then AI’s get to make significant impact to human lives. This is the essence of what Jeremy Knox calls the age of algorithmic cultures, ‘in which automated computer operations process data in such a way as to significantly shape the contemporary categorizing and privileging of knowledge, places, and people’ (Knox 2015: 5).
a clear rejection of scientific realism substituting a relation process ontology that points towards a indeterministic universe at the sub-atomic level and a form of quantum philosophy based on quantum mechanics and computing characterizing an era we are just entering. It will be transformative, dynamic, system-built ontology very different from our understanding of the digital, which itself has only got underway. A critical philosophy of the postdigital must be able to understand the processes of quantum computing, complexity science, and deep learning as they constitute the emerging techno-science global system and its place within a capitalist system that itself is transformed by these developments.
(Peters and Besley 2018)
Critical media literacy addresses some of these challenges in rapidly developing areas such as data literacy (D’Ignazio and Bhargav 2015; Koltay 2015, just to mention a few). However, functioning of computer systems is based on two equally important pillars: input data and system architecture. This article examines recent advances in data literacy with an accent to Big Data literacy. It moves on to internal workings of contemporary AI’s and associated concepts such as machine learning, neural networks, and deep learning. Finally, it explores the complex interplay between Big Data and AI and teases out some challenges for postdigital critical media literacy.
2 Data Literacy
Data literacy is a relative newcomer in literacy studies; these days, it is ‘in search of a name and identity’ (Koltay 2014: 401). Data literacy ‘cuts across disciplinary boundaries’ and traditional workplace roles; in the academia, for instance, ‘it is often difficult to separate data-related skills needed to become a successful researcher and to work as a data specialist’ (Koltay 2014: 404). In the context of critical media literacy D’Ignazio and Bhargav argue that ‘Big Data has an empowerment problem’ because of its four main characteristics:
- – Lack of Transparency: The data about people’s interactions with the world is generally collected with only token approval, if any at all, from the user. This denies the subject awareness that their actions are being recorded at the time the actions occur.
- – Extractive Collection: The data is collected by third parties and is not meant for observation or consumption by the people it is collected from (or about). This denies the subject agency in the data collection mechanism and interaction opportunities with the collector.
- – Technological Complexity: The data is analyzed with a variety of advanced algorithmic techniques, and discussed with highly technical jargon. This denies the subject an understanding of how any results were achieved, and how they might be critiqued.
- – Control of Impact: The data is used by the collector to make decisions that have consequences for the subject(s). This denies the subject participation in decisions that affect them. (D’Ignazio and Bhargava, 2015, emphasis from the original)
This list can be expanded by recent advances in educational research. Standardized tests, which are now routinely used in various contexts from admissions to final exams, produce large datasets about students – and these datasets suffer from various biases. Looking at ‘the role of various psychometric practices and testing theories, in particular item response theory, and their ability to link literacy practices and calculable psychological constructs’ Cormac O’Keefe suggests that large-scale digital assessments such as piaac do not merely produce data about ability – more importantly, they ‘perform the concept of ability into being’ (O’Keefe 2017:133). Ben Williamson looks into power relationships behind the development of educational data science. He asks an important question, ‘who owns big data?’ and shows its direct links to the broader question ‘who owns educational research?’ (Williamson 2016). In his analysis of ‘the Lytics Lab, Stanford University’s laboratory for research and development in learning analytics, and the Center for Digital Data, Analytics and Adaptive Learning, a big data research centre of the commercial education company Pearson’, he makes ‘an important central argument that educational data science has moved from non-profit academic laboratories to for-profit companies’ (Williamson 2017: 105).
In our recent book, Michael Peters and I argue that big data is crucial for enabling the digital logic that drives the single technical system of ‘algorithmic capitalism’ (Peters and Jandrić 2018: 32; see also Peters 2012; Peters and Bulut 2011; Braidotti 2015). Christian Fuchs describes a similar notion of ‘transnational informational capitalism’ (Fuchs 2011), and Jodi Dean speaks of ‘communicative capitalism’ in which ‘big data is interesting because of the way in which every kind of communicative interaction generates metadata consisting of location data, different layers of contacts and networks, and links between them. Now all our social substance is available to be enclosed, analyzed, and sold off’ (Dean, Medak, and Jandrić 2018). Astrid Mager examines ‘algorithmic ideology’, which is inscribed in data produced and used by computer code and broader computational logics, and concludes that ‘a fundamental debate about where to draw boundaries between the state and the market, how to set limits for corporate players, and how to sustain social justice is needed’ (Mager 2014: 37). Today’s (big) data is far from neutral, as ‘the complex systems of data production and representation co-constitute the very systems they purport to describe, and in this process, they often embed, replicate or reinforce pre-existing attitudes and prejudices’ (Jones 2018: 49). In the postdigital age, data actively co-constitutes our reality.
The approach from web science is to understand that the Web ecosystem is a composite open and dynamic system of humans and machines – referred to by Tim Berners-Lee as ‘social machines’ – a term that signals collective intelligence and motivates web users to collaboratively use and develop collective resources (Hendler & Berners-Lee, 2010). Education web science needs to examine, analyze, utilize and experiment with Internet-based forms of collective intelligence – a long-term development that runs counter to ideologies of individualism in educational policy, testing and assessment.
(Peters and Jandrić 2018: 62, italics form the original)
Emerging attempts in this area, such as digital humanities, have already shown some good results (see Jandrić 2017: 131). However, understanding of the Web as a dynamic system of humans and machines has also brought about many questions about the intersections between the material and the social worlds ‘where the human teacher’s agency comes up against the workings of data to conduct another, and different, kind of teaching which is neither human not machinic but some kind of gathering of the two’ (Jandrić 2017: 206). This gives rise to various sociomaterialist approaches which ‘conceptualise knowledge and capacities as being emergent from the webs of interconnections between heterogeneous entities, both human and nonhuman’ and ‘offer the prospect of being able to integrate the material technologies and media found in networked learning into a framework that encompasses people and machines in a symmetrical way’ (Jones 2018: 47). However, not everyone agrees with attempts at creating such frameworks. In June 2002 Steve Fuller and Bruno Latour staged a popular public debate with the following motion: ‘A strong distinction between humans and non-humans is no longer required for research purposes’ (Barron 2003: 78). By this day, the debate has not arrived even close to resolution (Fuller and Jandrić 2018). However, it clearly shows that the question of Big Data reaches all the way to fundamental questions about the changing relationships between humans and machines while they co-create the world as we know it.
Big Data is educators’ friend, because it creates the new window of opportunity including but not limited to educational Web science. Big Data is also educators’ foe, because of its in-built problems of representation including but not limited to the tendency to replicate and reinforce ideologies while presenting itself as fair (as in the case of standardized tests) and ideologically neutral (as in the case of educational research). Big Data is the current research frontier in areas from assessment theory to epistemology and ontology, and Big Data literacy – through its direct impact to the ways we produce, collect, and structure data – is an inherent part of this frontier.
3 AI Literacy
These days, AI is often called bombastic names such as ‘the next digital frontier’ (Bughin et al., 2017). However, computers have been around for a while, so we first need to establish what makes AI so special in relation to older digital technologies. According to Liza Daly, ‘artificial intelligence is the umbrella term for the entire field of programming computers to solve problems. I would distinguish this from software engineering, where we program computers to perform tasks.’ (Daly 2017) This simple definition describes an important paradigm change in inner workings of the computer. Traditional computers, including the most sophisticated expert systems of yesterday, consisted of long lines of code which determined their behaviour: for every input, such systems would do predetermined calculations and provide an output. In contrast, AI systems are provided with some initial rules of behaviour, and then they are ‘taught’ by large datasets. Then, computer independently establishes various connections between input data and produces ‘intelligent’ solutions to new problems in non-predetermined ways. This is the essence of machine learning, which is broadly defined as ‘the science of getting computers to act without being explicitly programmed’ (Ng 2018). Recent developments in machine learning are predominantly in areas of neural networks and deep learning. Neural networks are specific organisational model for machine learning, loosely modelled on neurons in human brains, where different parts of the network specializes for different tasks; deep learning is a class of neural networks with layered networked architecture.
The important thing with deep learning and machine learning in general is it needs a lot of data to train on, so a computer learns to do a task like recognizing an object in an image or identifying that there is a cancer cell or recognizing which word you’re saying when you’re speaking by looking at millions of examples, and one reason why neural nets didn’t catch on earlier is that we didn’t have that much data in the 90s.
(in Faggella 2017)
One important quality issue is data bias, which appears in different forms. These biases affect the (machine learning) algorithms that we design to improve the user experience. This problem is further exacerbated by biases that are added by these algorithms, especially in the context of recommendation and personalization systems.
(Baeza-Yates 2016)
AI systems do not only embed, replicate or reinforce attitudes or prejudices found in data – more importantly, they also recombine them and produce new biases. Creators and researchers of AI cannot directly predict or interfere with these processes; they can only change input variables such as architecture of neural network or input dataset and hope that their results will improve. However, this is easier said than done, and non-predictability remains an important in-built feature of AI.
As I write these words, AI’s non-predictability arriving from combinations of data bias and system architecture is a hot research topic. ibm research claims that ‘bias in AI system mainly occurs in the data or in the algorithmic model’; therefore, they claim, ‘it’s critical to develop and train these systems with data that is unbiased and to develop algorithms that can be easily explained’ ( ibm Research 2018). However, using unbiased data is often not enough for developing unbiased AIs, and companies are trying out other solutions. In September 2018 Wired’s Jessi Hempel has written about a new trend: ‘an auditing process that asks companies to open up their technology for evaluation’ (Hempel, 2018). Algorithm auditing does not follow any standard procedures, and ‘an audit doesn’t prove that a company has avoided all the unintended pitfalls of an algorithm. The auditor might not look at the right set of stakeholders, or pose the right set of questions.’ (Hempel, 2018) Nevertheless, concludes Hempel, ‘it’s a baby step toward a more transparent data future: If we cannot strip algorithms of all their bias, at least we should rid them of the bias we can identify.’ (Hempel, 2018) At the moment, AI literacy cannot be thought of without (big) data literacy and information literacy in more general. However, AI brings about a unique challenge of constantly changing and never completely disappearing biases arriving from two dialectically intertwined sources: system architecture and input data. For the time being these biases cannot be predicted or supressed analytically, and solutions such as auditing software are becoming increasingly similar to educational assessment of human beings.
4 The Postdigital Challenge of Critical Media Literacy
Critical media literacy is an educational response that expands the notion of literacy to include different forms of mass communication, popular culture, and new technologies. It deepens the potential of literacy education to critically analyze relationships between media and audiences, information, and power. Along with this mainstream analysis, alternative media production empowers students to create their own messages that can challenge media texts and narratives.
(Kellner and Share, 2007)
In place of conclusion, I will now build upon this definition using insights into data and AI literacy. In the age of traditional mass media such as radio and television, technological affordances could be taken for granted in our studies of relationships between information, popular culture, ideology, and power. However, we are now entering a new stage of technological development, where human attitudes, ideologies, and power relationships are not only expressed through technology, or even built into technology – these days, they also inform and direct technology’s autonomous behaviour. AI systems are obviously technological / digital, but their functioning is dialectically intertwined with the non-technological / human. In result, our postdigital reality ‘is hard to define; messy; unpredictable; digital and analog; technological and non-technological; biological and informational. The postdigital is both a rupture in our existing theories and their continuation.’ (Jandrić et al. 2018: 895) The postdigital challenge does not make earlier forms of critical media literacy irrelevant; instead, it updates them for the digitally saturated world. In words of Peters and Besley: ‘The postdigital does not describe a situation, condition or event after the digital. It is not a chronological term but rather a critical attitude (or philosophy) that inquires into the digital world, examining and critiquing its constitution, its theoretical orientation and its consequences.’ (Peters and Besley 2018) The crucial contemporary challenge of critical media literacy, therefore, is to accept the arrival of postdigital reality and reinvent existing theories and practices.
Looking more practically, it is a no-brainer that critical media literacy needs to involve deep understanding of issues pertaining to data and AI. Referring again to Peters and Besley, ‘the critique of digital reason has two elements: first, the mathematico-technical control systems that are part of the emerging global digital infrastructure within which we now exist, and second, the political economy of these systems – their ownership, acquisition, structure and ownership.’ (Peters and Besley 2018) These elements are dialectically intertwined: Big Data carries inevitable political, economic, and ideological baggage in the form of in-built biases, and AI algorithms exacerbate existing data biases and produce new ones. Postdigital biases are not only carried by technology, or disseminated through technology, or produced by people using technology. They are built into technology, and developed through technology, in new and fundamentally non-predictable ways. This brings about the second postdigital challenge of critical media literacy: finding a new balance between the technological aspects of data and AI literacy with the political aspects of data and AI literacy, and learning how to deal with non-predictability.
Whether we philosophically subscribe to one or another type of sociomaterialist symmetry between humans and non-humans, or we decide to follow more traditional understandings of humans as masters of technology, there is no doubt that AIs have fundamentally altered the nature of human control over technologies. Traditional practices such as software testing are now being replaced by far less analytic practices such as AI auditing – in place of mathematico-technical processes where humans control software, we now have techno-biological processes where humans teach software and then assess its independent behaviour. Thus we arrive to ‘the convergence and marriage of the two dominant world historical forces of digital and biological systems and the ways in which together they constitute the unsurpassable horizon for existence and becoming – the species evolution of homo sapiens and life in general.’ (Peters and Besley 2018) This brings into the fore various posthumanist accounts of human and social development, which translate into some deep philosophical questions. ‘While we speculate what kind of future world we will inhabit in coexistence with new forms of intelligent life, we should firmly focus on the questions what forms of intelligent life should be included in our collective decisions about the future and how we might raise them’ (Peters and Jandrić forthcoming 2019). The third postdigital challenge of critical media literacy is to find ways of active engagement with these questions. In the postdigital context, the question of literacy does not relate only to people; these days, we also need to start thinking what makes AIs literate and develop ways of raising literate thinking machines.
The postdigital challenge is a rupture and continuation in our existing approaches to critical media literacy. It is about the practical challenge of producing (more) balanced datasets, about understanding of inner workings of AIs, about grand philosophical questions such as equality and / or symmetry between human and non-human actors, and about conceptualisation, development, and understanding of new forms of intelligence we would like to live with in the future. We are now at the very beginning of the postdigital epoch, at its very infancy, and critical media literacy has an urgent duty to take an active role in all aspects of its development.
References
Baeza-Yates R. (2016). Data and algorithmic bias in the web. Proceedings of the 8th acm Conference on Web Science – WebSci ’16. Doi:10.1145/2908131.2908135.
Barron C. (2003). A strong distinction between humans and non-humans is no longer required for research purposes: A debate between Bruno Latour and Steve fuller. History of the Human Sciences, 16(2), 77–99.
Braidotti R . (2015). Introduction to a lecture by Luciana Parisi, automated cognition, algorithmic capitalism and the incomputable. http://cfh-lectures.hum.uu.nl/automated-cognition-algorithmic-capitalism-and-the-incomputable/. Accessed 15 October 2018.
Bughin J. ; Hazan E. ; Ramaswamy S. ; Chui M. ; Allas T. ; Dahlström P. ; Henke N. ; & Trench M. (2017). Artificial Intelligence the Next Digital Frontier?. Discussion Paper. McKinsey Global Institute. https://www.mckinsey.com/~/media/McKinsey/Industries/Advanced%20Electronics/Our%20Insights/How%20artificial%20intelligence%20can%20deliver%20real%20value%20to%20companies/MGI-Artificial-Intelligence-Discussion-paper.ashx. Accessed 15 October 2018.
D’Ignazio C. & Bhargav R. (2015). Approaches to Building Big Data Literacy. Bloomberg Data for Good Exchange Conference. 28-Sep-2015, New York City, NY, usa. https://dam-prod.media.mit.edu/x/2016/10/20/Edu_D’Ignazio_52.pdf. Accessed 11 October 2018.
Daly L . (2017). AI Literacy: The basics of machine learning. https://worldwritable.com/ai-literacy-the-basics-of-machine-learning-2e20f93e34b4. Accessed 15 October 2018.
Dastin J. (2018). Amazon scraps secret AI recruiting tool that showed bias against women. Reuters Business News, 10 October. https://www.reuters.com/article/us-amazon-com-jobs-automation-insight/amazon-scraps-secret-ai-recruiting-tool-that-showed-bias-against-women-idUSKCN1MK08G. Accessed 11 October 2018.
Dean J. ; Medak T. & Jandrić P. (2018). Embrace the antagonism, build the Party! The new communist horizon in and against communicative capitalism. Postdigital Science and Education, 1(1). Doi: https://doi.org/10.1007/s42438-018-0006-7.
Faggella D . (2017). The Rise of Neural Networks and Deep Learning in Our Everyday Lives – A Conversation with Yoshua Bengio. https://www.techemergence.com/the-rise-of-neural-networks-and-deep-learning-in-our-everyday-lives-a-conversation-with-yoshua-bengio/. Accessed 15 October 2018.
Fuchs C. (2011). Foundations of Critical Media and Information Studies. London and New York: Routledge.
Fuller S. & Jandrić P. (2018). The Postdigital Human: Making the history of the future. Postdigital Science and Education, 1(1). Doi: https://doi.org/10.1007/s42438-018-0003-x.
Hempel J. (2018). Want to prove your business is fair? Audit your algorithm. Wired, 5 September. https://www.wired.com/story/want-to-prove-your-business-is-fair-audit-your-algorithm/.
ibm Research (2018). AI and bias. https://www.research.ibm.com/5-in-5/ai-and-bias/. Accessed 15 October 2018.
Jandrić P. (2017). Learning in the Age of Digital Reason . Rotterdam: Sense.
Jandrić P. ; Knox J. ; Besley T. ; Ryberg T. ; Suoranta J. & Hayes S. (2018). Postdigital Science and Education. Educational Philosophy and Theory . Doi: https://doi.org/10.1080/00131857.2018.1454000.
Jones C. (2018). Experience and Networked Learning. In Bonderup Dohn N. ; Cranmer S. ; Sime J.A. ; M. de Laat & Ryberg T. (Eds.), Networked Learning: Reflections and Challenges, pp. 39–56. Springer International.
Kellner D. & Share J. (2007). Critical media literacy is not an option. Learning Inquiry, 1(1), 59–69. Doi: https://doi.org/10.1007/s11519-007-0004-2.
Knox J. (2015). Critical education and digital cultures. In Peters M. (Ed.), Encyclopedia of educational philosophy and theory. Singapore: Springer, 1–6. Doi: https://doi.org/10.1007/978-981-287-532-7_124-1.
Koltay T. (2015). Data literacy: in search of a name and identity. Journal of Documentation, 71(2), 401–415. Doi: https://doi.org/10.1108/JD-02-2014-0026.
Mager A. (2014). Defining algorithmic ideology: Using ideology critique to scrutinize corporate search engines. Triple C: Communication, Capitalism and Critique, 12, 28–39.
Ng A . (2018). Machine Learning. https://www.coursera.org/learn/machine-learning. Accessed 15 October 2018.
O’Keeffe C. (2017). Economizing education: Assessment algorithms and calculative agencies. E-Learning and Digital Media, 14(3), 123–137. Doi: https://doi.org/10.1177/2042753017732503.
Peters M. (2017). Deep learning, education and the final stage of automation, Educational Philosophy and Theory, https://www.tandfonline.com/doi/abs/10.1080/00131857.2017.1348928.
Peters M.A. (2012). Algorithmic capitalism and educational futures: Informationalism and the googlization of knowledge. TruthOut. http://truth-out.org/news/item/8887-algorithmic-capitalism-and-educational-futures-informational ism-and-the-googlization-of-knowledge. Accessed 15 October 2018.
Peters M.A. & Besley T. (2018). Critical Philosophy of the Postdigital. Postdigital Science and Education. Doi: https://doi.org/10.1007/s42438-018-0004-9.
Peters M.A. , & Bulut E. (Eds). (2011). Cognitive capitalism, education and digital labor . New York, NY: Peter Lang.
Peters M.A. & Jandrić P. (2018). The Digital University: A Dialogue and Manifesto . New York: Peter Lang.
Peters M.A. & Jandrić, P. (forthcoming, 2019). AI, Human Evolution, and the Speed of Learning. In Knox J. ; Wang Y. ; & Gallagher M. (Eds.), Artificial Intelligence and Inclusive Education: speculative futures and emerging practices. Springer Nature.
Shadbolt N. , Hall W. , Hendler J.A. , & Dutton W.H. (2013). Web science: A new frontier. Philosophical Transactions of the Royal Society A , 371: 20120512.
Williamson B . (2016). Who owns educational theory? Pearson, big data and the ‘theory gap.’Retrieved 7 July 2017 from https://codeactsineducation.wordpress.com/2016/01/19/who-ownseducational-theory/. Accessed 15 October 2018.
Williamson B. (2017). Who owns educational theory? Big data, algorithms and the expert power of education data science. E-Learning and Digital Media, 14(3), 105–122. Doi: https://doi.org/10.1177%2F2042753017731238.