Abstract
Algorithms are playing an ever-increasing role in the way that we understand and interact with our world. This paper provides an overview of Artificial Intelligence technologies, explores the promise of algorithms as well as the issues and considers the implications of algorithms on education and literacy practices.
Keywords
Algorithms, Literacy, AI, Artificial Intelligence, Education, Deep Learning, Bias, Algorithmic Discrimination, Media, Social Media, Data Literacy
Algorithms have an ever-increasing impact on our daily lives on both a conscious and a subconscious level. Have you stopped to consider the path that you took when you drove to work? Perhaps you recall a time that you communicated with a friend about a very specific topic and shortly afterwards you found yourself inundated with related ads on Facebook, Gmail and YouTube. Do you ever wonder why Netflix or YouTube make the recommendations that they do? Algorithms are everywhere. You use an algorithm when finding the fastest path to work. Smart speakers, email accounts and chat programs are listening in on your conversations so they can advertise to you. Netflix and YouTube aim to maximize your viewing time and have little concern about whether or not you like the content.
This article will examine the competing definitions of algorithms, the stakeholders who are most affected by algorithms, inequities caused by the power structures behind algorithms and what happens when algorithms go wrong. Finally, the educational aspects of algorithms will be explored. Algorithms have a significant impact on the way that research is conducted by students. Algorithms are also fundamentally changing how we consume media and directly affecting the way that we write. In order to adapt to the increasing role of algorithms in education, this article will explore creative content creation as a viable alternative to research and writing assignments that algorithms have turned into a less creative, more routinized process. Finally, this article will discuss why teachers must be aware of the influence of algorithms and work to promote algorithmic literacy among their students.
What are Algorithms?
At the most basic level, an algorithm is a set of rules or instructions to be followed (Tutt, 2017). We make use of algorithmic thinking in our everyday lives. We use an algorithm when we determine the fastest way to drive to work, when we navigate through a cafeteria in the most efficient way possible or when we follow a recipe to bake a cake (Wing, 2006). Algorithms are also a fundamental component of how computers function, providing a desired output based on inputs (Lloyd, 2019). The tasks performed by computer algorithms range from simple to complex. One example of a simple algorithm is a sort function, an algorithm that organizes entities numerically or alphabetically (Vöcking et al., 2010). Manufacturers embed simple algorithms within the electronics of nearly all modern appliances, performing basic functions like cooking rice or moderating the temperature of the fridge (Vöcking et al., 2010). Technologies such as GPS units, that are now commonplace in cars and mobile phones for navigation, use more advanced algorithms. The advanced algorithms in a GPS unit simultaneously consider many variables, such as the speed of the earth’s rotation and precise timings, in order to make informed predictions (Vöcking et al., 2010). Other famous algorithms of this type include the original PageRank algorithm used by Google to create a database of the internet and return relevant web pages. A simplified description of how PageRank functions is that it looks at the links between web pages and assigns scores based on the popularity of the originating website (Tutt, 2017). A page that is being linked to from the New York Times will receive a higher PageRank score than a page that is being linked to from a personal blog (Tutt, 2017). While the simple and advanced algorithms described above vary in complexity, they share several similarities. All the aforementioned algorithms might be called traditional algorithms. Traditional algorithms have been designed entirely by a programmer or group of programmers who have a concrete idea of the function of each line of code, and their output is deterministic, meaning that given the same set of inputs the algorithm will always produce the same result (Tutt, 2017).
The first functional use of an artificial intelligence (AI) algorithm emerged in the 1950s. Taking cues from recent advances in neurobiology, Marvin Minsky created a machine that used a randomly connected collection of electronic synapses to perform calculations (McCorduck & Cfe, 2004; Russell & Norvig, 2002). Many of today’s most complex algorithms rely on artificial intelligence, an umbrella term which encompasses machine learning, deep learning, neural networks and numerous algorithmic implementations such as reinforcement learning and convolutional neural networks (Zubarev, 2018).
AI-based algorithms are used to personalize online content for users (Giansiracusa, 2021). According to Burrell (2016), machine learning is typically used for classification tasks, such as organizing search results and making predictions about credit scores and the likelihood of a loan default and filtering search results. Vasily(2018) provides a simplified definition of machine learning, describing it as predicting results from incoming data with some AI thrown in. Deep reinforcement learning (deep learning) is another type of AI. Deep learning relies on neural networks, a type of algorithm which emulates how neurons in our brain function by reinforcing strong connections and trimming weak connections, and unlike traditional programming they achieve their results through weighting and repetition (Burrell, 2016). Within a given neural network, there are many layers of connections and, importantly, many of those layers are hidden. Because of the opacity that exists within a neural network, even the creator of the algorithm is unlikely to understand exactly how the algorithm reached its decision (Burrell, 2016; O’Neil, 2016). The opacity within AI algorithms can mask biases and make it difficult or impossible to establish accountability (O’Neil, 2016). This creates many moral and ethical ramifications that will be discussed in the following pages.
Natural language processing (NLP) is a subset of machine learning. It uses patterns in text and linguistic conceptualization to generate text on its own by analyzing a small block of text and guessing the next word (Giansiracusa, 2021). Machine learning has enabled the spread of fake news,
“and other types of malicious content. With increased usage of AI and machine learning, these attacks become harder to detect and contain, according to Symantec. In 2018, the company found 592 million unique IP addresses connected to more than 1.5 billion malicious emails. Researchers from Stanford University found more than 1.5 million AI-generated fake accounts on Twitter and generated almost 100,000 tweets that generated at least 1,000 retweets in a week. While there have been some advancements in AI, it’s difficult to measure its true impact on a variety of different industries and businesses. The technology that the market is investing in is advancing so quickly that it is difficult to properly evaluate.”
Many individuals believe that they could never be tricked by AI generated writing. However, the italicized text above was generated through machine learning[1]. The best description of language model outputs was made by Twitter user Julian Togelius (@togelius) who Tweeted “GPT-3 often performs like a clever student who hasn’t done their reading trying to bullshit their way through an exam. Some well-known facts, some half-truths, and some straight lies, strung together in what first looks like a smooth narrative” (Togelius, 2020). As NLP systems have improved, it can become increasingly challenging to notice inconsistencies in generated text and to differentiate the errors made by an AI system with errors made by a human author, although differentiating between a computer and a human becomes easier when texts are longer or specific questions are asked (Tegmark, 2017).
NLP systems have been associated with the generation of “fake news” content. Giansiracusa (2021) explains that advertising revenue is tied to the number of views that a webpage receives, encouraging content creators to create blog posts and news stories that are sensationalist, exaggerated and sometimes untrue to encourage web traffic. The impact of fake news stories can be amplified by increased attention as a result of algorithms that repost content and aggregate news stories (Giansiracusa, 2021).
In 1950, Alan Turing devised “the imitation game” a test which is now known as the “Turing Test”, to distinguish between humans and machines based upon a natural language conversation between a computer participant (A) a human participant (B) with a human observer (C) (Turing, 1950). MIT physicist and machine learning researcher Max Tegmark believes that the most advanced NLP systems are taking steps towards passing this test, although he cautions that the test is “testing human gullibility more than true artificial intelligence” (Tegmark, 2017, p. 119). While computers and their associated algorithms have not reached a state that might be considered “intelligence”, the social power that algorithms possess can influence human agency (Lloyd, 2019) and the effectiveness of algorithms in creating deceptive content such as fake news will require us to re-examine our critical literacy practices.
Interpreting Algorithms
We interpret and interact with algorithms in different ways, based upon our individual awareness of algorithms, our personal technical expertise, whether we have access to the algorithm’s code, and the complexity of an algorithm’s underlying code. Algorithmic culture has become a larger part of our everyday lives and it influences the choices and decisions that we make on an everyday basis (Lloyd, 2019). For those not immersed in the world of algorithms, the results of algorithms often appear to be magic or go unnoticed. This type of user is unlikely to notice subtle changes in Amazon recommendations based on search histories or the influence that past searches have made on their Google search results. They are unaware of the datafication of our lives, a process where our everyday practices are measured and recorded, allowing for data analysis, including tracking and prediction (Mertala, 2020). A private conversation about cars might be overheard by a smart speaker and lead to an increased quantity of automobile advertisements. Lloyd (2019), notes that this group of users knowingly or unknowingly give up authority to algorithms when they perform searches or interact with technology. Assuming that this group of naïve computer users is composed of Baby Boomers and GenXers is a common assumption, non-digital-natives who grew up without exposure to advanced technologies and the powerful influence of algorithms should be less likely to possess high levels of digital literacy (Henderson et al., 2020). However, Henderson at al. (2020) has found that age is a worse predictor of digital literacy than gender, ability, and socioeconomic status. If simple exposure to technology and algorithms does not increase literacy, what can be done to educate today’s generation of digital natives?
For those with a deeper understanding of algorithmic interactions, two definitions emerge. The first definition is given by the individuals who create algorithms. For mathematicians, programmers, engineers and alike, algorithms are viewed as a computer function that provides a desired output based on a series of inputs (Lloyd, 2019). Algorithms are viewed less positively by media theorists, sociologists and others who study algorithms from an outsider’s perspective who believe that algorithms contain inherent cultural biases and that they lack accountability and transparency for the decisions they make whose impact can range from mundane to life-changing (Lloyd, 2019; O’Neil, 2016).
Algorithms from an Insider’s perspective
For computer scientists, mathematicians, and engineers; algorithms are often described as a component of computational thinking. Computational thinking focuses on the mathematical process of abstraction and includes algorithms, data structures, state machines, programming languages, logistics and semantics, heuristics, control structures, communication and architectures (Wing, 2006) Computer scientists, engineers and other creators of algorithms interact with algorithms from the inside. For the creators of algorithms, algorithms are often viewed with extreme positivity. Vöcking et al. describe algorithms as “beautiful ideas for tackling or solving computational problems more efficiently” (2011, p. vi). Computer scientists and engineers rightfully praise algorithms for their ability to solve complex problems and the many scientific achievements that would not be possible without AI and advanced algorithms, including the mapping of the human genome, prediction of weather events and the development of medicines and vaccines (Vöcking et al., 2010; Wing, 2006). Algorithms and data collection are advertised as a way to improve the education sector by Holmes et al. (2016) who argue that Educational Data Mining (EDM) can be used to track students’ behaviours and identify students that are at risk of dropping out Additionally, AI can used to provide personalized tutoring to students and facilitate group work by identifying students that would work well as a team based on cognitive abilities and interests (Holmes et al., 2016).
While tech industry insiders are optimistic about the power of algorithms, they are also acutely aware of the potential negative impacts of technology. Steve Jobs famously restricted his children from using iPads (Bilton, 2014) and an increasing number of Silicon Valley tech employees are banning children’s use of technology at home and sending their children in tech-free schools (Weller, 2018). The promotion of technology by powerful tech figures while simultaneously limiting their families exposure to technology, indicates that technology use might also have negative impacts.
Algorithms from the outside
In the movie Jurassic Park, Jeff Goldblum’s character states “your scientists were so preoccupied with whether or not they could, they didn’t stop to think if they should.” (Spielberg, 1993) Media theorists, sociologists and proponents of social justice have an opposing view of what data is and they see a darker side to algorithms. Many in this group view our data as an extension of ourselves. Cheney-Lippold (2017) explains that humans exist within the digital realm as digital entities made up of bits and our digital entities are categorized by algorithms. Tegmark (2017) extends the view that humans are data even further, envisioning a future in which data would enable humans to exist after death. The framing of humans as data entails that data needs to be handled responsibly in the same way that people are; however, this view is incompatible with large tech companies who view data as a commodity (Raffaghelli, 2020). Categorizations made by Google are often wrong, with Cheney-Lippold (2017) providing the example that women who use Google to conduct research on topics within male-dominated fields are more likely to be gendered as male by Google’s algorithm.
Cathy O’Neil coined the term “weapons of math destruction” to refer to a group of algorithms that are opaque, large scale and damaging (2016). O’Neil’s criticisms of algorithms are shared by other theorists who are concerned with the reach of algorithms, a lack of accountability, and the destruction that algorithms have caused and might cause in the future; topics that will be explored within the following sections.
In addition to O’Neil’s specific criticisms, there is also a general concern about the growth in usage of AI-powered bots to spread election propaganda and other types of fake news as well as the possibility that large companies who are in possession of large amounts of human data could make use of their large data sets and AI to manipulate behaviour.(2020)
The use of AI within education is a contentious issue. Mertala (2020) is concerned with the datafication of education proposed by Holmes, arguing that there are negative social consequences to the collection of attendance and test scores, which are often used as a way to compare schools and measure student and teacher performance. Holmes et al.’s (2016) support of EDM to identify at-risk students is also problematic for Raffaghelli et al. (2020) who are critical of the machine learning systems used to identify vulnerable students, noting that the labelling stays with the student throughout their studies and might cause a student to become more vulnerable. Raffaghelli et al. (2020) go on to suggest that any analysis of vulnerability should be considered within the context of student agency and that data practices need to be transparent and developed with the participation of staff and students.
Algorithms are opaque
AI algorithms are extraordinarily difficult for external researchers to understand and analyze for several reasons. Algorithms falling under the umbrella of AI have been described as a “black box” because, in most cases, the source code for the algorithm is proprietary and not available to the public (Burrell, 2016; Lloyd, 2019). Burrell (2016) breaks down the opaque nature of algorithms into three distinct categories: intentional opacity, opacity due to technical illiteracy, and opacity due to scale. Intentional opacity often occurs because companies view their algorithms as intellectual property and closely guard their algorithms from any type of external inspection (O’Neil, 2016). The second category of algorithmic opacity is technical illiteracy, with Burrel (2016) noting that reading and writing code and understanding the design of algorithms is a specialized skill that the majority of the population does not possess. However, even if the design of an algorithm is understood it is unlikely that a human programmer could determine how a specific result was achieved as “both the inputs (data about humans) and outputs (classifications) can be unknown and unknowable” (Mittelstadt et al., 2016, p. 6). Examples of opacity due to scale include the algorithms that underlie Google’s search functionality which are so large and complex that teams of engineers each have an understanding of a small component of the search function and no single person understands the algorithm in its entirety Because algorithms are opaque, there is no mechanism to audit an algorithm’s findings, if the results generated by the algorithm are unfair or discriminatory (Mittelstadt et al., 2016; O’Neil, 2016).The opacity of algorithms makes the results of algorithms difficult to challenge and reduces the agency of individuals who the algorithm is affecting (Mittelstadt et al., 2016)
Discriminatory Algorithms and the Power Structures that Drive them
The incongruent views of algorithms between scientists and theorists could be related to the power differential that exists between the owners of algorithms and the users of algorithms.
According to Lloyd (2019), power structures have an influence on the creation of algorithms themselves. Many of the most advanced and influential algorithms in use today are developed by a group of companies coined “FAANG” composed of Facebook (now Meta), Amazon, Apple , Netflix, and Google (Alphabet) (Hobbs, 2020). All the companies within FAANG are based in The United States of America but have a global influence. (Lloyd, 2019)The programmers, engineers and computer scientists that work for the FAANG group of companies are predominantly high socio-economic status males and when computer code is generated, it is influenced by the social power and positionality of the coder (Lloyd, 2019). In addition to the biases held by the programmers themselves, AI-based algorithms generally require training data, consisting of inputs that can be used to teach the algorithm the compilation of datasets used to train AI algorithms, and these datasets are not neutral (Leander & Burriss, 2020). When facial detection systems were first created, computer scientists used their own faces to train the algorithms and, as a result, facial detection systems are most effective at identifying white male faces (O’Neil, 2016). Modern implementations of facial detection algorithms have used datasets such as the faces of US military employees, imparting a gender bias into the functioning of AI systems (Leander & Burriss, 2020). While algorithms have been shown to give preferential treatment to individuals that resemble their creators, they have also been found to discriminate against outsiders. A notable and particularly egregious example is Level of Service Inventory–Revised (LSI-R), a statistics-based sentencing model that relies on a questionnaire and recidivism data to help determine sentencing in at least twenty-four US States (O’Neil, 2016). According to O’Neil (2016) the questionnaire contains questions that are inadmissible in court and the LSI-R model consistently recommends harsher sentencing for African Americans even when the crimes are similar.
The “Whoops” Factor – Algorithms that have Unintended Influence/Impact
As algorithms become more and more complex, they do not always function in ways their creators intended. Microsoft’s “TAY” an acronym for “thinking about you” was an online chatbot released by Microsoft in 2016 (Lee, 2016). According to Microsoft (2016), TAY was created by “mining relevant public data and by using AI and editorial developed by a staff including improvisational comedians” and was designed to interact with 18 – 24-year-olds in the US. In less than a day, TAY transformed from a friendly chat bot with a teenage persona to vehement Trump supporter that generated racist, pro-nazi and anti-feminist tweets (Hunt, 2016). The TAY experiment lasted for less than 24 hours before the bot was taken down and Microsoft provided an official apology (Microsoft, 2016). Malfunctioning algorithms can also lead to disastrous effects, including the loss of life. An example of algorithms leading to loss of life is illustrated by the recent Boeing 737 max crashes. The crashes were caused by several factors, first a malfunctioning angle of attack indicator (AOA) sent false data to the airplane’s flight computer. The incorrect AOA data led algorithms within the flight computer’s Maneuvering Characteristics Augmentation System (MCAS) to make adjustments to the flight control surfaces, causing the plane’s nose to pitch down, leading to the Lion Air Flight 610 crash on October 29, 2018, and Ethiopian Airlines Flight 302 crash on March 10, 2019, resulting in the deaths of 346 people (Gelles, 2019; Perell, 2020). Computer algorithms may have caused Wall Street’s trillion-dollar “Flash Crash” of May 6, 2010 (Tegmark, 2017). The Flash crash of May 6, 2010 caused the US stock indices to fall by nearly 9% within 15 minutes leading to the loss of over a trillion dollars of market value (Lin, 2013). High frequency trading (HFT) which uses algorithms to make stock trades more quickly than any human broker are a primary cause of the crash (Lauricela et al., 2010). Also implicated, was a single, home-based, trader who manipulated the market by using a computer program to create fake sell orders. (Lynch & Meidema, 2015). Tegmark (2017) believes that one of the root causes of accidents is that algorithms do not understand the data that they are interacting causing them to make predictions that a human would immediately recognize as invalid and he argues that all algorithms require both validation and verification.
Another unintended consequence of algorithms is that algorithms exacerbate the disadvantages faced by marginalized populations who may be denied access to loan applications, college admissions and employment based on their zip code, which is often a proxy for race and socioeconomic status (O’Neil, 2016). Mittelstadt et al. (2016) define this type of algorithmic discrimination as “unfair outcomes”, which occur when algorithms make data-driven conclusions that are discriminatory. As discussed within the “Algorithms are Opaque” section, it can be difficult to prove that algorithms are acting in a discriminatory manner. To reduce the harm caused by discriminatory algorithms, O’Neil (2016) recommends that algorithms be tested with individuals from diverse backgrounds to determine if the algorithms are biased and believes that government regulation is also required.
Algorithms and Education
How Technology and Algorithms are Transforming Education
For as long as computers have existed, education technologists have made claims that computers, the internet and student access to new technologies will transform education and create equitable educational outcomes (Reich, 2020). However, increased access to technology, such as computers, has not been found to improve educational outcomes (Waluyo, 2019; Starkey & Zhong, 2019). Battle (1999) finds that students from higher socioeconomic backgrounds are more likely to use computers for research purposes and that they receive a greater educational benefit from computer use. There is a real threat that the increased use of technology will increase the digital divide which the OECD (2001) describes as “the gap between individuals, households, businesses, and geographic areas at different socioeconomic levels concerning both their opportunities to access information and communication technologies and their use of the Internet for a wide variety of activities” (p.5). Better teacher training has been found to improve the benefits that students receive from the use of technology reducing the digital divide (Starkley et al., 2016; Warcheauer et al., 2016). Reich (2020) recommends that educators immerse themselves in new technologies and teach students how to use new technologies, rather than continuing to teach in old ways and using technology as a tool. A similar strategy is recommended for increasing artificial intelligence literacy among students. To increase artificial intelligence literacy, Reich (2020) advocates that students learn about how AI functions and ethical considerations when working with AI rather than just learning how to use AI-based programs (Ng et al., 2021). In order to prepare students for a world filled with algorithms and AI educators will first need to understand these technologies so that they can provide effective artificial intelligence literacy instruction.
How Algorithms shape the media that we consume
As storage and processing power have become affordable, there is no limit to the amount of information that can be stored and analyzed (Cheney-Lippold, 2017). Algorithms make use of this large body of information to assign age, gender and race to users based on their search history and identified interests (Cheney-Lippold, 2017). Our internet histories can have a direct influence on what we read. When we perform a google search to find more information about a topic or for suggestions for the next book to read, Google considers our stored personal data and search keywords before returning its results (Kraus-Friedberg, 2021). The results that we see are also reduced by our quest for efficient retrieval of information. Bhatt and MacKrenzie (2019) found that higher education students in the fields of STEM, the humanities and business all used ritualized behaviours in their searches related to course assignments. A broad analysis of computer users found that during a Google search, somewhere between 30% and 40% of all users click on the first result, and two-thirds of users click on one of the first three results (Bailyn, 2022; Fay, 2022). Google’s search algorithm has made it easy to find information. Unfortunately, the same filtering that makes Google searches so efficient also works to filter out alternate perspectives that might be less popular or that run contrary to the profile that Google has established for us based on our previous search history. While the specific details of the PageRank algorithm is a closely guarded secret, Google has released the details of how its YouTube Recommendation system works. On YouTube, the effects of prior searches are especially apparent because Google’s recommendation algorithm has been finely tuned to maximize the amount of time that users spend consuming video content (Giansiracusa, 2021). Giansiracusa’s recommendation algorithm findings are mirrored in Google’s own research presentation. At the time of Google’s 2016 conference presentation, the algorithm for video recommendations was described as “generally a simple function of expected watch time per impression.” that “is constantly being tuned based on live A/B testing results” (Covington et al., 2016, p. 5) Figure 3 of Google’s presentation Deep Neural Networks for YouTube Recommendations. Proceedings of the 10th ACM Conference on Recommender Systems, 191–198 indicates that a user’s viewing history has an impact on the type of videos that are recommended and the ranking of the recommended videos. To counter the bias that exists within internet research, Bhatt and MacKenzie (2019) believe that students need to learn about algorithms to better understand the way that their online experiences are shaped and to make students aware that any internet research contains bias.
How Algorithms are Transforming Writing
Algorithms have transformed societal writing patterns. The widespread use of mobile phones has replaced simple writing tasks, such as to-do lists, grocery lists and directions. While early mobile phones required users to provide input through typing, improvements in the machine learning algorithms that power voice to text translation have increased the likelihood that information is dictated to our devices rather than typed. According to a Stanford study, speech recognition using Baidu’s Deep Speech 2 deep learning-based algorithm was three times faster than typing on an iPhone and resulted in fewer errors (Ruan et al., 2017). According to the Ruan et al. (2017) study, speech recognition could transcribe text at a rate of 100 words per minute, a rate that is significantly faster than students in the study could handwrite or type on a computer keyboard (Horne et al., 2011).
Technologies designed to improve a user’s writing, such as Grammarly’ and ProWritingAid’, have also made improvements in their ability to find and correct errors in writing. An action research study from Indonesia found that the use of Grammarly improved essay writing, increasing the number of students meeting the success criteria from 62% to 80% with key improvements in the categories of grammar and diction (Karyuatry, 2018). It is conceivable that programs designed to correct spelling and grammar errors could become so proficient that they would find and correct virtually every error within a given text. When programs reach this critical threshold, the impact could be as great as the impact of the calculator on mathematics. Although spelling and grammar knowledge would still be important, assessment of written works would have a greater focus on creativity and originality and a reduced focus on spelling, punctuation and syntax.
Tools such as Grammarly and ProWritingAid correct what you type, the newest generation of tools do the typing for you. Based on natural language processing and deep learning AI models, autocomplete technologies provide word suggestions as you type (Tegmark, 2017). One example of an autocomplete technology is Google’s Smart Compose, introduced to Gmail around 2009 as a way to make typing emails faster, it has since been added to other Google applications including the Google Docs application used in many schools (Google, 2022). According to information presented by Google at the 2019 Knowledge, Discovery and Datamining Conference, “the core of Smart Compose is a large-scale neural language model” (Chen et al., 2019, p. 2287). Smart Compose makes use of large language models that have been trained using machine learning, by looking at millions of sentences that have been written by others, the algorithm predicts the words that are most likely to come next in your sentence and offers them as suggestions that can be automatically added to your writing by pressing the tab key (Chen et al., 2019). Technologies such as Smart Compose are problematic for educators because the suggestions provided reflect the most common words written by adults. When students accept the recommendations made by Smart Compose it may change the flow and style of a student’s writing. More significantly, text completion algorithms have the potential to impart an unintended bias in writing. Google (2022) acknowledges the possibility of unintended bias by posting the following disclaimer on its smart compose web page: As language understanding models use billions of common phrases and sentences to automatically learn about the world, they can also reflect human cognitive biases. Being aware of this is a good start, and the conversation around how to handle it is ongoing. Google is committed to making products that work well for everyone and are actively researching unintended bias and mitigation strategies (2022).
In the coming years, educators will need to cope with writing that has been generated entirely by algorithms. In 2005, Massachusetts Institute of Technology (MIT) students created a rudimentary program called “SCIgen” that generated nonsensical research papers. The students submitted the fake papers and were invited to speak at the World Multiconference on Systemics, Cybernetics and Informatics (WMSCI) to present their entirely generated paper “Rooter: A Methodology for the Typical Unification of Access Points and Redundancy,” (Conner-Simons, 2015). Springer Publishing, one of the largest academic publishing houses, announced that they had removed over 120 papers generated by SCIgen, and had created a free tool to detect papers created by SCIgen (Conner–Simons, 2015). Modern algorithms relying on advanced language models produce text that is much more convincing than the text generated by SCIgen. Given an input of several words, the natural language processing algorithms (such as the aforementioned GPT-J) are able to generate text of any length at speeds that are significantly faster than any human can generate content. An example recalled by Giansiracusa (2021), was a GPT-3 bot that posed as a human on Reddit’s /r/AskReddit forum and responded to questions in under one minute with six paragraph long responses. As research and the generation of text becomes simplified, I predict that there will be a change in the way that knowledge is produced and presented within education, with research being aided by algorithms and the production of text becoming more efficient. Bhatt and MacKenize (2019) advocate for a shift away from ritualized student research towards the creation of original content. A view is compatible with the prediction that any job that can be performed on a computer will eventually be taken over by computers (Reich, 2020; Tegmark, 2017)
The need for Data Literacy
Henderson et al. (2020) challenged the concept of “digital, or the idea that those who have been brought up during a time of advanced technology are automatically experts in technology use . In their study, young adults (19-29) were found not to demonstrate advanced digital skills and assumptions of technological expertise and digital fluency was damaging to educational development. The rapid growth in the way that AI driven systems have shaped the internet has not been reflected in “current educational curricula, practices and theories” (Leander & Burriss, 2020, p. 1264). An example of this is the concept of a shared text, texts are no longer static, they can be customized to the individual and may change over time we can no longer assume that texts are a shared experience (Leander & Burriss, 2020). According to Henderson et al (2020), engaging students in internet research is transformative and can have a lasting impact on students’ digital lives beyond school. Mertala (2020) has found that the education sector has enacted formal data literacy pedagogies that focus on improving students’ data literacy when conducting research, but that they neglect to teach students about data literacy in their everyday lives. Henderson et al. (2020) advocate the use of a digital literacy toolkit that includes the development of skills in a participatory manner and encourages students to investigate “dataveillance , internet wellbeing, and artificial intelligence/machine learning” as well as consider personal privacy issues. Before such a toolkit could be implemented, frameworks are needed to guide educators in teaching students about AI (Ng et al., 2021).
Traditional approaches to literacy and the analysis of media will need to be revised in response to the growing impact of algorithms. Although media educators have pedagogical tools to critically analyze texts, Leander and Burriss (2020) express concern that new media is often entangled with machines and, for this reason, the methods used to analyze text need to expand. The definition of the hidden curriculum, lessons which are taught and learned in an unconscious manner (Kentli, 2009) needs to be broadened to include data literacy, as the materials used within a classroom, such as computers and software, might also teach students about hidden messages (Mertala, 2020). The definition of literacies also needs to be expanded to include computer coding. According to Vee (2017), coding parallels literacy as it is “a socially situated, symbolic system that enables new kinds of expressions as well as the scaling up of pre-existing forms of communication” (p. 4). Within a new definition of literacy that includes computer coding, critical computer literacy should also be explored. While understanding technology is an excellent first step, users of technology must also recognize the impact of AI and take action towards AI that is unjust (Leander & Burriss, 2020). As teachers we need to promote data literacy among our students. According to Raffaghelli (2020) Strong data literacy skills enables students to express themselves, it enhances diversity, and helps students contribute towards social justice issues.
Algorithms and the Future Direction of Education
Tegmark (2017) predicts that in the future of employment, the skills that are least likely to be replaced by AI include social intelligence, creativity, and working in an unpredictable environment. Reich (2020) calls for educational reforms, noting that any metric that can be easily accessed by a computer is likely related to a task that can be easily performed by a computer and not required within the labour market with examples including bank tellers and airline check-in employees. One potential way for students to express creativity and social intelligence is through content creation through platforms such as TikTok, a video sharing platform that lets users record brief clips that can then be edited to include filters and background music. While TikTok has a reputation among parents of providing little educational value and questionable content (Quirt, 2020) the videos created on the platform show a high level of creativity and resonate with today’s youth in a way that Anderson (2020) describes as “creative chaos” due to the wide variety of content and the uncertainty of what will appear next (p. 5). Using a thematic analysis approach, Li et al. (2021) found that TikTok videos and, in particular, TikTok dances were one of the most effective ways to freely convey public health information related to Covid-19 based upon the number of likes and shares. TikTok can also be used to gauge student sentiment. Using a thematic analysis approach, Literat (2021) studied the content of 1,930 TikTok videos containing the hashtag #onlineschool finding that students attending school virtually had significant concerns about mental health issues and the way that their home life would be perceived by others. In its current state, TikTok is not a viable solution for education, as some videos contain explicit content and there are also concerns about online privacy and the algorithms used by the platform to determine which videos will be displayed next (Anderson, 2020). However, there are also many unexpected positives to TikTok, with Literat (2021) noting that students created videos to help others with math concepts and that the TikTok comments section was used for peer support and “mental health check-ins on videos tagged #ADHD” (p. 7). As educators, we need to be aware of the wide sweeping changes that algorithms have made to the education system and embrace new methods of communication in order to resonate with students.
Conclusion
Algorithms will continue to become an increasingly prevalent part of our daily lives. While some scholars have called for the regulation of algorithms (O’Neil, 2016; Tutt, 2017) the global spread of the internet makes regulation difficult to implement on a global scale. As computers become faster and algorithms become more advanced, I believe that the impact of algorithms and artificial intelligence will continue to grow. Algorithms have the ability to make our lives easier by simplifying communication. Algorithms can increase our media enjoyment by providing us content that is curated to meet our interests. But algorithms have also been shown to reflect the worst human qualities, including callousness when making life impacting decisions, racism, and the pursuit of profit above all else. The impacts of algorithms on education are widespread, with algorithms affecting how we receive, process, and communicate information. Educators will need to adapt to the changing employment climate brought on by algorithms and also teach students the algorithmic literacy skills that are necessary to understand the influence of algorithms.
References
Battle, P. A., Juan. (1999). Home Computers and School Performance. The Information Society, 15(1), 1–10. https://doi.org/10.1080/019722499128628
Bhatt, I., & MacKenzie, A. (2019). Just Google it! Digital literacy and the epistemology of ignorance. Teaching in Higher Education, 24(3), 302–317.
Bilton, N. (2014, September 10). Steve Jobs Was a Low-Tech Parent. The New York Times. https://www.nytimes.com/2014/09/11/fashion/steve-jobs-apple-was-a-low-tech-parent.html
Burrell, J. (2016). How the machine ‘thinks’: Understanding opacity in machine learning algorithms. Big Data & Society, 3(1), 205395171562251. https://doi.org/10.1177/2053951715622512
Cheney-Lippold, J. (2017). We are data: Algorithms and the making of our digital selves. New York University Press.
Conner-Simons, A. (2015). How three MIT students fooled the world of scientific journals. MIT News | Massachusetts Institute of Technology. https://news.mit.edu/2015/how-three-mit-students-fooled-scientific-journals-0414
Covington, P., Adams, J., & Sargin, E. (2016). Deep Neural Networks for YouTube Recommendations. Proceedings of the 10th ACM Conference on Recommender Systems, 191–198. https://doi.org/10.1145/2959100.2959190
Fay, O. (2022, March 31). The Value Of Google Result Positioning. Poll the People. https://pollthepeople.app/the-value-of-google-result-positioning-3/
Foucault, M. (1980). Power/knowledge: Selected interviews and other writings, 1972-1977. Vintage.
Gelles, D. (2019, March 22). Boeing 737 Max: What’s Happened After the 2 Deadly Crashes. The New York Times. https://www.nytimes.com/interactive/2019/business/boeing-737-crashes.html, https://www.nytimes.com/interactive/2019/business/boeing-737-crashes.html
Giansiracusa, N. (2021). How Algorithms Create and Prevent Fake News: Exploring the Impacts of Social Media, Deepfakes, GPT-3, and More. Apress. https://doi.org/10.1007/978-1-4842-7155-1
Henderson, M. J., Shade, L. R., & Mackinnon, K. (2020). EVERY CLICK YOU MAKE: ALGORITHMIC LITERACY AND THE DIGITAL LIVES OF YOUNG ADULTS. AoIR Selected Papers of Internet Research.
Hobbs, B. (2020, November 6). FAANG stands for five very successful tech companies that can move the stock market—Here’s what to know about investing in them. Business Insider. https://www.businessinsider.com/what-is-faang
Holmes, W., Luckin, R., & UCL Knowledge Lab. (2016). Intelligence unleashed: An argument for AI in education. Pearson : UCL Knowledge Lab.
Horne, J., Ferrier, J., Singleton, C., & Read, C. (2011). Computerised assessment of handwriting and typing speed. 28(2), 16.
Hunt, E. (2016, March 24). Tay, Microsoft’s AI chatbot, gets a crash course in racism from Twitter. The Guardian. https://www.theguardian.com/technology/2016/mar/24/tay-microsofts-ai-chatbot-gets-a-crash-course-in-racism-from-twitter
Jeppesen,S., Hoechsmann, M., ulthiin, i., VanDyke, D., McKee, M. (2022). The Capitol Riots: Digital Media, Disinformation, and Democracy Under Attack. Routledge.
Karyuatry, L. (2018). Grammarly as a Tool to Improve Students’ Writing Quality: Free Online-Proofreader across the Boundaries. JSSH (Jurnal Sains Sosial Dan Humaniora), 2(1), 83. https://doi.org/10.30595/jssh.v2i1.2297
Kraus-Friedberg, C. (2021). I Could Just Google It: Algorithms, Information Literacy and Database Instruction. Michigan Academician, 47(3), 74. ABI/INFORM Global.
Leander, K. M., & Burriss, S. K. (2020). Critical literacy for a posthuman world: When people read, and become, with machines. British Journal of Educational Technology, 51(4), 1262–1276.
Lee, P. (2016, March 25). Learning from Tay’s introduction. The Official Microsoft Blog. https://blogs.microsoft.com/blog/2016/03/25/learning-tays-introduction/
Li, Y., Guan, M., Hammond, P., & Berrey, L. E. (2021). Communicating COVID-19 information on TikTok: A content analysis of TikTok videos from official accounts featured in the COVID-19 information hub. Health Education Research.
Literat, I. (2021). “Teachers Act Like We’re Robots”: TikTok as a Window Into Youth Experiences of Online Learning During COVID-19. AERA Open
Lloyd, A. (2019). Chasing Frankenstein’s monster: Information literacy in the black box society. Journal of Documentation, 75(6), 1475–1485. https://doi.org/10.1108/JD-02-2019-0035
McCorduck, P., & Cfe, C. (2004). Machines who think: A personal inquiry into the history and prospects of artificial intelligence. CRC Press.
Mertala, P. (2020). Data (il) literacy education as a hidden curriculum of the datafication of education. Journal of Media Literacy Education, 12(3), 30–42.
Microsoft. (2016, April 14). Meet Tay—Microsoft A.I. chatbot with zero chill. https://web.archive.org/web/20160414074049/https://www.tay.ai/
Mittelstadt, B. D., Allo, P., Taddeo, M., Wachter, S., & Floridi, L. (2016). The ethics of algorithms: Mapping the debate. Big Data & Society, 3(2), 205395171667967. https://doi.org/10.1177/2053951716679679
Ng, D. T. K., Leung, J. K. L., Chu, S. K. W., & Qiao, M. S. (2021). Conceptualizing AI literacy: An exploratory review. Computers and Education: Artificial Intelligence, 2, 100041. https://doi.org/10.1016/j.caeai.2021.100041
OECD Publications. 2001. Understanding the digital divide.
http://www.oecd.org/decument/51/0,3343,en_2649_33757_1814131_1_1 _1_,00.html
O’Neil, C. (2016). Weapons of math destruction: How big data increases inequality and threatens democracy. Broadway Books.
Perell, D. (2020). Why Did The Boeing 737 Max Crash? David Perell. https://perell.com/essay/boeing-737-max/
Raffaghelli, E. (2020). Is data literacy a catalyst of social justice? A response from nine data literacy initiatives in higher education. Education Sciences, 10(9), 233.
Raffaghelli, J. E., Manca, S., Stewart, B., Prinsloo, P., & Sangrà, A. (2020). Supporting the development of critical data literacies in higher education: Building blocks for fair data cultures in society. Springer.
Reich, J. (2020). Failure to disrupt: Why technology alone can’t transform education. Harvard University Press.
Ruan, S., Wobbrock, J. O., Liou, K., Ng, A., & Landay, J. A. (2017). Comparing Speech and Keyboard Text Entry for Short Messages in Two Languages on Touchscreen Phones. 1(4), 23.
Russell, S., & Norvig, P. (2002). Artificial intelligence: A modern approach.
Spielberg, S. (1993). Jurassic Park. Universal Pictures. https://www.imdb.com/title/tt0107290/
Starkey, L., Sylvester, A., & Johnstone, D. (2017). Negotiating Digital Divides: Perspectives From the New Zealand Schooling System. Journal of Research on Technology in Education, 49(1–2), 31–42. https://doi.org/10.1080/15391523.2017.1292161
Starkey, L., & Zhong, J. (2019). The effect of netbook ownership on children’s academic achievement in mathematics, reading, and writing. Journal of Computer Assisted Learning, 35(1), 34–41. https://doi.org/10.1111/jcal.12308
Tegmark, M. (2017). Life 3.0: Being human in the age of artificial intelligence. Vintage.
Togelius, J. (2020, July 17). GPT-3 often performs like a clever student who hasn’t done their reading trying to bullshit their way through an exam. Some well-known facts, some half-truths, and some straight lies, strung together in what first looks like a smooth narrative. [Tweet]. Twitter. https://twitter.com/togelius/status/1284131360857358337
Turing, A. M. (1950). I.—COMPUTING MACHINERY AND INTELLIGENCE. Mind, LIX(236), 433–460. https://doi.org/10.1093/mind/LIX.236.433
Tutt, A. (2017). An FDA for algorithms. Admin. L. Rev., 69, 83.
Vöcking, B., Alt, H., Dietzfelbinger, M., Reischuk, R., Scheideler, C., Vollmer, H., & Wagner, D. (2010). Algorithms unplugged. Springer Science & Business Media.
Vöcking, B., Alt, H., Dietzfelbinger, M., Reischuk, R., Scheideler, C., Vollmer, H., & Wagner, D. (Eds.). (2011). Algorithms Unplugged. Springer Berlin Heidelberg. https://doi.org/10.1007/978-3-642-15328-0
Waluyo, B. (2019). The Effects of ICT on Achievement: Criticizing the Exclusion of ICT from World Bank’s Education Sector Strategy 2020. Malaysian Online Journal of Educational Technology, 7(2), 71–87. https://doi.org/10.17220/mojet.2019.02.005
Weller, C. (2018). Silicon Valley parents are raising their kids tech-free—And it should be a red flag. Business Insider. https://www.businessinsider.com/silicon-valley-parents-raising-their-kids-tech-free-red-flag-2018-2
Wing, J. M. (2006). Computational thinking. Communications of the ACM, 49(3), 33–35.
Zubarev, V. (2018). Machine Learning for Everyone. Vas3k Blog. http://vas3k.com/blog/machine_learning/
[1] The italicized text was generated by the author. An NLP system available at https://textsynth.com/playground.html used the GPT-J 6B model and the words “Machine learning has enabled the spread of fake news” as the input text.
Current Issues
- Media and Information Literacy: Enriching the Teacher/Librarian Dialogue
- The International Media Literacy Research Symposium
- The Human-Algorithmic Question: A Media Literacy Education Exploration
- Education as Storytelling and the Implications for Media Literacy
- Ecomedia Literacy
Leave a Reply