5 Peer Reviewed Articles
This annotated bibliography includes peer reviewed research, pre-print research, and professional blogs for educators. The annotations are still a work in progress, so thank you for your patience! Where abstracts or author summaries are available, I have included these.
Anderson, J. & Rainie, L. (2023). As AI Spreads, Experts Predict the Best and Worst Changes in Digital Life by 2035. Pew Research Center https://www.pewresearch.org/internet/wp-content/uploads/sites/9/2023/06/PI_2023.06.21_Best-Worst-Digital-Life_2035_FINAL.pdf?mibextid=Zxz2cZ
This report covers results from the 16th “Future of the Internet” canvassing that Pew Research Center and Elon University’s Imagining the Internet Center have conducted together to gather expert views about important digital issues. This is a nonscientific canvassing based on a nonrandom sample; this broad array of opinions about where the potential influence of current trends may lead society between 2023 and 2035 represents only the points of view of the individuals who responded to the queries. Pew Research Center and Elon’s Imagining the Internet Center sampled from a database of experts to canvass from a wide range of fields, inviting entrepreneurs, professionals and policy people based in government bodies, nonprofits and foundations, technology businesses and think tanks, as well as interested academics and technology innovators. The predictions reported here came in response to a set of questions in an online canvassing conducted between Dec. 27, 2022, and Feb. 21, 2023. In all, 305 technology innovators and developers, business and policy leaders, researchers and activists responded in some way to the question covered in this report.
D’Agostino, S. (2023, June 5). How AI tools both help and hinder equity, Inside Higher Ed. https://www.insidehighered.com/news/tech-innovation/artificial-intelligence/2023/06/05/how-ai-tools-both-help-and-hinder-equity
The technology promises to assist students with disadvantages in developing skills needed for success. But AI also threatens to widen the education gap like no tech before it.
Eaton, S. E., Dawson, P., McDermott, B., Brennan, R., Wiens, J., Moya, B., Dahal, B., Hamilton, M., Kumar, R., Mindzak, M., Miller, A., & Milne, N. (2023). Understanding the Impact of Artificial Intelligence on Higher Education. Calgary, Canada. https://hdl.handle.net/1880/116624
Eaton, S. (2023).Academic Integrity and Artificial Intelligence: Research Project Update. Learning, Teaching, and Leadership [Blog]. https://drsaraheaton.wordpress.com/2023/06/22/academic-integrity-and-artificial-intelligence-research-project-update/
This blog post contains numerous links to peer reviewed research, presentations, and other resources for college educators on the use of AI in the classroom.
Eloundou, T., Manning, S., Mishkin, P., & Rock, D. (2023). Gpts are gpts: An early look at the labor market impact potential of large language models. arXiv preprint arXiv:2303.10130. https://arxiv.org/abs/2303.10130
From section 4.3: “Our findings indicate that the importance of science and critical thinking skills are strongly negatively associated with exposure, suggesting that occupations requiring these skills are less likely to be impacted by current language models. Conversely, programming and writing skills show a strong positive association with exposure, implying that occupations involving these skills are more susceptible to being influenced by language models.”
Gadiraju, V., Kane, S., Dev, S., Taylor, A., Wang, D., Denton, E., & Brewer, R. (2023, June). ” I wouldn’t say offensive but…”: Disability-Centered Perspectives on Large Language Models. In Proceedings of the 2023 ACM Conference on Fairness, Accountability, and Transparency (pp. 205-216). https://research.google/pubs/pub52358/
Large language models (LLMs) trained on real-world data can inadvertently reflect harmful societal biases, particularly toward historically marginalized communities. While previous work has primarily focused on harms related to age and race, emerging research has shown that biases toward disabled communities exist. This study extends prior work exploring the existence of harms by identifying categories of LLM-perpetuated harms toward the disability community. We conducted 19 focus groups, during which 56 participants with disabilities probed a dialog model about disability and discussed and annotated its responses. Participants rarely characterized model outputs as blatantly offensive or toxic. Instead, participants used nuanced language to detail how the dialog model mirrored subtle yet harmful stereotypes they encountered in their lives and dominant media, e.g., inspiration porn and able-bodied saviors. Participants often implicated training data as a cause for these stereotypes and recommended training the model on diverse identities from disability-positive resources. Our discussion further explores representative data strategies to mitigate harm related to different communities through annotation co-design with ML researchers and developers.
Lorentzen, A. and Bonner, E. (2023, February 12). Customizable ChatGPT AI Chatbots for Conversation Practice. The FLT Mag. https://fltmag.com/customizable-chatgpt-ai-chatbots-for-conversation-practice/
Since OpenAI’s AI can also be accessed and interacted with using a little bit of computer code called an API, we decided to go the extra step and develop a simple application using Unity, some 3D characters and some C# coding that would give the AI a customizable voice, accent and appearance. This way our students could not only create their own personalized AI conversation partners, but also choose how they look and sound. Finally, using Google Speech-to-Text for the students and Text-to-Speech for the AI, our students spoke with their AI directly and heard and saw the AI respond in kind.
Mollick, E. R. (2023, February 17). My class required AI. Here’s what I’ve learned so far. One Useful Thing. https://www.oneusefulthing.org/p/my-class-required-ai-heres-what-ive?subscribe_prompt=free
I fully embraced AI for my classes this semester, requiring students to use AI tools in a number of ways. This policy attracted a lot of interest, and I thought it worthwhile to reflect on how it is going so far. The short answer is: great! But I have learned some early lessons that I think are worth passing on.
Mollick, E. R., & Mollick, L. (2023). Using AI to implement effective teaching strategies in classrooms: Five strategies, including prompts. https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4391243
This paper provides guidance for using AI to quickly and easily implement evidence-based teaching strategies that instructors can integrate into their teaching. We discuss five teaching strategies that have proven value but are hard to implement in practice due to time and effort constraints. We show how AI can help instructors create material that supports these strategies and improve student learning. The strategies include providing multiple examples and explanations; uncovering and addressing student misconceptions; frequent low-stakes testing; assessing student learning; and distributed practice. The paper provides guidelines for how AI can support each strategy, and discusses both the promises and perils of this approach, arguing that AI may act as a “force multiplier” for instructors if implemented cautiously and thoughtfully in service of evidence-based teaching practices.
Mollick, E. R., & Mollick, L. (2022). New modes of learning enabled by AI chatbots: Three methods and assignments. Available at SSRN. https://papers.ssrn.com/sol3/papers.cfm?abstract_id=4300783
Chatbots are able to produce high-quality, sophisticated text in natural language. The authors of this paper believe that AI can be used to overcome three barriers to learning in the classroom: improving transfer, breaking the illusion of explanatory depth, and training students to critically evaluate explanations. The paper provides background information and techniques on how AI can be used to overcome these barriers and includes prompts and assignments that teachers can incorporate into their teaching. The goal is to help teachers use the capabilities and drawbacks of AI to improve learning
Sayers, D. (2023, May 25). A simple hack to ChatGPT-proof assignments using Google Drive. Times Higher Education. https://www.timeshighereducation.com/campus/simple-hack-chatgptproof-assignments-using-google-drive
Sayers explains how he uses Google Docs version history to track potential misuse of generative AI in student assignments (note from Liza: this may prove more difficult now that generative AI writing assistant is embedded in Google Docs, but it still may be a good idea to try).
Zamfirescu-Pereira, J. D., Wong, R. Y., Hartmann, B., & Yang, Q. (2023, April). Why Johnny can’t prompt: how non-AI experts try (and fail) to design LLM prompts. In Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems (pp. 1-21). https://dl.acm.org/doi/pdf/10.1145/3544548.3581388
Pre-trained large language models (“LLMs”) like GPT-3 can engage in fluent, multi-turn instruction-taking out-of-the-box, making them attractive materials for designing natural language interactions. Using natural language to steer LLM outputs (“prompting”) has emerged as an important design technique potentially accessible to non-AI-experts. Crafting effective prompts can be challenging, however, and prompt-based interactions are brittle. Here, we explore whether non-AI-experts can successfully engage in “end-user prompt engineering” using a design probe—a prototype LLM-based chatbot design tool supporting development and systematic evaluation of prompting strategies. Ultimately, our probe participants explored prompt designs opportunistically, not systematically, and struggled in ways echoing end-user programming systems and interactive machine learning systems. Expectations stemming from human-to-human instructional experiences, and a tendency to overgeneralize, were barriers to effective prompt design. These findings have implications for non-AI-expert-facing LLM-based tool design and for improving LLM-and-prompt literacy among programmers and the public, and present opportunities for further research.