Cyberethics, Data Surveillance, and the Classroom: Addressing Corporate Control of Generative AI to Protect Student Learning Environments

Research output: Contribution to conferencePosterpeer-review

Abstract

Generative AI (GenAI) is rapidly being incorporated into higher education assignments, often with lesson plans promoted without mention of data privacy (Walker et al., 2023). This proposal argues that unlike previous educational technologies, GenAI presents a significantly elevated risk to student privacy and intellectual freedom due to industry consolidation.

The massive resources required to train GenAI algorithms create conditions for a natural monopoly, consolidating control among a few very large technology corporations (e.g., Alphabet/Google, Microsoft, Meta) (Vipra & Myers West, 2023). This contrasts sharply with past educational technologies, which were often managed in-house or by much smaller educational technology firms. GenAI's reliance on continuous, two-way data exchange via APIs means student behavioral data is shared with companies whose core business model is rooted in surveillance capitalism.

Consequently, requiring GenAI use forces students into an ethically fraught situation: they must consent to corporate surveillance and give up intellectual freedom as a condition for course completion, often through complex and ineffective terms of service (Tokson, 2024). We outline the structural privacy challenges and provide practical, transdisciplinary strategies—including leveraging privacy literacy instruction from librarians—to empower students and mitigate the ethical risks of required GenAI use in the classroom.

Tokson, M. (2024). Government purchase of private data. Wake Forest Law Review, 59(1), 269–324. https://www.wakeforestlawreview.com/wp-content/uploads/2024/10/Tokson_Final.pdf

Vipra, J. & Myers West, S. (2023). Computational power and AI. AI Now Institute. https://ainowinstitute.org/wp-content/uploads/2023/09/AI-Now_Computational-Power-an-AI.pdf

Walker, K. L., Bodendorf, K., Kiesler, T., de Mattos, G., Rostom, M., & Elkordy, A. (2023). Compulsory technology adoption and adaptation in education: A looming student privacy problem. Journal of Consumer Affairs, 57(1), 445-478. https://doi.org/10.1111/joca.12506
Original languageAmerican English
StatePublished - Feb 26 2026
EventScholarship of Teaching and Learning Commons Conference - Savannah, GA, United States
Duration: Feb 26 2026Feb 28 2026

Conference

ConferenceScholarship of Teaching and Learning Commons Conference
Country/TerritoryUnited States
CitySavannah, GA
Period02/26/2602/28/26

UN SDGs

This output contributes to the following UN Sustainable Development Goals (SDGs)

  1. SDG 4 - Quality Education
    SDG 4 Quality Education
  2. SDG 16 - Peace, Justice and Strong Institutions
    SDG 16 Peace, Justice and Strong Institutions

Scopus Subject Areas

  • Education
  • Economics and Econometrics
  • Artificial Intelligence
  • Management of Technology and Innovation

Disciplines

  • Economics
  • Artificial Intelligence and Robotics

Fingerprint

Dive into the research topics of 'Cyberethics, Data Surveillance, and the Classroom: Addressing Corporate Control of Generative AI to Protect Student Learning Environments'. Together they form a unique fingerprint.

Cite this