Smart Glasses, Academic Freedom, and the Risks of Frictionless Learning
Exploring governance challenges of inclusivity, surveillance, and synthetic media in higher education
How to cite this learning scenario
Smart Glasses Lab (2025). Smart Glasses, Academic Freedom, and the Risks of Frictionless Learning (2025). www.AI4education.org. Licensed under a Creative Commons Attribution 4.0 International License.
abstract
This scenario is based on Arantes, Welsman & Marland (2025), “Smart glasses’ boost to inclusivity could cost us our academic freedom”. The article highlights the tension between the accessibility potential of smart glasses and the profound governance risks they introduce.
Smart glasses promise frictionless participation in higher education: students can instantly translate texts, transcribe speech, or reorganise content for their needs. This is life-changing for students with disabilities, dyslexia, or neurodiverse learning styles. However, indistinguishable from ordinary eyewear, these devices open new vectors for cheating, surveillance, and manipulation. As the article warns, smart glasses could normalise hyper-monitoring in classrooms, capturing eye movement, attention levels, and emotional cues. They also generate data that can be weaponised into deepfakes or used for doxing. Such risks threaten the very conditions of academic freedom, as staff and students self-censor for fear of hidden recording. While policies such as GDPR and the AI Act address privacy in broad terms, they fall short in tackling academic freedom and the psychosocial effects of surveillance. The absence of anticipatory governance risks embedding distrust, silence, and reputational harms into everyday teaching. This scenario asks whether universities can establish frameworks that protect inclusivity without sacrificing freedom. It underscores that “a little friction” in learning is not a barrier but a safeguard, and that governance must evolve before the technology becomes ubiquitous.
Smart glasses promise frictionless participation in higher education: students can instantly translate texts, transcribe speech, or reorganise content for their needs. This is life-changing for students with disabilities, dyslexia, or neurodiverse learning styles. However, indistinguishable from ordinary eyewear, these devices open new vectors for cheating, surveillance, and manipulation. As the article warns, smart glasses could normalise hyper-monitoring in classrooms, capturing eye movement, attention levels, and emotional cues. They also generate data that can be weaponised into deepfakes or used for doxing. Such risks threaten the very conditions of academic freedom, as staff and students self-censor for fear of hidden recording. While policies such as GDPR and the AI Act address privacy in broad terms, they fall short in tackling academic freedom and the psychosocial effects of surveillance. The absence of anticipatory governance risks embedding distrust, silence, and reputational harms into everyday teaching. This scenario asks whether universities can establish frameworks that protect inclusivity without sacrificing freedom. It underscores that “a little friction” in learning is not a barrier but a safeguard, and that governance must evolve before the technology becomes ubiquitous.
“Frictionless learning is a nice idea. But when it comes to academic freedom, a little friction is not a flaw: it is a vital safeguard.”
Frictionless learning is a nice idea. But....
In 2027, a leading Australian university implemented a policy explicitly allowing students to use smart glasses across all teaching spaces and assessments. The decision was influenced by research such as Arantes, Welsman & Marland (2025), which emphasised both the inclusivity potential and the dangers of inaction.
At first, the benefits were undeniable. Students with dyslexia used the glasses to support learning in real time, while hearing-impaired students accessed seamlessspeech to text captions in lectures. Vision-impaired students reported unprecedented autonomy through instant text-to-speech overlays. The institution promoted the policy as a landmark achievement in equity and accessibility.
Yet, as the article had warned, governance blind spots soon surfaced. In high-stakes exams, several students were caught using smart glasses to discreetly photograph and transmit questions to external tutors. Attempts to pivot to viva voce assessments quickly failed when staff realised the same devices could be used for covert prompts during oral exams.
In classrooms, staff began noticing shifts in behaviour. Discussions on politically or ethically charged issues grew muted. Some students later admitted they were reluctant to voice opinions for fear of being secretly recorded and potentially misrepresented online. The indistinguishable nature of the glasses created a climate of uncertainty: who was recording, and how might that footage be used?
The risks escalated when a senior lecturer was targeted in a deepfake scandal. Using data harvested from classroom footage, a video was circulated online that falsely depicted the lecturer making racist remarks. The deepfake, created with just a handful of images, spread rapidly across social media, triggering harassment and reputational damage. The lecturer took leave, citing fears for their safety.
A taskforce was established, drawing explicitly on the concerns raised. While privacy frameworks offered some protections, they did not address the erosion of academic freedom through normalised surveillance. Nor did they cover the psychosocial toll on staff silenced by fear of being deepfaked or doxed.
By 2030, the paradox was clear: smart glasses had simultaneously expanded inclusivity and entrenched surveillance. Students who needed the technology thrived, yet staff and peers increasingly engaged in self-censorship. Academic freedom, as Arantes, Welsman & Marland (2025) had predicted, was weakened by habituated surveillance.
The scenario forces universities to confront the trade-off: can inclusivity be delivered without embedding pervasive risks? Or must higher education accept the silencing of academic voices as the hidden cost of digital equity?
Research Topics
Research Questions
How does the presence of smart glasses influence classroom participation and willingness to speak freely?
What forms of governance best balance inclusive accessibility with the need to protect academic freedom?
How can higher education institutions anticipate and mitigate risks of reputational harm through deepfakes?
How wearables reshape academic freedom and intellectual risk-taking.
The governance of deepfakes and doxing in higher education workplaces.
Policy design for invisible cheating technologies.
Data Collection
Activity 1: Compare and Contrast Frameworks
Task: Using disability inclusion and academic freedom frameworks, map your institution’s current smart device / wearbales policies. Where do tensions emerge?
Activity 2: Deepfake Preparedness Audit
Task: Survey staff about their awareness of deepfake risks and available protections. Identify institutional gaps.
Activity 3: Ethnographic Reflection Groups
Task: Facilitate student focus groups to explore feelings about being observed through smart glasses. How do perceptions of hidden recording alter dialogue and learning?