Higher Education in the Age of Artificial Intelligence
Timo Seidl (Technical University of Munich)
www.timoseidl.com
I. The Frontier is Jagged but the Hype is Real — So Let’s Get Real Too
II. Teaching: Why we need more Damokles, less Prokrustes, and a bit of Hercules
III. Research: Let’s Not Let a Serious Revolution Go to Waste
Let me be clear: there is an attention economy surrounding claims about AI’s transformative impact, which sometimes drowns out nuance and caution. Benchmarks never tell the full story, and there are well-founded objections even to the most popular ones. And there are profound risks — from extraction to economic upheavals, from safety to the concentration of power.
But if most benchmarks point in the same direction and if many serious — and often initially skeptical — people agree that something profoundly changed sometime around late 2025, you should perhaps at least pay attention—and dip your toes in it. Yet, large parts of academia still seem to cycle through the first two stages of grief, mixing outright denial with misdirected anger.
Three Tropes of Skepticism
Deepities Seemingly profound dismissals that are trivially true but profoundly misleading
Non-Sequiturs Treating a refutation of the hype as a refutation of the capability
Humanist Copium Insisting, by sentimental fiat, that only humans can really think, judge, or create
Like ‘love is just a word’, statements like ‘LLMs just predict the next token’ are deepities in the sense that Daniel Dennett used the word: seemingly profound on a first reading, but ambiguous and shallow on a second. They are trivially true but profoundly misleading.
Yes, love is just another word like igloo or institution—but that tells us nothing about whether the emotion is real. Similarly, LLMs do predict next tokens — but sufficiently good prediction requires modelling the world that language is about, not to mention that newer architectures also reason, plan, backtrack.
‘Artificial intelligence, if we’re being frank, is a con: a bill of goods you are being sold to line someone’s pockets. A few major well-placed players are poised to accumulate significant wealth by extracting value from other people’s creative work, personal data, or labor, and replacing quality services with artificial facsimiles. (…) We call this type of con “AI hype”’ (Bender and Hanna 2025, 4).
The hype is real: tech companies oversell, extract, and exaggerate—and yes, tech bros can be obnoxious. But the capabilities are also real—working with these models often is like working with wizards. To paraphrase Tibor Rutar, if you’re not astonished, you’re probably not using it (correctly).
Treating a refutation of the hype as a refutation of the capability is simply a non-sequitur. As Kahneman and Tversky put it long ago, ‘A refutation of a caricature can be no more than a caricature of a refutation.’
‘I’m not interested in reading something that nobody wrote. I read because I want to understand how somebody sees something, and there’s no “somebody” inside the synthetic text-extruding machines.’ Emily Bender (June 2025)
‘Social scientists produce useful, meaningful knowledge because of their capacity for judgment. Judgment and discernment are fundamentally interpretive tasks that by definition cannot be outsourced to a machine.’ Erin Lockwood (March 2026)
‘Where real thinking involves organic associations, speculative leaps, and surprise inferences, AI can only recognize and repeat embedded word chains, based on elaborately automated statistical guesswork. (…) AI, not the human mind, is the weak, narrow, crude machine.’ n+1 magazine (Fall 2025)
Just declaring that AI cannot do certain things is not a serious intellectual position, it’s copium — reminiscent of insisting that China ‘can only copy,’ and arguably born of the same refusal to, well, take the competition seriously. I find two things particularly problematic:
First, placing so much emphasis on the source rather than value of knowledge. Kevin Bryan put it best: ‘I get the desire for artisanal, hand-crafted research, with the matrices hand-inverted. But our job is to move the frontier of knowledge, not self-actualization.’
Second, the blatant hypocrisy of holding AI to the highest possible standards if we constantly let humans get away with epistemic murder. As Alexander Kustov put it provocatively, ‘If we applied the same skepticism to human-produced research that we apply to AI outputs, we’d shut down half the journals tomorrow.’
Reliability Remains fragile: reliability lags capability, hallucination is rarer but unsolved, and even frontier models still often make basic errors.
Performance Degrades on long tasks, in complex environments, and wherever success is hard to verify at scale.
Adaptability No persistent learning across tasks, poor transfer to messy real-world contexts, and limited ability to correct course within a task.
Booksmart, not streetsmart Models ace benchmarks but make strange, unintuitive errors, lacking the practical judgment, common sense, and accumulated experience that come from actually doing things.
The meh factor AI output often still feels just a bit meh: competent but mediocre, fluent but flat, correct but lifeless.
Adapted from/Inspired by Posts by Tomas Pueyo, Colin Fraser and Alberto Romero
Adapted from/Inspired by Posts by Tomas Pueyo, Colin Fraser and Alberto Romero
Adapted from/Inspired by Posts by Tomas Pueyo, Colin Fraser and Alberto Romero
‘Too much academic discourse treats AI as a taboo to be avoided or a fad to be ignored. We need to take our heads out of the sand: these tools are becoming a competitive necessity, and pretending otherwise is a disservice to the next generation [and to science itself].’ — Kiran Garimella
‘I use technology in order to hate it properly.’ — Nam June Paik
AI is already transformative — regardless of whether progress soon hits a wall, the bubble bursts, or ‘sparks’ of a more general intelligence kindle a fire. So instead of getting surprised and blind-sided by events, let us start to think seriously, proactively, and, yes, critically about what AI means for teaching and research in higher education. And don’t just reflect on it, talk about it, design around it, play with it. Even if you hate AI: start hating it properly.
Many have raised the alarm about a ‘cheating vibe shift’, the ‘death of the student essay’, and universities’ ‘losing battle against AI cheating’. Student and faculty surveys are clear: students use generative AI all the time, they use it for tests, they feel unsure about how to use it; faculty, meanwhile, are concerned and feel left in the dark.
There is no question: cheating needs to be reined in, exams designed to prevent it, and a sanction regime established. But fighting the fever should not distract us from addressing the underlying infection: the transformation of cognitive labor and the risk of cognitive atrophy.
‘College is just how well I can use ChatGPT at this point.’
‘I spend so much time on TikTok. Hours and hours, until my eyes start hurting, which makes it hard to plan and do my schoolwork. With ChatGPT, I can write an essay in two hours that normally takes 12.’
‘I really like writing. Honestly, I think there is beauty in trying to plan your essay. You learn a lot. You have to think, Oh, what can I write in this paragraph? Or What should my thesis be? [But] an essay with ChatGPT, it’s like it just gives you straight up what you have to follow. You just don’t really have to think that much.’
Quotes by various students, cited in Walsh 2025
‘Learning results from what the student does and thinks and only from what the student does and thinks. The teacher can advance learning only by influencing what the student does to learn.’ — Herbert Simon quoted in Ambrose et al. (2010).
‘The Problem with AI is not that it encourages cheating. The problem with AI is that it discourages learning. (…) We’ve been focused on how students use AI to cheat. What we should be more concerned about is how AI cheats students.’ — Nicolas Carr
Why we cannot (really) outsource thinking
To the extent that AI reads and writes for you, it thinks for you. Thinking is inextricably linked up with the labor of learning: sifting through data, brooding over research papers, writing and rewriting an argument.
You can outsource the seemingly tedious parts of research—literature reviews, data cleaning, drafting—but only at the cost of compromising the creative parts: novel ideas, cogent explanations, original insight. The dull and the creative are, for better or worse, inseparable.
Why we should (really) not outsource thinking
Outsourcing tasks involving critical thinking and complex language is not like outsourcing arithmetic or navigation (although those come at a cost too). ‘Language’, as the poet W.H. Auden put it, ‘is the mother, not the handmaiden, of thought’.
Without Google Maps, we get disoriented in unfamiliar places. Without learning how to think, we risk losing cognitive autonomy, becoming disoriented and dependent. Like illiterates in the library of Alexandria, we become surrounded by knowledge we cannot access on our own.
The Curious Case of William A.
In 2024, a Tennessee student graduated with a 3.4 GPA—but was unable to read or spell his own name. To complete written assignments, he dictated topics into speech-to-text software, pasted the output into ChatGPT to generate a paper, then ran it through Grammarly.
His school argued that since he could produce the requested output, he’d received an adequate education.
A federal appeals court disagreed: the accommodations had ‘simply done the work for him.’ The school had concentrated on whether assignments were completed—not whether he had learned.
‘You’re asking me to go from point A to point B, why wouldn’t I use a car to get there?’ Student cited in Shirky (April 2025)
You are confusing the output of education with its product. Using AI to do your assignments, as Ted Chiang put it, is like bringing a forklift to the weight room. You go to the gym to get fit, not to lift weights. What you want to do instead is maximise ‘time under tension’. Build cognitive muscle by thinking, slow and hard.
Much as the mirror provides affordances for vanity, LLMs provide affordances for cognitive lethargy. Like sugary snacks at the checkout, they exploit first-order desires at the expense of second-order ones (Frankfurt 1971). Students may want to learn but they also want to finish their assignments. Procrastinating or cutting corners does not require LLMs but is aided and abetted by them. Why use an ‘e-bike for the mind’ if you have a delivery app at your fingertips?
LLMs not only promote ‘metacognitive laziness’: a reduced tendency to plan, monitor, and evaluate one’s own thinking (Fan et al. 2025). They also erode the very thing that makes learning rewarding. Intellectual and personal growth, as Thomas Pfau reminds us, ‘can only ever be the fruit of sustained personal effort’. For education to be more than a ‘relentless series of logistical challenges’, students must experience the struggle and labor of learning.
‘I’ve become lazier. AI makes reading easier, but it slowly causes my brain to lose the ability to think critically or understand every word.’
“Yeah, it’s helpful, but I’m scared that someday we’ll prefer to read only AI summaries rather than our own, and we’ll become very dependent on AI.’
‘I literally can’t even go 10 seconds without using Chat when I am doing my assignments. I hate what I have become because I know I am learning NOTHING, but I am too far behind now to get by without using it. I need help, my motivation is gone. I am a senior and I am going to graduate with no retained knowledge from my major.’
Students cited in Shirky (April 2025)
Outsourcing thinking erodes thinking. In a large field experiment with high schoolers, unrestricted access to GPT-4 improved practice scores but led to worse performance on unassisted exams. A similar experiment with software developers found that AI assistance undermined the acquisition of the very skills it automated.
In Thinking—Fast, Slow, and Artificial, Shaw and Nave (2026) experimentally document what they call ‘cognitive surrender’: the adoption of AI outputs with minimal scrutiny, overriding intuition (System 1) and deliberation (System 2). Given access to ‘System 3’ (AI), participants often use it and rarely override it. When the AI was right, accuracy increased, but when the AI was wrong, accuracy dropped sharply, while confidence remained high. The ‘rise of thinking machines’, it turns out, can lead to the ‘decline of thinking people’.
‘Political writers have established it as a maxim, that, in contriving any system of government (…) every man ought to be supposed a knave, and to have no other end, in all his actions, than private interest. By this interest we must govern him, and, by means of it, make him, notwithstanding his insatiable avarice and ambition, co-operate to public good.’ David Hume, Essays Moral, Political, Literary, 1777
Let’s not kid ourselves, the AI-enabled ‘democratization of cheating’ means that certain types of assignments are no longer viable—at least not without some extra effort. We need better detection and sanctioning, more oral and in-person assessment, and creativity in designing new assignment formats.1 This is not because all students cheat or want to cheat—but because some do. When designing rules, we need to think of the knaves.
Yet a supposedly AI-proof constitution for knaves has two central problems:
It ignores that assignments are not just assessment formats but learning vehicles. Making curricula cheat-proof by eliminating certain assignments is like draining the pool to prevent drowning.
It ignores that designing around incentives only crowds out intrinsic motivation—distrust breeds distrust which undermines the disposition to learn (Bowles 2016).
‘I write entirely to find out what I’m thinking, what I’m looking at, what I see and what it means.’ Joan Didion
‘It’s a cliche, but it’s true: learning how to write is learning how to think. Writing a paper isn’t just the process of translating thoughts that already exist in your head onto the page; rather, it’s only when you have to commit your thoughts to writing that you are forced to clarify them to yourself.’ Daniel Greco
‘The demise of writing matters, because writing is not a second thing that happens after thinking. The act of writing is an act of thinking.’ Derek Thompson
‘The one-two punch of reading and writing is like the serum we have to take in a superhero comic book to gain the superpower of deep symbolic thinking, and so I have been ringing this alarm bell that we have to keep taking the serum.’ Cal Newport
‘Writing compels us to think — not in the chaotic, non-linear way our minds typically wander, but in a structured, intentional manner. By writing it down, we can sort years of research, data and analysis into an actual story, thereby identifying our main message and the influence of our work.’ Nature Reviews Bioengineering
‘[AI] doesn’t add to your thinking. It replaces the part where you’re supposed to struggle. Struggling with sentences is the point. That’s how the thinking sharpens.’ Mark Blyth
In the myth, Damocles envied the king — until he saw the sword hanging above the throne. Instead of giving up on written or other hard-to-cheat-proof assignments, we should create similar Damocles effects: pair every submission with an accountability moment that hangs over the process like the mythological sword.
AI may be able to write a paper for you. But it can’t save you from embarrassing yourself if you have to explain, defend, and communicate it.
In the myth, Procrustes offered travellers a bed — then stretched or amputated them to fit it. Fixed semester lengths, standardized exam formats, rigid credit structures often do something similar: they force the learning process to fit inherited institutional furniture.
AI is like a forcing function that reveals how poorly learning actually fits these constraints.
In the myth, the young Hercules meets two figures at a crossroads: Vice, who offers ease and pleasure, and Virtue, who offers a harder path leading to genuine accomplishment. With AI, we face this choice over and over.
The, well, Herculean task is to get students to want to learn and convince them (and ourselves) that most everything worth having is on the other side of hard — and to explore and provide the tools and ideas that make the hard choice easier.
Example Learning Opportunities is a Claude Code skill for deliberate skill development, developed by Cat Hicks. Drawing on well-established findings in learning science, it deliberately interrupts the speed of a typical agentic coding session to help you reflect on and explore your generated work. Instead of just answering, the AI pauses and waits, pushing against its own default fluency to create space for the learner’s own thinking. Techniques include active generation (predictions, explanations, sketches), retrieval practice (check-ins, teach-it-back, self-testing), deliberate pauses (spacing, reflection), and explicit metacognition (self-assessment, gap identification) (on personalized AI-tutoring, see also Chung et al. 2026).
Even before the advent of AI, science was in serious trouble. Industrialized fraud (paper mills), the abject epistemic and moral failure of how we currently do peer review, ever-increasing administrative loads, and incentive structures that militate against risk-taking — have all contributed to a production-progress paradox: scientific production is booming. Scientific progress is not.
Whatever AI will bring, it is arriving at a time when science is already immunocompromised.
To think that AI will, by itself, speed up scientific progress tout court is, as Narayanan and Kapoor put it, like ‘adding lanes to a highway when the slowdown is actually caused by a tollbooth.’
AI can undoubtedly make individual scientists more productive—even if it makes their research interests narrower (e.g. Hao et al. 2026). But more productive scientists will only lead to more scientific progress if we reform the institutions and norms of academic publishing. As of today, AI mostly acts as an ‘accelerant for the existing optimization machine, (…) offering a faster way to produce the artifacts the system rewards’—perpetuating and intensifying the existing reward structure, attention scarcity, and pressure toward safe, incremental work.
As it becomes easier and easier to produce decent — or at least decent-looking — papers, we need to start thinking about the epistemic and attentional externalities of ‘monstrous scientific overproduction’: each paper requires attention from editors, reviewers, and readers, and any further increase risks overwhelming scientists’ collective capacity for validation and digestion (for good starts, see here and Munger et al. 2026).
Like other forms of pollution, the answer must involve some form of tax on production, so that publishing only pays if a paper’s contribution justifies the burden it places on others. Whether that tax takes the form of stricter norms, submission caps, or incentives that reward depth over volume is an open question. But the direction seems clear: we need fewer and better papers—and maybe rethink the paper format altogether!
While AI disrupts and destabilizes the current publishing equilibrium, it can also help — if supported by the right institutional changes — to move us toward a new one of fewer and better papers. The key is to stop thinking of AI as a tool allowing you to produce more and start thinking of it as a collaborator allowing you to do better.
Claes Bäckman, for example, has set up a collection of Claude Code skills that helps you with everything from spelling and academic style to tables and figures, and even provides (journal-specific) adversarial reviews. Tools like this can dramatically speed up feedback and democratize access to expertise.
If set up right, AI can place you in rooms full of experts: theorists, methodologists, editors who are intimately familiar with your work and can not just write, code, and iterate through different analyses, but also provide pushback, ask follow-up questions, and surface connections you missed.
Ideas
Or why it’s all about parkour vision
Expertise
Or why really knowing stuff is still key
Taste
Or why we could all use some phronesis
Voice
And why you need to find your own
Where people walking see walls, handrails, and gaps, parkour athletes see vertical floors, pathways, spaces to be jumped. This is called ‘parkour vision’ — seeing potential for interaction and exploration where others see nothing at all. Jasmine Sun argues that there is an analogous thing called ‘software vision’ that allows programmers to see ‘software-shaped problems’ everywhere. Social scientists need something similar: the ability to see ‘research-shaped problems’.
As AI makes it easier to find answers, the most interesting ideas and insights will come from those who can ask the right questions — or ask them in the right way. This fundamentally requires the ability to recast old problems in new ways, reframe something obvious as puzzling (or the other way around), or ask what it would mean for something familiar if we assumed something unfamiliar about it (Peirce 1997). Mills (2000) called this the ‘sociological imagination’ — but we can also call it parkour vision for social scientists.
Just as AI makes execution less important, it makes deep expertise more important. AIs still frequently get things wrong, miss or misunderstand context, go down irrelevant rabbit holes, start from questionable assumptions, and get stuck in local optima. As Kevin Baker put it, ‘absurdity is not self-correcting’ — and it requires competent researchers to detect when something is either plain wrong or just doesn’t smell right.
Only having real expertise allows you to tell, as John List put it, the difference between ‘nearly right’ and ‘right’, which will be ‘more valuable than ever’. Expertise also allows you to use AI much more effectively, pointing it in the right directions, helping it avoid unproductive lines of reasoning, or providing essential context. Just like an experienced gamer can do a lot more with the same cheat code, or an experienced piano player much more with a music model, expert researchers benefit disproportionally from AI.
‘If it’s clear and useless, it’s useless. It’s organized and useless, it’s useless. It’s persuasive and useless, it’s useless.’ Larry McEnerney
‘If anyone can produce a competent empirical paper on any topic, the bottleneck moves to identifying which questions are important to ask in the first place.’ Seva Gunitsky
An increasingly important skill, especially at a time of great social and technological upheaval, is the ability to ask questions that truly matter, and interpret results in ways that connect them to what truly matters.
Aristotle called this phronesis or practical wisdom—which he distinguished from scientific knowledge (episteme) or technical skill (techne). Isaiah Berlin called it ‘a sense of reality’ (Berlin 2019). We can just call it taste or discernment.
Frontier LLMs are surprisingly poor writers, in some ways worse than earlier, much less capable models. With their rough edges sanded down during post-training, their style is strangely disembodied, emotionally lobotomized, compulsively even-handed. This is not straightforward to fix as the very whimsicality that makes good writing may not be what you—or your customers—want from a model in other domains. In addition to writing bland prose, LLMs, perhaps due to their very architecture, also shrink linguistic diversity and amplify ‘certain sophisticated linguistic patterns while narrowing the range of individual expression’ (Sourati et al. 2025, 34).
Given that many people are already getting sick and tired of AI slop, having your own voice, your own distinct style may be a major competitive advantage: a breath of fresh air in a symbolic landscape suffocating from sameness, a well-recognized gift to a variety-starved world.
The great paradox we are facing is this: parkour vision, deep expertise, taste, a unique voice are all built through the labor of learning — tedious reading of technical texts, long days (and nights) with data, the productive suffering of writing, the accumulated ‘scar tissue’ from many trials and errors. You need to learn about different methods, theories, and domains in order to get a sense of what could make sense — and to assess whether something does. You can’t prompt what you can’t verbalize, and you can’t use what you don’t understand.
The problem, however, as Timothy Burke put it, is that AI is ‘brutally short-circuiting the processes by which people gain enough knowledge (…) to use the potential of generative AI correctly.’ Rightly or wrongly, junior researchers will increasingly feel that they have little incentive to spend long hours acquiring these skills—nor do senior researchers have much incentive to invest in training them.
The AI debate, Derek Thompson has recently pointed out, is so fixated on how AI will out-skill humans that ‘we miss the many ways that technology can de-skill us.’ As people read and write less, they think less. Whether or not we are witnessing the AI-accelerated ‘dawn of a post-literate society’, the threat of cognitive atrophy is already too serious to ignore.
The stakes are not just practical but political. As Frank Herbert warned in Dune: ‘Once men turned their thinking over to machines in the hope that this would set them free. But that only permitted other men with machines to enslave them.’ Ultimately, the risk is a reversal of the Enlightenment as Kant understood it: a self-incurred return to the inability to think independently.
As this cognitive revolution unfolds, universities will have to strike a delicate balance between setting their students up for success in a fast-changing labor market, which will require teaching them how to use and collaborate with AI; and shoring up their cognitive autonomy, which will require teaching them the ‘habits of thinking and reflecting with other people’ that were once at the heart of what is called liberal education in America and Bildung in Germany.
‘Man has, as it were, become a kind of prosthetic God. When he puts on all his auxiliary organs he is truly magnificent; but those organs have not grown on to him and they still give him much trouble at times. Nevertheless, he is entitled to console himself with the thought that this development will not come to an end precisely with the year 1930 A.D. Future ages will bring with them new and probably unimaginably great advances in this field of civilization and will increase man’s likeness to God still more. But in the interests of our investigations, we will not forget that present-day man does not feel happy in his Godlike character.’
Sigmund Freud, Civilization and Its Discontents (1930)