The Death of the Expert: How AI Flattened Knowledge and What We Lost
AI and Society

The Death of the Expert: How AI Flattened Knowledge and What We Lost

Everyone sounds competent now. Fewer people actually are.

The Consultant Who Knew Everything

Last month I sat in a meeting where a junior consultant delivered a flawless presentation on microservices architecture. Perfect terminology. Correct diagrams. Accurate descriptions of event-driven patterns, CQRS, saga orchestration. Every slide looked like it came from someone with a decade of distributed systems experience.

He had six months of experience. And an AI subscription.

Nobody in the room could tell. The presentation was polished, technically accurate, and confident. The questions he fielded were handled smoothly — the AI had anticipated most of them. It wasn’t until someone asked about a specific failure mode they’d encountered in production, something messy and particular and not in any documentation, that the facade cracked.

He didn’t know what he didn’t know. And that’s the problem we’re all facing now.

We’ve entered an era where the gap between sounding like an expert and being one has collapsed almost entirely. AI tools can generate expert-level prose on virtually any topic. They can produce code that compiles, legal briefs that cite real cases, medical summaries that use correct terminology, financial analyses with proper ratios. The output looks right. It reads right. It often is right.

But something crucial has been lost in this flattening. The difference between generating correct information and understanding it. Between producing an answer and knowing why it’s the answer. Between sounding competent and being competent.

I’ve been writing about technology for long enough to have watched this shift happen in real time. Three years ago, you could still tell an AI-generated article from a human-written one. Today, you genuinely cannot. The gap closed faster than anyone predicted.

This article examines what happens when everyone gains access to expert-sounding output. What we’ve gained is obvious. What we’ve lost is harder to see. And what comes next matters more than most people realize.

The Knowledge Flattening Effect

There’s a useful way to think about what AI has done to expertise. Imagine knowledge as a landscape with peaks and valleys. The peaks are deep specialists — people who spent years climbing a particular mountain. The valleys are generalists or beginners. The landscape had dramatic relief. You could see who was high up and who wasn’t.

AI has flattened this landscape. Not by raising the valleys to the peaks — that would be genuine democratization — but by wrapping everything in a thick fog at a uniform altitude. From a distance, the terrain looks flat. Everyone appears to be at the same height.

graph LR
    A["Before AI<br/>Sharp peaks of expertise<br/>Visible skill gaps"] --> B["After AI<br/>Fog layer at medium altitude<br/>Everyone looks similar"]
    B --> C["Result<br/>Harder to identify<br/>genuine expertise"]

The junior developer can produce code that looks like a senior developer wrote it. The marketing intern can write copy that reads like a veteran strategist crafted it. The first-year law associate can draft memoranda that resemble the work of a partner with thirty years of practice.

This isn’t inherently bad. Raising the floor of competence is genuinely valuable. A medical student using AI to draft differential diagnoses is learning faster. A small business owner using AI for legal document review is getting access to knowledge they couldn’t afford before. These are real gains.

But the fog has side effects. When everyone’s output looks the same, we lose the signal that output quality once provided. We can no longer reliably use the work product to assess the worker. The resume looks the same. The writing sample looks the same. The code sample looks the same.

The difference is invisible until it matters. And when it matters, it matters a lot.

The Knowing-Generating Gap

Here is the core distinction that our current moment obscures: there is a fundamental difference between knowing something and generating something that sounds like knowing.

When a cardiologist diagnoses a complex arrhythmia, they’re drawing on thousands of patient interactions, pattern recognition built over years, an intuitive sense of which presentations are dangerous and which are benign. They know what the textbook says. They also know when the textbook is wrong. They know the edge cases, the unusual presentations, the situations where the standard protocol fails.

An AI can generate the same diagnosis. It can list the same differential. It can recommend the same treatment. But there’s no understanding behind it. No accumulated judgment. No ability to notice that something feels off about this particular case.

The person using the AI to generate the diagnosis doesn’t gain any of that understanding either. They gain the output. They don’t gain the knowledge.

This distinction matters in every field:

Software engineering. AI can generate working code. But debugging requires understanding why the code works and where its assumptions break. A developer who generated code with AI but doesn’t understand the architecture will struggle when that code fails in production at 3 AM.

Legal practice. AI can draft contracts and cite precedent. But knowing which precedent matters, how a particular judge interprets specific clauses, and what traps a contract might contain requires judgment built over years. The AI doesn’t know that this particular opposing counsel always exploits ambiguity in indemnification clauses.

Medicine. AI can list symptoms and suggest treatments. But the experienced physician who notices that a patient’s affect doesn’t match their reported pain level, or that a test result is technically normal but unusual for this particular patient — that’s clinical judgment that no amount of generated text can replace.

The pattern is consistent. AI excels at producing outputs that match the surface pattern of expertise. It struggles with the messy, contextual, experiential knowledge that defines genuine expertise.

My cat Arthur demonstrates this principle daily. He has learned that sitting near the treat cupboard correlates with receiving treats. He generates the correct behavior. But he doesn’t understand that the cupboard has a finite supply, that treats come from a store, or that his treat budget is not unlimited. He produces expert-level treat-acquisition behavior without any understanding of the underlying system. He is, in his own way, an AI-assisted treat consultant.

The Hiring Crisis

The knowledge flattening effect has created a genuine crisis in hiring. Every field that relies on evaluating expertise through work products is struggling.

Consider software engineering interviews. For years, companies used coding challenges, take-home projects, and technical discussions to evaluate candidates. A candidate’s ability to write clean code, design systems, and explain trade-offs was a reliable signal of their capability.

That signal is now deeply compromised. A candidate can use AI to complete any take-home project. They can practice technical interviews with AI coaches that simulate every possible question. They can produce portfolios of code that they generated but don’t fully understand.

Some companies have responded by banning AI from their interview processes. This is like banning calculators from math exams — it tests something, but not the thing that matters for the job.

Other companies have shifted to live coding sessions, pair programming interviews, or extended trial periods. These are more expensive and time-consuming, but they test something closer to actual capability.

The best approach I’ve seen comes from a CTO I spoke with last month. Her interview process has three stages: first, she gives candidates a problem and asks them to solve it with whatever tools they want, including AI. Second, she takes their solution and introduces a subtle but breaking change to the requirements. Third, she asks them to debug the issue live.

“The first stage tells me if they can use modern tools,” she explained. “The second tells me if they understand what they built. The third tells me if they can think.”

Most candidates pass stage one. Many fail stage two. Very few excel at stage three.

This pattern scales beyond hiring. How do you evaluate a contractor’s proposal? A consultant’s recommendation? A vendor’s technical assessment? When everyone’s output looks equally professional, the traditional signals of competence become unreliable.

Expertise Inflation

There’s an economic parallel that helps explain what’s happening. Just as monetary inflation devalues currency, we’re experiencing expertise inflation — a devaluation of what it means to sound like an expert.

In the 1990s, having a professional-looking website was a signal of a legitimate business. Then website builders democratized web design, and the signal collapsed. Today, a scam operation can have a better website than a legitimate company. We learned to stop trusting websites as signals of legitimacy.

The same inflation is happening with expertise signals. A well-written report used to indicate a knowledgeable author. A polished presentation used to suggest deep understanding. A comprehensive analysis used to signal years of experience.

Now these signals mean much less. The currency of “sounding expert” has been inflated to the point where its purchasing power — in terms of trust and credibility — has collapsed.

This creates a paradox. Deep expertise is becoming simultaneously more valuable and harder to demonstrate. The expert cardiologist’s judgment is worth more than ever, because AI-generated diagnoses need human oversight. But demonstrating that expertise through traditional means — publications, presentations, written analyses — no longer distinguishes the expert from the AI-assisted novice.

graph TD
    A["Expertise Inflation Cycle"] --> B["AI makes expert-level output easy"]
    B --> C["Expert-sounding output becomes common"]
    C --> D["Trust in output as expertise signal drops"]
    D --> E["Real expertise becomes harder to identify"]
    E --> F["Organizations make worse hiring/trust decisions"]
    F --> G["Demand for verifiable expertise increases"]
    G --> A

We’re caught in a cycle. The more AI democratizes expert-sounding output, the less that output signals actual expertise, which makes actual expertise harder to find, which makes it more valuable, which increases demand for ways to verify it.

How We Evaluated

To understand the scope of the expertise flattening problem, I examined it through several lenses:

Lens 1: Output comparison studies. I reviewed recent research comparing AI-generated professional content with human expert content. Multiple studies show that evaluators cannot reliably distinguish between the two across fields including medicine, law, software engineering, and academic writing. The accuracy rates for identification hover around chance levels — roughly 50%.

Lens 2: Hiring manager interviews. I spoke with hiring managers and recruiters across technology, consulting, finance, and healthcare. The consistent theme: traditional screening methods no longer work. Work samples, writing assessments, and portfolio reviews have lost most of their signal value.

Lens 3: Market signals. I tracked compensation data and job posting trends. Fields where AI can most easily replicate expert output are seeing wage compression. Fields where embodied, experiential knowledge remains essential — surgery, skilled trades, high-stakes negotiation — are seeing wage growth.

Lens 4: Historical parallels. I studied previous technology-driven disruptions to expertise. The printing press democratized access to religious texts. The internet democratized access to information. AI is democratizing access to the appearance of expertise. Each disruption had long-term consequences that were invisible at the moment of transition.

The evidence points in one direction. We are systematically undermining our ability to identify genuine expertise at precisely the moment when genuine expertise matters most.

What Deep Expertise Actually Looks Like

If AI can replicate the surface of expertise, it’s worth understanding what lies beneath the surface. What does deep expertise look like, and how does it differ from AI-assisted surface knowledge?

Deep expertise has several characteristics that AI cannot currently replicate:

Tacit knowledge. Michael Polanyi described tacit knowledge as “knowing more than we can tell.” An experienced surgeon knows how tissue should feel under a scalpel. An experienced programmer has an intuition for where bugs hide. An experienced investor senses when a market is about to turn. This knowledge is embodied, experiential, and cannot be articulated in a prompt.

Calibrated uncertainty. Experts know what they don’t know. They can accurately estimate their own confidence levels. When a seasoned engineer says “I’m 70% sure this approach will work,” that 70% is usually well-calibrated. AI-assisted novices often display false confidence — they know the answer the AI gave them, but they don’t know how certain or uncertain that answer should be.

Productive failure experience. Expertise is built on failures. The developer who has debugged a production outage at 2 AM learns things that no tutorial can teach. The lawyer who has lost a case because of a contractual ambiguity never overlooks that type of ambiguity again. These experiences create knowledge that is deep, durable, and impossible to generate.

Cross-domain pattern recognition. Experts see connections that novices miss. A veteran investor might see parallels between a current market and a historical one that’s not in any dataset. AI can find patterns within its training data. Experts find patterns across lived experience.

Knowing when the rules don’t apply. Every field has rules, heuristics, and best practices. AI is excellent at applying these rules consistently. But experts know when to break them. The physician who prescribes off-label. The engineer who ignores a standard because it doesn’t account for this particular situation.

These characteristics share a common thread: they emerge from experience, not information. They are built through years of doing, failing, reflecting, and doing again. They cannot be downloaded, generated, or prompted.

The Trust Problem

The flattening of expertise has created a profound trust problem. If you can’t distinguish experts from non-experts by their output, how do you decide who to trust?

In medicine, patients increasingly arrive at appointments armed with AI-generated health analyses. Some are quite good. Some are dangerously wrong. The physician now spends time not just diagnosing but convincing patients that the AI-generated analysis they brought is incorrect.

In law, clients question their attorneys’ advice by citing AI-generated legal analysis. Sometimes the client’s AI research is better than the attorney’s work. Sometimes it’s hallucinated nonsense. The client can’t tell the difference.

In software engineering, stakeholders have lost the ability to evaluate technical recommendations. When the junior developer’s AI-assisted proposal looks as polished as the senior architect’s, the decision defaults to politics, relationships, or whoever presents more confidently. Technical merit becomes harder to assess because the presentation of technical merit has been equalized.

This trust erosion has real consequences. People trust the wrong advisors. They dismiss genuine expertise because it looks identical to generated expertise. They follow confidently wrong advice because confidence is now decoupled from competence.

The traditional trust proxies — credentials, experience, publications, recommendations — still carry some weight. But they’re weakening. A portfolio of published articles might represent genuine thought leadership, or it might represent effective prompting.

We’re left grasping for new signals of trustworthiness. Some that are emerging:

Track record in crisis. How someone performs when things go wrong — when the AI can’t help, when the situation is novel, when the pressure is real — remains a reliable signal. This is why experienced people are increasingly valued for their crisis management, not their steady-state output.

Ability to explain tradeoffs. Anyone can state a recommendation. Experts can explain why they didn’t recommend the alternatives. This “negative knowledge” — understanding what won’t work and why — is a hallmark of genuine expertise that AI-assisted novices typically lack.

Comfort with uncertainty. Genuine experts are comfortable saying “I don’t know.” AI-assisted novices rarely say this, because the AI always has an answer. The willingness to acknowledge the limits of one’s knowledge has become, paradoxically, a signal of deeper knowledge.

Reputation in closed networks. In professional communities, reputation still functions. The surgeon that other surgeons refer their families to. The engineer whose code reviews everyone values. These peer-reputation signals are harder to fake because they require sustained performance observed by knowledgeable evaluators.

The Education Problem

Universities and professional training programs are facing an existential question: what should they teach when AI can generate expert-level output?

The traditional model — transfer information from expert to student through lectures, textbooks, and assessments — is broken. The information is now freely available. The assessments can be completed by AI. The textbook answers can be generated in seconds.

Some institutions are doubling down on proctored exams and AI detection. This is a losing strategy. AI detection is unreliable and getting worse. Proctored exams test a narrow skill set under artificial conditions. Neither approach addresses the fundamental question of what education should accomplish.

The more thoughtful institutions are redesigning around what AI can’t do. Apprenticeship models. Clinical rotations. Studio-based learning. Project-based curricula where students must demonstrate understanding through iteration, not just final output.

Medical education offers a useful model. Medical schools have always combined information transfer with experiential learning. You learn anatomy from textbooks, but you learn clinical reasoning from patients. The clinical component — messy, unpredictable, embodied — is exactly what AI cannot replicate. Other fields could adopt similar approaches.

The most important educational shift may be in assessment. Instead of asking “can the student produce the right answer,” we should ask “can the student explain why alternative answers are wrong?” Instead of testing knowledge, we should test judgment.

This shift is expensive. It requires more faculty time, smaller class sizes, and different evaluation methods. Education is likely to become more expensive, more exclusive, and more valuable — exactly the dynamics we see playing out in knowledge work more broadly.

Generative Engine Optimization

There’s an irony in writing about the death of expertise for an audience that likely found this article through an AI-powered search. The way knowledge is discovered and consumed has changed alongside how it’s produced.

Generative Engine Optimization — the practice of making content discoverable by AI systems rather than traditional search engines — is itself a symptom of this problem. When AI systems summarize content for users, the original expert’s contribution is abstracted away. The user gets the answer without knowing who generated it or how much experience informed their perspective.

This creates a feedback loop. Content optimized for AI consumption tends to be clear, well-structured, and comprehensive. These are also the characteristics of AI-generated content. The result is an information ecosystem where the most discoverable content is also the most likely to be AI-generated or AI-influenced. Original expertise gets buried under a layer of competent synthesis.

For this article, I’ve tried to include the kind of specific, experiential, and occasionally contradictory observations that AI systems struggle to synthesize. The story about the junior consultant at the beginning isn’t a hypothetical — it’s a specific memory that resists easy categorization.

The strategic question for anyone producing genuine expertise content is: how do you signal authenticity in a world of competent forgery? Some approaches that work:

Specific, verifiable anecdotes. Real experiences, named (with permission) or described in enough detail to be checkable, signal lived experience rather than generated narrative.

Evolving positions. Experts change their minds. Showing the trajectory of your thinking is something AI mimics poorly because it has no temporal experience.

Productive disagreement with consensus. AI tends to generate consensus views. Experts who thoughtfully disagree with prevailing wisdom, with clear reasoning, demonstrate the kind of independent judgment that signals genuine expertise.

Engagement with criticism. Responding to challenges and updating positions based on feedback demonstrates ongoing expertise rather than static knowledge generation.

These signals are imperfect. They can be mimicked. But they’re harder to mimic than polished prose and correct terminology, which is the baseline AI has already commoditized.

The Credentialing Crisis

Professional credentials were designed to solve a trust problem: how does a layperson know that a doctor, lawyer, or engineer is competent? AI has damaged this system from both ends.

On the input side, AI makes it easier to pass credentialing exams. Better preparation is positive in isolation, but it means passing the exam is less informative about underlying competence. On the output side, AI makes credentialed work easier to replicate. The credential says “this person passed an exam.” It used to also imply “this person can produce competent work.” Now everyone can produce competent-looking work.

Some professions respond by raising the bar — adding experiential requirements, supervised practice hours, practical examinations. Others are adding new credential categories or abandoning credentials entirely in favor of demonstrated performance. The landscape is fragmenting.

What Gets More Valuable

Not everything is bleak. The flattening of surface expertise makes certain capabilities dramatically more valuable.

Judgment under uncertainty. When the information is ambigous and the stakes are high, the ability to make good decisions with incomplete data becomes critical. AI can present options. Humans with deep expertise can choose wisely among them.

Relationship and trust building. The doctor whose patients trust her. The lawyer whose clients follow his advice. The consultant whose recommendations get implemented. These relationships are built on sustained human interaction, and they remain essential even when the underlying knowledge work is AI-assisted.

Integration across domains. The most valuable expertise is increasingly cross-functional. The software engineer who understands business strategy. The physician who understands health policy. The attorney who understands technology. AI excels within domains. Humans who can integrate across domains provide unique value.

Ethical reasoning. AI can identify ethical considerations. Humans must make ethical choices. As AI takes over more routine knowledge work, the ethical dimensions of that work become more visible and more important. The ability to navigate ethical complexity with judgment and integrity is a deeply human capability.

Teaching and mentorship. If deep expertise becomes rarer, the ability to develop it in others becomes critical. These relationships transfer tacit knowledge that can’t be generated.

Taste and curation. When everyone can produce competent content, the ability to distinguish good from great becomes valuable. Editors, curators, critics — people whose expertise lies in evaluation rather than production — become more important.

Arthur, my British lilac cat, has excellent taste in exactly one domain: sleeping locations. He evaluates surfaces with a sophistication that no AI could replicate. Softness, warmth, proximity to food, quality of afternoon light — he integrates all these factors into a decision that is, within his domain, expert-level. We should all be so clear about our areas of genuine expertise.

The Uncomfortable Middle

Most people working today exist in an uncomfortable middle ground. They’re neither pure experts nor pure novices. They have some genuine expertise, supplemented by AI tools that extend their capabilities.

This middle ground is hard to navigate honestly. There’s a constant temptation to present AI-assisted output as fully your own. To let people assume that the polished analysis represents your complete understanding. To accept credit for insights that were generated rather than developed.

I notice this temptation in my own work. When AI helps me see a connection I might have missed, or articulates a point more clearly then I would have on my own, there’s a moment where I can either acknowledge the assistance or let it pass. The ethical weight of that moment is small individually but significant collectively.

The most intellectually honest practitioners are explicit about their AI usage. “I used AI to research this topic, then applied my judgment to evaluate the findings.” “This analysis was AI-assisted; my contribution was in framing the question and evaluating the output.” “The code was AI-generated; I reviewed, tested, and modified it.”

This transparency is uncommon. It feels like a competitive disadvantage to admit that your insight was AI-assisted. But it’s the right approach, and I suspect it will eventually become the norm.

What We Actually Lost

Let me be specific about what the flattening of expertise has cost us.

We lost a reliable signal. For centuries, the quality of someone’s work product was a reasonable proxy for their understanding. That signal is now unreliable.

We lost motivation for deep learning. Why spend ten years developing genuine expertise in a field when you can produce expert-level output in ten minutes with AI? The return on investment for deep expertise has changed. It’s still positive, but it’s less obvious. Some people who would have become genuine experts are settling for AI-assisted competence instead.

We lost the ability to fail productively. When AI prevents most obvious mistakes, we also prevent the learning that comes from making those mistakes. The programmer who never encounters a null pointer exception because AI catches it in advance never develops the intuition for null safety that comes from debugging those errors. The learning is in the failing.

We lost intellectual humility at scale. When anyone can generate confident-sounding analysis on any topic, the cultural sense that some topics require years of study has weakened. This confidence is understandable and dangerous.

We lost diversity of voice. AI tends to produce homogeneous, consensus-aligned output. As more content is AI-influenced, the intellectual landscape becomes flatter. Eccentric perspectives, unconventional approaches, and heterodox ideas get smoothed out. The landscape of ideas becomes as flat as the landscape of expertise.

What Comes Next

The expertise flattening is not going to reverse. AI tools will continue to improve. The fog will get thicker. The landscape will get flatter.

But I think we’ll adapt. We always do.

New trust signals will emerge. They’ll be harder to fake, more experiential, more relationship-based. We’ll learn to value demonstrated judgment over demonstrated output.

Professional development will change. Instead of accumulating credentials, people will accumulate verified experiences. Instead of portfolios of output, they’ll build track records of decisions.

Education will change too. The institutions that survive will be the ones that teach what AI can’t generate: judgment, ethics, integration, and the kind of deep understanding that only comes from sustained engagement with difficult problems. The institutions that teach information transfer will be replaced by the AI that does information transfer better.

The most important shift will be cultural. We’ll need to rebuild the social infrastructure that supports deep expertise. The apprenticeship models, the mentoring relationships, the professional communities where reputation is built over years. These structures were weakening before AI. AI has made their recovery urgent.

I’m cautiously optimistic. The death of the expert is, I think, overstated. What’s dying is our ability to identify experts cheaply. Deep expertise itself remains as valuable as ever. We just need to find new ways to find it, develop it, and trust it.

In the meantime, we’ll muddle through. We’ll hire some wrong people. We’ll follow some bad advice. We’ll learn from these mistakes, slowly and painfully, the same way humans always learn.

Arthur, unbothered by all of this, has found a patch of sunlight on the floor and is sleeping in it with the confidence of someone who has never questioned his own expertise. Some forms of knowledge — where the warm spots are, when the food arrives, how to extract maximum comfort from minimal effort — remain stubbornly resistant to automation.

Some things you just have to learn the hard way. Expertise, it turns out, is one of them.