{"id":8006,"date":"2026-02-25T14:00:00","date_gmt":"2026-02-25T14:00:00","guid":{"rendered":"https:\/\/stuartglover.com\/?p=8002"},"modified":"2026-02-25T19:22:29","modified_gmt":"2026-02-25T19:22:29","slug":"the-quiet-ones-what-ai-still-cant-do","status":"publish","type":"post","link":"http:\/\/iamglover.com\/?p=8006","title":{"rendered":"The Quiet Ones: What AI Still Can\u2019t Do, and Why That\u2019s Actually Good News"},"content":{"rendered":"<p style=\"text-transform:uppercase; font-size:0.75em; font-weight:bold; color:#27AE60; letter-spacing:0.1em;\">Opinion \/ Artificial Intelligence<\/p>\n<p style=\"font-size:1.15em; font-style:italic; color:#555555;\"><strong>Everyone&#8217;s obsessed with what AI can do. I&#8217;m more interested in what it can&#8217;t.<\/strong><\/p>\n<hr \/>\n<p><strong style=\"font-size:1.1em;\">There&#8217;s a pattern I&#8217;ve noticed in how people talk about AI, and it&#8217;s starting to bother me.<\/strong><\/p>\n<p>Every week, a new benchmark gets shattered. Every month, a capability that was &#8220;years away&#8221; arrives early. The tone of AI coverage has settled into a kind of breathless incrementalism &#8212; always zooming in on what just became possible, never pausing to look at the shape of what remains impossible.<\/p>\n<p>I think that&#8217;s a mistake. Not just intellectually, but practically. If you want to understand where AI is actually going &#8212; and where your own edge might lie &#8212; the most useful question isn&#8217;t &#8220;what can it do now?&#8221; It&#8217;s &#8220;what does it still genuinely struggle with, and why?&#8221;<\/p>\n<p>So here are five things that, as of early 2026, AI still can&#8217;t do well. I&#8217;m not saying &#8220;never.&#8221; I&#8217;m saying &#8220;not yet, and the reasons are more interesting than you might think.&#8221;<\/p>\n<h2>1. Care About the Outcome<\/h2>\n<p>AI systems are extraordinarily good at optimising for a target. What they can&#8217;t do is genuinely care whether the target was the right one. This sounds abstract, but it plays out in very concrete ways.<\/p>\n<p>A lawyer who cares about their client doesn&#8217;t just answer the question asked &#8212; they notice the question behind the question. A doctor who cares about their patient doesn&#8217;t just process symptoms &#8212; they notice the patient&#8217;s face. AI, for all its power, is still fundamentally reactive. It responds to what you give it. It doesn&#8217;t have skin in the game.<\/p>\n<p>This is not a small gap. Most of the truly valuable work humans do &#8212; parenting, leadership, mentorship, great medicine, great teaching &#8212; depends not just on capability, but on genuine investment in the outcome. That&#8217;s still ours.<\/p>\n<h2>2. Navigate a Room<\/h2>\n<p>Embodied, real-time social intelligence &#8212; reading a room, sensing when someone has checked out of a meeting, knowing that now is not the moment to push &#8212; remains stubbornly difficult for AI systems. Language models are trained on text. Text is what people decided to write down. It is, almost by definition, not the full picture.<\/p>\n<blockquote style=\"border-left: 4px solid #27AE60; padding-left: 1.2em; font-style: italic; font-size: 1.1em; color: #27AE60; margin: 1.5em 2em;\"><p>\n&#8220;Text is what people decided to write down. It is, almost by definition, not the full picture.&#8221;\n<\/p><\/blockquote>\n<p>The unspoken negotiation that happens in any human interaction &#8212; the pauses, the micro-expressions, the sense that something is off &#8212; is still largely invisible to AI. Multimodal models are making inroads, but real-time social navigation at a human level remains a frontier, not a solved problem.<\/p>\n<h2>3. Be Consistently Wrong in a Useful Way<\/h2>\n<p>This one sounds strange, so let me explain. When a human expert is wrong, their errors are often diagnostic. They reveal assumptions, blind spots, theoretical commitments. You can learn something from the shape of a human&#8217;s mistake.<\/p>\n<p>AI errors are different. They&#8217;re often confident, plausible, and structurally unrelated to the truth. A hallucination doesn&#8217;t reveal a belief &#8212; it reveals a statistical pattern gone sideways. There&#8217;s no &#8220;why&#8221; to interrogate in the way there is with human error.<\/p>\n<p>For fields where learning from failure is central &#8212; scientific research, strategy, engineering &#8212; this matters more than it might seem. Human mistakes are data. AI hallucinations are mostly just noise.<\/p>\n<h2>4. Have a Reputation to Protect<\/h2>\n<p>Trust, in human society, is built and maintained through accountability. A professional who gives bad advice suffers consequences &#8212; reputation damage, lost clients, in some cases legal liability. Those consequences shape behaviour, and that shaping is part of what makes advice trustworthy.<\/p>\n<p>AI has no reputation to protect. It cannot be embarrassed. It won&#8217;t lose sleep over a bad call. This is not just a philosophical observation &#8212; it has structural implications for how much we should trust AI outputs in high-stakes domains, and how we should design the systems around them.<\/p>\n<p>The humans who sit between AI and its outputs &#8212; the ones who sign their name at the bottom, who take the call, who stand in the room &#8212; are not redundant. They are the accountability layer. For now, that layer matters enormously.<\/p>\n<h2>5. Want Something New<\/h2>\n<p>AI can recombine. It can synthesise. It can extrapolate patterns in directions you might not have thought to look. What it cannot do &#8212; at least not yet, and perhaps not in any meaningful sense &#8212; is want something that isn&#8217;t implicit in its training.<\/p>\n<p>The truly disruptive ideas in human history didn&#8217;t come from pattern-matching against the existing data. They came from people who were obsessed with a problem nobody else thought was worth solving, or who saw a connection that the prevailing paradigm had made invisible. That kind of directed, motivated originality &#8212; the kind that bends the curve of history &#8212; still has a distinctly human signature.<\/p>\n<h2>Why This Is Actually Good News<\/h2>\n<p>I want to be clear: I&#8217;m not writing this as a comfort blanket. I&#8217;m not trying to reassure you that humans are fine and nothing will change. A lot will change. A lot already has.<\/p>\n<p>But there&#8217;s something genuinely useful in mapping the gaps. Because if you know what AI can&#8217;t do &#8212; really can&#8217;t, structurally, not just &#8220;not yet&#8221; &#8212; then you know where to invest. You know which skills are appreciating in value rather than depreciating. You know where to put yourself.<\/p>\n<p>Care. Social intelligence. Accountability. Reputation. Genuine curiosity. These aren&#8217;t soft skills. In an AI-saturated economy, they are increasingly the hard ones.<\/p>\n<hr \/>\n<p style=\"text-align:center; font-size:1.15em; font-style:italic; font-weight:bold; color:#2C3E50;\">The race isn&#8217;t to outrun the machine. It&#8217;s to become more deeply, irreducibly human.<\/p>\n<hr \/>\n<p style=\"font-size:0.8em; color:#888888; font-style:italic;\">Tags: Artificial Intelligence &bull; Opinion &bull; Future of Work &bull; Technology &amp; Society<\/p>\n","protected":false},"excerpt":{"rendered":"<p>Everyone&#8217;s obsessed with what AI can do. I&#8217;m more interested in what it can&#8217;t. Five things AI still genuinely struggles with in 2026 &#8212; and why understanding the gaps is the smartest move you can make.<\/p>\n","protected":false},"author":2,"featured_media":0,"comment_status":"open","ping_status":"open","sticky":false,"template":"","format":"standard","meta":{"footnotes":""},"categories":[3,23],"tags":[],"class_list":["post-8006","post","type-post","status-publish","format-standard","hentry","category-artificial-intelligence","category-technology"],"jetpack_featured_media_url":"","_links":{"self":[{"href":"http:\/\/iamglover.com\/index.php?rest_route=\/wp\/v2\/posts\/8006","targetHints":{"allow":["GET"]}}],"collection":[{"href":"http:\/\/iamglover.com\/index.php?rest_route=\/wp\/v2\/posts"}],"about":[{"href":"http:\/\/iamglover.com\/index.php?rest_route=\/wp\/v2\/types\/post"}],"author":[{"embeddable":true,"href":"http:\/\/iamglover.com\/index.php?rest_route=\/wp\/v2\/users\/2"}],"replies":[{"embeddable":true,"href":"http:\/\/iamglover.com\/index.php?rest_route=%2Fwp%2Fv2%2Fcomments&post=8006"}],"version-history":[{"count":1,"href":"http:\/\/iamglover.com\/index.php?rest_route=\/wp\/v2\/posts\/8006\/revisions"}],"predecessor-version":[{"id":8007,"href":"http:\/\/iamglover.com\/index.php?rest_route=\/wp\/v2\/posts\/8006\/revisions\/8007"}],"wp:attachment":[{"href":"http:\/\/iamglover.com\/index.php?rest_route=%2Fwp%2Fv2%2Fmedia&parent=8006"}],"wp:term":[{"taxonomy":"category","embeddable":true,"href":"http:\/\/iamglover.com\/index.php?rest_route=%2Fwp%2Fv2%2Fcategories&post=8006"},{"taxonomy":"post_tag","embeddable":true,"href":"http:\/\/iamglover.com\/index.php?rest_route=%2Fwp%2Fv2%2Ftags&post=8006"}],"curies":[{"name":"wp","href":"https:\/\/api.w.org\/{rel}","templated":true}]}}