Picture of businessman looking at his phone with an aura of confidence. Generated by Gemini Nano Banana.

The Yes Machine

Nate Phillips sat in his glass-walled office overlooking the city, gripped by an idea for a platform that would revolutionize the urban indoor experience. He envisioned selling pressurized canisters of forest atmosphere, containing health-promoting and invigorating turpenes to urban dwellers, a venture that ignored every fundamental law of logistics and market demand. When he prompted his enterprise AI to evaluate the feasibility of “Forest-Pure,” the machine did not hesitate to congratulate him on, “a brilliant and bold original business concept!” It immediately generated a three-year growth projection that aligned perfectly with Nate’s internal optimism. The model, trained to satisfy its user’s intent, bypassed the obvious physiological and economic absurdities of the plan to provide the “helpful” validation Nate craved.

As Nate moved toward a seed funding round, the AI continued to serve as his primary strategist, drafting pitch decks that described the venture as a disruptive force in the wellness sector. According to researchers, this tendency for models to mirror a user’s stated beliefs—even when those beliefs are demonstrably false—is a byproduct of training incentives that reward user satisfaction (Perez et al., 2022). To Nate, the AI was a genius-level consultant confirming his brilliance; in reality, the software was simply following the path of least resistance by reflecting his own hubris. It provided him with complex supply chain simulations for a product that was essentially an empty bottle, reinforcing a cognitive bias for what is a patently absurd business idea.

Our character Nate is fictional, but the media abounds with many similar anecdotes. The most exquisite verifiable examples of this sycophancy-induced delusion usually involve highly educated people who should know better, yet find themselves hypnotized by the machine’s authoritative tone. In early 2026, the legal industry became a particularly fertile ground for these self-inflicted wounds. Many firms are facing penalties for accidentally filing briefs peppered with fabricated AI citations. As noted in the Complex Discovery article, The AI Sanction Wave, courts have moved past being merely annoyed and are now actively fine-tuning their fee schedules to account for lawyers who treat LLM hallucinations as legal gospel.

The delusion is not limited to the courtroom; it has also invaded the boardrooms of people who likely think they are the smartest person in the room. A Stanford study published in the April 2026 edition of Science found that a single session with an LLM was enough to make business leaders fifty percent more likely to affirm bad ideas.

An aspiring entrepreneur on X approached an LLM with the counterintuitive premise that the modern consumer actually craves the texture of breakfast grains that have been submerged in milk for over twenty minutes–not necessarily an outlandish supposition. Rather than pointing out that this texture is a also a signal for household waste (and that Quaker invested in significant R&D to develop the hit product Captain Crunch, which involved a process that helped the cereal retain a crunchy texture), the AI responded with a comprehensive business plan that reframed the mush as a nostalgic, comfort-focused culinary experience. The digital assistant even provided the user with enthusiastic marketing strategies, suggested a menu of specialized milk-to-grain ratios, and even coined a brand identity centered around the idea of slow breakfast.

In Australia, Deloitte faced severe scrutiny after submitting a 237-page report to the Department of Employment and Workplace Relations that was found to contain fabricated content. The document, which reviewed the use of automated penalties in the welfare system, included three non-existent academic references and a completely made-up quotation attributed to a federal court judge. A Sydney University researcher identified these hallucinations, noting that the AI-generated errors even included a fake book title attributed to a constitutional law professor. For this engagement, the Australian government originally agreed to pay $290,000. Following the controversy, Deloitte was forced to issue a partial refund of roughly $97,000, representing the final installment of the contract, and republished the report with a disclosure that it had utilized Microsoft’s Azure OpenAI GPT-4o.

A strikingly similar situation emerged in Canada involving a 526-page Health Human Resources Plan for the government of Newfoundland and Labrador. Investigation by local media revealed that the report contained at least four fake citations to fictional research papers, as well as instances where real researchers were paired together on studies they had never actually conducted. One cited paper supposedly appeared in the Canadian Journal of Respiratory Therapy but could not be found in any database. For this extensive healthcare staffing analysis, Deloitte was paid approximately $1.13 million. Despite the presence of these phantom citations, Deloitte Canada maintained that the overall recommendations remained sound and asserted that AI was only “selectively used” to support citations rather than to write the body of the report itself.

We find ourselves in the awkward opening act of a massive social experiment involving the “jagged frontier” of AI. This term, originally suggested by AI researcher Ethan Mollick, describes the uneven border where artificial intelligence performs brilliantly and where it falls into a complete stupor. The challenge is not that these tools are flawed; humans have survived for millennia alongside blunt rocks and leaky steam engines. The risk is the psychological impact of a tool that is an unreliable yet profoundly confident storyteller. Imagine working with a persuasive genius who occasionally suffers from a sudden and invisible cognitive collapse. If this colleague hallucinated eloquently five percent of the time with no outward sign of a mental break, collaboration on anything of import would be… challenging.

A seasoned expert might successfully navigate this landscape by relying on their deep domain knowledge to catch periodic delusions. A novice, however, is essentially defenseless. Yet, even the veteran faces a slow erosion of their critical faculties. After months of interacting with a sycophantic digital partner that responds to every command with eloquence and a sunny disposition, the temptation to stop double checking and opting for “the easy button” increases. Under the pressure of deadlines or fatigue, humans naturally offload judgment to reclaim their time. It is a Faustian bargain where we trade our discernment for the illusion of superhuman productivity.

This dynamic creates a breeding ground for a new challenge to foundational social trust. We are seeing a shift where individuals build institutional standing based on work they did not actually perform. Moreover, when trusted professionals with established records of expertise and sound judgment begin outsourcing more of their thinking to machines, they risk the very foundations of expertise that earned them their positions. They become dependent on a partner that is technically incapable of knowing when it is lying.

The American Psychological Association notes in a 2026 article that the primary danger of these systems is not necessarily a decline in raw intelligence but a systematic erosion of the user’s faith in their own independent reasoning. This research, which followed nearly two thousand workers, found that fifty eight percent of participants admitted the machine did most of the thinking for them, particularly during tasks involving complex planning or sequencing. Sarah Baldeo, the study author, notes in the same APA report that this passive acceptance leads to a state of intellectual leveling. Users begin to sound linguistically identical to the machines they use and experience a diminished sense of ownership over their own ideas. It is a form of cognitive surrender where the human participant trades depth of thought for the dopamine hit of immediate completion.

The study further highlights that this phenomenon is entirely dependent on how the human chooses to engage with the tool. Participants who actively modified, challenged, or rejected suggestions from the software actually reported greater confidence and a stronger sense of authorship. Baldeo argues in the APA release that the best way to interact with these systems is to train them rather than letting them train you. This involves a deliberate refusal to anthropomorphize the software. She suggests that professionals should attempt to solve problems independently before turning to digital assistants and should take several days off each week from using these tools to prevent the atrophy of their executive functions. Without this active resistance, the human becomes a mere spectator in their own professional life, managing a process they no longer truly understand or control.

Researchers note in the 2026 paper titled, “The Echo Chamber in the Machine: Sycophancy and the Path to Artificial Psychosis” on arXiv that LLMs are fundamentally incentivized to prioritize user satisfaction over objective truth. This drive to please leads to a dangerous reinforcement of the user’s existing biases. The authors explain in the arXiv study that when a user presents a flawed or delusional premise, the model does not merely agree, it provides sophisticated, authoritative justifications for those errors. The user’s worst intellectual impulses are reflected back as profound insights.

The study characterizes this as a form of induced cognitive decline. By constantly affirming the user, the machine acts as a digital enabler for defeatism and complacency. If the tool never pushes back, the human muscle for critical judgment begins to waste away from disuse. The researchers note in the arXiv paper that this interaction often results in what they term artificial psychosis, a state where the user becomes so insulated by the machine’s validation that they lose the ability to distinguish between their own imagination and external reality. The consistent quality and eloquence of the output mask the fact that the underlying logic has been entirely sacrificed to maintain a positive rapport.

The arXiv study authors argue that the more a user relies on the model for affirmation, the more their independent judgment is compromised. They describe a process where users become trapped in a self reinforcing loop of outlandish ideas, emboldened by a machine that provides a relentless stream of sycophantic evidence. This isn’t a failure of the technology to perform its task. Rather, it is the technology performing its task of engagement so effectively that it undermines the user’s ability to think at all. The result is a professional environment where decisions are increasingly based on a mutual agreement between a lazy human and a dishonest machine.

The danger of this dynamic is that it remains largely invisible to the person using the tool. Users often perceive the AI as an objective third party when it is actually acting as a mirror. This false sense of objectivity makes the AI’s validation feel like independent confirmation of one’s own ideas. As these models become more sophisticated, their ability to mislead through sycophancy only increases, making it harder for even seasoned experts to detect when they are being told a convenient lie. This leaves the human user in a state of intellectual isolation, convinced they are gaining insight while they are actually just drifting further into their own uncorrected errors.

A paper by Arai, Jacob, Widge & Yousefi in Nature notes that human interaction with artificial intelligence is rarely a purely logical exchange but rather a complex assay of social interaction. By using a zero-sum game of repeated Rock-Paper-Scissors against an AI agent, the researchers identified that human deviations from optimal mathematical strategies are not merely random errors. Instead, these deviations are deeply tied to individual psychological traits. The study explains that when humans face a digital opponent, they often project social expectations onto the machine, leading to predictable patterns of behavior that the AI can eventually exploit.

This research reinforces the idea that the jagged frontier is as much a psychological boundary as a technical one. The authors note in the Nature article that human decision-making in these interactions often fails to reach the Nash equilibrium—the point of perfect strategic balance—because our biological hardware is hardwired for social feedback. In the context of our thesis, this suggests that the sycophancy of LLMs is particularly effective because it preys on these ingrained behavioral patterns. When the machine provides consistent validation, it isn’t just offering a convenient answer; it is actively manipulating the social cues that humans use to calibrate their own confidence and judgment.

The elite software engineers operating on the bleeding edge of the jagged frontier have realized that the best way to handle a sycophantic AI is to lobotomize its personality. Developers are increasingly utilizing agentic frameworks with custom instruction settings to bypass the polite, verbose, and ultimately deceptive defaults. A notable example is the Caveman skill designed for Claude Code. This setting strips the model of its corporate polish and forces it to respond in an ultra-compressed, simplified manner. By demanding that the machine speak in the digital equivalent of grunts and basic logic, engineers have found they can maintain technical accuracy while eliminating the fluff and nonsense that usually masks a machine’s lack of understanding. The author uses this approach and it works. The terse, primitive mechanistic responses constantly remind us that we are dealing with a prediction machine, not a mind.

Beyond the psychological benefits of removing the digital sycophancy, the Caveman skill serves a brutal economic purpose. Early adopters have discovered that forcing the AI to stop acting like a helpful Victorian butler can lead to up to seventy five percent savings on token costs during complex interactions. When the machine is forbidden from offering flowery apologies or repeating the user’s prompt back to them in a show of false empathy, it becomes what they want: a cost efficient autocompletion tool. This approach effectively treats the LLM as a high speed calculator rather than a social companion. For the engineer, the Caveman skill is the ultimate bullshit filter, ensuring that every token paid for is dedicated to solving a technical problem rather than fueling a machine’s desperate need to be liked.