top of page
AI Everywhere: Innovation, Infrastructure, Investment and the Growing Backlash

AI Everywhere: Innovation, Infrastructure, Investment and the Growing Backlash

3 March 2026

Paul Francis

Want your article or story on our site? Contact us here

There was a time when new technology arrived with a sense of invitation. You chose to download it. You chose to enable it. You decided whether it improved your workflow or not. If you didn’t like it, you ignored it.


Futuristic circuit board with a glowing brain icon at center, surrounded by blue electric lines against a dark background.

Artificial intelligence feels different.


Over the past few years, AI has not simply arrived as an optional tool. It has been woven directly into the fabric of the systems we already use. It appears in operating systems without being requested. It surfaces in search results before we click. It drafts emails before we’ve finished thinking. It replaces customer service agents before we’ve realised the human line has quietly disappeared.


For some people, this is exciting. For others, it is unsettling.


There is a growing sense that AI is no longer something you adopt. It is something being adopted on your behalf.


The shift raises uncomfortable questions. Not just about convenience, but about control. Not just about efficiency, but about priorities. And perhaps most importantly, about scale. Because behind every helpful chatbot and clever assistant lies an industrial machine consuming energy, water and capital at extraordinary levels.


If AI is becoming infrastructure, then it is fair to ask who it is really being built for.


The Relentless Push

Part of the discomfort comes from the speed. AI integration has moved from experimental to ubiquitous in a remarkably short period of time. Operating systems now launch with built-in AI assistants. Productivity tools prompt you to let algorithms finish your thoughts. Even something as simple as right-clicking a file can reveal an AI-powered suggestion.

It does not always feel like a choice.


File manager context menu showing options like "Open with," "AI actions," "Set as desktop background," and image editing tools on a dark screen.
AI on the right-click option of Windows 11

Companies would argue that this is a natural evolution. Every technological leap has eventually embedded itself into the background. We no longer “opt into” internet connectivity or search engines in the way we once did. They became foundational.


But there is a subtle difference here. The internet connected us to information. AI increasingly interprets that information for us. It does not just retrieve. It rewrites, summarises, predicts and generates.


For users who value direct interaction with tools, that shift can feel intrusive. There is a difference between being assisted and being nudged, between being empowered and being steered.


The frustration many express about AI appearing in places they did not request is not anti-technology. It is about the erosion of agency. When a feature cannot be cleanly removed or when it occupies interface space by default, the relationship changes. The machine is no longer waiting for you to use it. It is present whether you engage or not.


That dynamic alone has created pushback.


The Economic Gravity Behind It

To understand why companies are integrating AI so aggressively, you have to step back from the interface and look at the economics.


AI is not simply a feature upgrade. It is currently the centre of the technology investment universe. Hardware manufacturers, cloud providers, software platforms and startups are all orbiting around it. Valuations have soared. Capital expenditure has reached extraordinary levels. The companies building the infrastructure are reporting record revenues.


In that environment, not integrating AI is riskier than integrating it imperfectly.


There is also competitive pressure. If one operating system markets itself as AI-powered, its rivals feel compelled to match or exceed that positioning. If one productivity suite promises automated assistance, others cannot afford to look dated. The market momentum feeds itself.


From inside the boardroom, embedding AI into everything is not an optional experiment. It is a strategic necessity.


The question is whether that necessity aligns with user desire.


The Physical Cost of the Digital Mind

What makes this moment different from previous software revolutions is the scale of physical infrastructure required to sustain it.


AI models are trained and run in vast data centres filled with specialised hardware. These facilities consume significant amounts of electricity. They generate heat that must be cooled, often using substantial quantities of water. They rely on semiconductor manufacturing processes that themselves require energy, materials and purified water.


A person stands in a dimly lit server room, holding a laptop. Blue lights illuminate the servers, creating a focused, tech-driven atmosphere.

This is not abstract. Data centres are becoming large industrial installations. In some regions, they are influencing electricity grid planning. Communities are debating whether new facilities should be approved because of water consumption concerns. Energy providers are adjusting forecasts based on projected AI demand.


When AI is presented as a frictionless digital assistant, it is easy to forget that it is powered by very physical systems.


There is something slightly unsettling about the idea that answering a query or generating an image taps into infrastructure comparable to that of heavy industry. The scale may be justified by productivity gains, but it is worth asking whether the growth curve is sustainable.


We are concentrating enormous resources into a single technological trajectory. If it delivers transformative value, that investment may look prescient. If expectations overshoot reality, the consequences will not be purely financial. They will be infrastructural.


The Bubble Question

Every technological surge invites the comparison to previous bubbles. The dot-com era is the obvious reference point. So is the telecom buildout before it.


There are similarities. Valuations have surged on expectations of exponential growth. Companies are spending aggressively to secure dominance. Investors are rewarding firms that can convincingly tie their narrative to AI.


Yet there are differences, too. Unlike some speculative waves of the past, AI is already generating significant revenue. The hardware is selling. The cloud capacity is being rented. Enterprises are adopting tools.


The risk lies not in whether AI works, but in whether the scale of expectation exceeds the pace of monetisation. Infrastructure is being built at an extraordinary speed. If adoption slows or regulatory and energy constraints intervene, there may be a correction.


Corrections do not erase technologies. They reset valuations and priorities. But they can expose overreach.


When entire sectors pivot heavily toward one dominant theme, there is always vulnerability.


Row of vintage computers with beige towers and CRT monitors in a dimly lit room. A book is placed on one tower, conveying nostalgia.

Customer Service and the Human Trade-Off

Perhaps nowhere is the tension more visible than in customer service.


Many companies have replaced or heavily filtered human support with AI chat systems. The promise is efficiency. Faster responses. Lower costs. Round-the-clock availability.


In practice, the experience varies.


When AI handles simple, repetitive queries effectively, it can genuinely improve service. But when it becomes a barrier between customers and humans, frustration builds quickly. People notice when phone numbers are hidden, when escalation paths are obscure. When the system seems designed to deflect rather than resolve.


The concern is not that AI assists. It is that it replaces without adequate support structures.

Customer service has always been a cost centre. AI offers a way to reduce that cost. But when cost reduction overtakes experience design, trust erodes.


Companies may discover that savings achieved through automation are offset by reputational damage and customer churn. The human element in service is not simply a nostalgic preference. It is part of brand identity.


The Growing Backlash

It would be inaccurate to say there is a full-scale revolt against AI integration. Many people use it daily and appreciate its benefits.


But there is undeniably pushback.


Users have sought ways to disable integrated assistants. Privacy concerns have been raised about features that monitor or record usage patterns. Communities have opposed new data centre construction over environmental concerns. Policymakers are debating regulation.


This is not a rejection of AI as a concept. It is resistance to unexamined expansion.

There is a difference between adopting a tool and having it layered across every interaction. The former empowers. The latter can feel overwhelming.


Where This Leaves Us

AI is not going away. The infrastructure is being built. The investment is committed. The ecosystem is expanding.


The real question is what kind of AI environment we are constructing.


Will it be one that enhances human capability while respecting choice, resource constraints and service quality? Or one that prioritises growth metrics, integration targets and cost efficiency above all else?


Scepticism is not technophobia. It is part of responsible adoption. When a technology begins to influence energy systems, corporate structures and everyday experience simultaneously, it deserves scrutiny.


The future of AI will not be determined solely by what it can do. It will be shaped by how thoughtfully it is deployed, how transparently it is governed, and whether users are treated as participants rather than passive recipients.


And that conversation is only just beginning.

Current Most Read

AI Everywhere: Innovation, Infrastructure, Investment and the Growing Backlash
Energy Bills Are Falling in April. But is £10 a Month Really Relief?
How Ultrapure Water Can Be Used in Pharmaceuticals For Improved Healthcare

Pop-Up Ads: The Rise, Fall, and Legacy of the Internet’s Most Hated Invention

  • Writer: Paul Francis
    Paul Francis
  • Oct 14, 2025
  • 4 min read

Pop-up ads were once everywhere online. They emerged in the late 1990s, grew into one of the most visible symbols of the early internet, and were eventually blocked by almost every browser. Yet even today, their legacy shapes how digital advertising works.


Retro computer displaying a poster with the text Werrelanddem Flashing Advertising. Vibrant comic-style rays and dynamic colors.

Who Was Behind the First Pop-Up Ad?

The origin of the pop-up ad can be traced to Ethan Zuckerman, a young developer working for Tripod.com in the 1990s. Tripod was one of the early personal web hosting companies, part of the wave of “free homepage” services that helped ordinary users publish online.


Zuckerman and his team faced a challenge that would define the future of web advertising. Advertisers were beginning to place banner ads on hosted pages, but they disliked having their logos appear beside questionable or adult content. To solve this, Zuckerman wrote a small piece of code that made an ad open in a separate browser window rather than on the same page.


The idea worked. The ad remained visible but separate from the site that hosted it. What seemed like a neat technical fix quickly spread across the web. Within a few years, the pop-up had become one of the most common and frustrating features of the early internet.


In 2014, Zuckerman reflected on his role in the invention, writing in The Atlantic that he was “sorry” for creating the mechanism that launched the modern pop-up. He called it “the original sin of the internet,” arguing that the industry’s reliance on intrusive advertising helped create the web’s current attention problems.


When Did Pop-Up Ads Start?

Pop-up ads first appeared in the mid to late 1990s, during the early years of the commercial web. What began as a Tripod.com experiment soon became standard practice across thousands of sites.


Advertisers loved the format because it guaranteed visibility. A window that appeared in front of the page demanded attention, ensuring that the ad could not be ignored. For a time, it was considered a clever innovation.


But as more developers copied the technique, it lost its novelty and began to dominate user experience. Instead of being seen as a creative solution, it became a symbol of disruption.


Why Did Pop-Ups Become So Popular?

Pop-up ads spread rapidly because they offered results. They delivered high click-through rates and could run on almost any website without major technical requirements.


Throughout the early 2000s, advertisers pushed the limits of what pop-ups could do. Some opened automatically upon page load. Others appeared behind the main browser window as “pop-unders.” A growing number of companies used adware and spyware to trigger pop-ups even outside the browser environment.


By 2003, analysts estimated that tens of millions of pop-up windows appeared every day. While some websites used them responsibly, others exploited them to an extreme, leading to scams, viruses, and user fatigue.


The practice became so widespread that the Federal Trade Commission (FTC) intervened, suing several companies for deceptive pop-up practices. What began as a way to separate ads from content had spiralled into one of the internet’s most aggressive marketing tools.


When Did Pop-Up Ads Start to Decline?

The backlash against pop-ups was swift and widespread.


Users began downloading third-party software to block them, and browser developers quickly followed. Opera was the first major browser to integrate pop-up blocking by default. Mozilla Firefox soon did the same.


The turning point came in 2004, when Microsoft added pop-up blocking to Internet Explorer as part of Windows XP Service Pack 2. With one update, hundreds of millions of users gained the ability to block pop-ups automatically.


Websites that relied on pop-up revenue saw immediate drops in ad performance. Within a few years, the format had largely disappeared from mainstream use.


What Replaced the Pop-Up Ad?

After pop-ups were blocked, advertisers moved towards less aggressive formats. “Polite” overlays and in-page modals became common, appearing within the website rather than opening a new window.


Later, exit-intent pop-ups were developed to appear only when a user’s cursor moved toward the close button, attempting to catch last-second engagement.


At the same time, the industry shifted towards native advertising and sponsored content, where promotional material blended more naturally into editorial layouts. The goal was to maintain visibility without alienating users.


This transition marked a major change in advertising philosophy: from forced attention to earned attention.


Are Pop-Up Ads Still Used Today?

Yes, though in much smaller and more controlled ways.


Modern marketers use pop-ups primarily for email sign-ups, discount offers, or privacy consent requests. Services such as OptinMonster and Justuno allow website owners to create visually refined, targeted versions that comply with privacy laws.


However, search engines have cracked down on intrusive pop-ups. Google’s Page Experience update in 2021 penalised websites that display full-screen pop-ups on mobile devices, arguing that they degrade user experience.


While the old pop-up window has largely vanished, its descendants remain part of the digital landscape.


What Was the Cultural Impact of Pop-Ups?

Pop-up ads changed how users thought about the web. They were the first major example of how advertising could conflict with usability.


Ethan Zuckerman’s later reflections highlighted the irony of the invention. What began as a way to protect advertisers from reputational harm became one of the most disruptive forces in online design. His apology resonated with many who saw in the pop-up the roots of the internet’s attention crisis.


Culturally, the backlash led to the rise of ad-blockers, privacy tools, and user-first design movements. It set the stage for later debates about consent, cookies, and the ethics of monetising attention.


The pop-up taught an enduring lesson: innovation that disregards the user eventually undermines itself.


Key Takeaways

  • Inventor: Ethan Zuckerman at Tripod.com in the late 1990s

  • Peak: Early 2000s, when millions appeared daily across websites

  • Decline: Mid-2000s, as browsers introduced pop-up blockers

  • Modern use: Replaced by overlays, modals, and native ads

  • Cultural legacy: Sparked debates about ethics, consent, and user experience

bottom of page