top of page
Why You Should Not Trust Your Car’s Automatic Systems Completely

Why You Should Not Trust Your Car’s Automatic Systems Completely

12 February 2026

Paul Francis

Want your article or story on our site? Contact us here

Most modern drivers assume that if a feature is labelled “automatic”, it will take care of itself. Automatic lights. Automatic braking. Automatic lane correction. The car feels intelligent, almost watchful.


Car dashboard at night with blurred city lights in the background. Speedometer glows blue. Display shows 8:39. Moody, urban setting.

But there is a quiet issue that many drivers are unaware of, and it begins with something as simple as headlights.


The automatic headlight problem

In fog, heavy rain or dull grey daylight, many cars will show illuminated front lights but leave the rear of the vehicle dark. From inside the car, everything appears normal. The dashboard is lit. The automatic light symbol is active. You can see light reflecting ahead.


However, what often happens is that the vehicle is running on daytime running lights rather than full dipped headlights. On many cars, daytime running lights only operate at the front. The rear lights remain off unless the dipped headlights are manually switched on.

The system relies on a light sensor that measures brightness, not visibility. Fog does not always make the environment dark enough to trigger full headlights. Heavy motorway spray can reduce visibility dramatically while still registering as daylight. The result is a vehicle that is difficult to see from behind, especially at speed.


Under the Highway Code, drivers must use headlights when visibility is seriously reduced. Automatic systems do not override that responsibility. In poor weather, manual control is often the safer choice. It is a small action that can make a significant difference.


Automatic emergency braking is not foolproof

Automatic Emergency Braking, often referred to as AEB, is one of the most widely praised safety technologies in modern vehicles. It is designed to detect obstacles and apply the brakes if a collision appears imminent.


In controlled testing, it reduces certain types of crashes. But it is not infallible. Cameras and radar can struggle in heavy rain, low sun glare, fog, or when sensors are obstructed by dirt or ice. Some systems have difficulty detecting stationary vehicles at high speed. Others may not recognise pedestrians at certain angles.


It is a safety net, not a guarantee.


Lane assist is not autopilot

Lane keeping systems gently steer the car back into its lane if it detects a drift. On clear motorways with bright road markings, they can work well.


On rural roads, in roadworks, or where markings are faded, they can disengage or behave unpredictably. Drivers may not even realise when the system has switched off. Over time, there is a risk that drivers become less attentive, assuming the vehicle will correct mistakes.

It will not.


Cars drive on a wet highway during sunset. The sky is golden, and trees line the road. The scene is viewed through a windshield.

Adaptive cruise control still requires full attention

Adaptive cruise control maintains speed and distance from the car ahead. It is comfortable on long motorway journeys.


However, it does not anticipate hazards like a human driver. It can brake sharply when another vehicle exits your lane. It may not react appropriately to a fast vehicle cutting in. Most importantly, it does not read the wider context of traffic conditions.


It reduces workload, but it does not remove responsibility.


Blind spot monitoring is not perfect

Blind spot indicators are helpful, especially in heavy traffic. They provide an extra warning when another vehicle is alongside you.


But motorcycles, fast approaching cars, or vehicles at unusual angles can sometimes escape detection. Sensors can also be affected by weather or dirt. A physical shoulder check remains essential.


Cameras distort reality

Reversing cameras and parking sensors have reduced low-speed bumps and scrapes. They are undeniably useful.


Yet cameras distort depth perception, and small or low obstacles can be difficult to judge accurately. Relying entirely on the screen rather than physically checking surroundings is one of the most common causes of minor accidents.


The bigger risk is complacency

There is a growing concern among safety researchers about automation complacency. When systems work well most of the time, drivers begin to relax. Attention drifts. Reaction times lengthen.


Modern vehicles are safer than ever, but the technology is designed to support an attentive driver. It is not designed to replace one.


The word “assist” appears frequently in the naming of these systems for a reason. They assist. They do not assume control.


Automatic lights, braking, steering correction and cruise systems are impressive pieces of engineering. They reduce risk. They improve comfort. But they still require a human driver who understands their limits.


Trusting technology is reasonable. Trusting it completely is not.

Current Most Read

When AI Crosses the Line: Why the Grok Controversy Has Triggered a Regulatory Reckoning
A World Cup Under Pressure: How American Politics Could Shape FIFA 2026
Why the ‘Driverless’ Narrative Is Failing the Freight Industry

When Social Media Stops Feeling Real: How AI Slop Is Reshaping Online Life

  • Writer: Paul Francis
    Paul Francis
  • Feb 3
  • 4 min read

Scroll through almost any major social media platform today and something feels different. Feeds that once mixed personal updates, news, and carefully made creative work are increasingly filled with strange images, repetitive videos, and emotionally charged scenes that feel artificial, exaggerated, or simply nonsensical.


A serious figure in a white shirt holds a sign saying "Like, share, support" against a purple, abstract background with black splashes.

This growing wave of low-effort, AI-generated content has become known online as “AI slop”. It is not a technical term, but it captures a shared frustration. Content that is cheap to produce, designed for fast emotional reaction, and optimised for engagement rather than meaning.


What began as novelty has quietly turned into saturation, and many users are beginning to push back.


What people mean when they say “AI slop”


A person in white clothing walks on stormy water, with a boat and distressed people behind him. Dark, cloudy sky and lightning set a dramatic mood.

AI slop usually refers to images and videos generated quickly using artificial intelligence tools, often with little care for realism, coherence, or ethics. Common examples include fake images of children in distress, miraculous acts staged for sympathy, animals in improbable danger, or surreal religious and military scenes designed to provoke emotion.


The aim is not accuracy or storytelling. The aim is reaction. Likes, shares, comments, and watch time.


Because modern algorithms reward engagement above all else, this type of content spreads easily. It requires no filming, no editing skills, and no real-world accountability. A single creator can generate dozens of posts a day, testing which ones trigger the strongest response.


Why platforms quietly benefit from the flood

Major platforms have not resisted this trend. In many cases, they have encouraged it.

Companies like Meta and Google have openly described artificial intelligence as the next phase of social media. Built-in image generators, video tools, and AI filters are now standard features, making content creation faster and more accessible than ever.


From a business perspective, AI slop is efficient. It keeps users scrolling, costs very little to host, and scales infinitely. Whether the content is meaningful is largely irrelevant to the system that distributes it.


Research into platform feeds suggests that a noticeable proportion of content shown to new users is already low-quality AI-generated media, particularly in short-form video formats where speed matters more than depth.


The growing sense of backlash

While AI slop performs well numerically, sentiment around it is shifting.

Under many viral posts, the most visible comments are no longer admiration but irritation. Users point out obvious flaws, complain about deception, or express exhaustion at constantly having to question what is real.


In some cases, comments criticising the content receive more engagement than the content itself. This creates a strange feedback loop where outrage still fuels visibility, further embedding the very material people want less of.


A small but notable part of this backlash has taken shape through online accounts dedicated to highlighting absurd or manipulative AI-generated posts. One such account, run by a young student in France, catalogues extreme examples of AI slop circulating on platforms like Facebook. The account has drawn attention to how easily such content gains traction without scrutiny. You can find it here: https://x.com/FacebookAIslop


The existence of accounts like this reflects a wider mood rather than a single campaign. A sense that something about the online environment is slipping out of balance.


The mental toll of constant artificiality

Researchers studying online behaviour warn that the impact of AI slop is not just annoyance.



Constant exposure to content that is fake, exaggerated, or meaningless can reduce attention span and discourage critical thinking. Verifying authenticity requires effort. Over time, many users simply stop checking.


This has led some academics to describe a “brain rot” effect. Not because individual videos are harmful, but because the overall environment trains people to consume quickly, react emotionally, and move on without reflection.


Even content that is obviously fake can contribute to this erosion by normalising a feed where nothing needs to make sense to succeed.


When slop turns into something more serious

Beyond irritation, AI-generated content can carry real risks.


A man in a blue shirt performs CPR on another man in red, outdoors. A first aid kit is nearby. Comic style, urgency depicted with "Puff Puff!"

Recent controversies involving AI tools being used to digitally alter images of real people, including women and children, show how quickly low-quality content can cross into abuse. In other cases, fake videos and images have been used to shape political narratives, creating the illusion of public support or emotional response that may not exist.


This is especially concerning as many people now rely on social media as their primary source of news and information.


At the same time, several major platforms have reduced human moderation, relying more heavily on automated systems and user reporting. This makes it harder to respond quickly or consistently to emerging harms.


Where this leaves us

AI-generated content is not going away. The tools are improving, the costs are falling, and platforms remain financially aligned with volume over quality.


The question is not whether AI will be part of online culture, but whether digital spaces can retain any sense of trust, creativity, or shared reality if everything becomes synthetic, disposable, and engagement-driven.


For many users, the frustration is not about technology itself, but about what it is being used for. The fear is not of AI creativity, but of an internet increasingly filled with noise, manipulation, and content designed to exploit attention rather than inform or inspire.


If there is a shift coming, it will likely come not from platforms, but from users deciding what they are willing to tolerate in their feeds, and what they quietly stop engaging with.


*All images generated on Leonardo AI

bottom of page