A pile of endless side projects

A tiny (and dusty) blog on stuff I do

AI everywhere, hype, laziness and human stupidity

2026-04-15

Introduction and preamble

This post was quite difficult to structure and write, it was refined over several iterations, starting from a much shorter draft. A sneak peek into an important point that I will discuss later - and a provocation - is the idea that someone could say “it would have been so much quicker to write all this stuff using an AI tool!”. This line of thought, and the temptation offered by it, proves immediately the point that we, as humans, are really efficient at being lazy, by design. Doing stuff needs energy. And time.

(This post is fully organic and human-made. If you know me you already knew, if you didn’t, happy reading!)

Another thing worth mentioning before starting is that I had to select the angle to use, the perspective I offer. I decided that this will not be a technical article. I don’t want to look at AI from this point of view. There will be a section with some examples related with software engineering, but I promise it’s just a few of them and they won’t be about LLMs1 inner workings.

As a teaser, to hook you already with the main point I will make in a while:

people using AI tools daily are getting more and more stupid

(If your attention span has been destroyed by the modern world, you can skip there, but I would love for you to read everything!)

AI everywhere, annoyingly

Since the LLM/AI boom of the past couple of years (~2023), I am getting more and more frustrated, disappointed, worried and tired with AI being everywhere. Lots of people are using this technology to do extremely trivial things, partly because tech companies are putting it everywhere, creating artificial hype about it being “the future” and the “Brand New Amazing Thing” that everyone should adore and use, from writing a grocery list to planning a vacation, from writing mails to searching online. For everything.

I’m concerned with this on so many different levels, and apparently I need to vent out a bit2, writing this post about these interesting times.

I am sure that I am not the only one with similar thoughts, but I have the feeling that the amount of excitement (either genuine or artificially generated) on the topic far overwhelms the less enthusiastic and more mindful opinions.

In the next section I will quickly mention the “hype cycle” stirred up by all companies, linked to marketing phenomena. This is tangential to my main point, but it’s worth mentioning that - for several reasons, some of them maybe a bit scary - AI is being pushed literally everywhere. I don’t need to tell you this, if you have eyes you should have noticed.

Corporations, FOMO and the “powered by AI” epidemic

An insane amount of corporations are trying to stuff AI everywhere. If you pick a random company and open its website, you will see that it’s either “powered by AI”, an “AI-first company” or another permutation like that. A quite simple (and boring) reason is FOMO, a term that means literally “Fear Of Missing Out”, in this case being scared of falling behind other competitors, not showing how innovative one is3.

🗣 «If X, Y and Z are putting AI in their product, we must also do it, or else nobody will come to us! We will look so ancient!»
đŸ‘€ «But Franco, we are a pizzeria
 đŸ•Â»
🗣 «Who cares! Let’s generate the menu using AI! Let’s put a chatbot on our website to help customers choose the pizza! Amazing idea! So cool!»

I think this (very simplified) example is clear. One of the main reasons why AI is everywhere is simply
 marketing. This answers most of these questions:

But tech corporations aren’t the only ones going crazy. I have the feeling that the amount of people excited about all of this (including vibe coders and AI bros) outweighs the critics. A quick glance at LinkedIn (that to be fair, always has been a dumpster of people congratulating each other about the nothingness) will reveal everyone so hyped! AI is the future! Amazing! Look at those stocks going up! We will make billions!4 5

Of course, excited people are always more vocal compared with the ones not joining the hype train, so this is (hopefully) just a feeling and my concern is not justified. Anyway here I am, worried, nevertheless.

Other potentially scariest answer(s)

If marketing is one big reason, we can speculate for hours about what is passing inside the brain of billionaires, CEOs and everyone convinced that replacing the workforce with AI agents, bots and tools would be a wonderful idea.

There are several counterarguments and digressions that can stem from this idea, but even assuming that all of this isn’t science fiction, that every job doable with a computer could be replaced by an AI tool (already quite a big stretch), in this amazing future we are describing, there are a couple of tiiiny problems. To mention just a few of them:

Ok. Now, let’s put all these digressions aside and let’s assume that - like all trends - this too shall pass, and that this is only a phase. Companies will understand that putting AI everywhere is not profitable at all, the AI bubble will burst (one can only hope). Let’s move on.

If we are also ignoring this point of view, what is then my main concern with people using personal AI tools for daily activities?

Final boundaries and assumptions

Several disclaimers follow! The entire topic is so broad to discuss that I must set up some boundaries to restrict the scope of this, and then make some assumptions, to set the tone.

Finally, to reduce and simplify the topic I will also make some big assumptions (that are not true at all), that would lead to another direction. Here follows a list of very negative things about the current AI world that I will deliberately ignore and forget for the rest of this post.

All disclaimers done, let’s dive directly into my main (and I guess controversial?) point.

Always more stupid people10

I am so tired of people using AI tools on a daily basis without giving it a second thought. In general. From the simple things that require just a small fraction of thinking11 to more complex tasks that require deep reasoning and creativity.

Expected rebuttal: “but wait, Alberto, AI is a tool, it depends on how you use it, it has a lot of potential if used properly!”

I will go into this topic later, I agree partially. But still, my thought is that:

people using AI tools often to produce something
are getting more stupid day after day

“Wow, chill out! Strong opinion, you are exaggerating, as always!”

Yes, absolutely, this post is a concerned rant about the whole AI hype. I am not writing for the first time in years to provide a moderate opinion. That would be a waste of time.

I also think that AI usage will bring a net negative to the entire world on the long run, on a greater measure when compared to other “innovative things” of the past decade that were hyped too much, that brought nothing or close to it in return (blockchain, IoT, domotics12, NFT
)

To dig into the counterargument above is: treating AI as a tool like the others, adopting the proper mindset, keeping a critical eye while using it, should be fine, right? I partially agree with this, I can see some utility, but my fear is: even a person using it with moderation and with the proper mindset, on the long run might get used to it (read: slowly addicted) and in the end using it for every single thing, daily.

That being said, the section below mentions some usage examples that I can understand, and in which I can see some benefits, if done with care.

“It’s a tool, it depends on how you use it”

I will start by immediately making a counterargument to myself. I believe there are cases in which AI tools might be helpful in doing some tasks, but with some big caveats. Let’s make some concrete examples, all related with software and coding13.

There are for sure many other scenarios in which AI can help, but notice how I am avoiding the elephant in the room: writing code or creating something from scratch.

From casual usage to replacing your ability to think

I wrote above that “people are getting more and more stupid”, then I made some examples that are “reasonable usage of AI”. Am I contradicting myself? Isn’t my main opinion a bit too much? Nope.

We are getting to the core of my point: when we use AI to do pure creative work the entire thinking process is being outsourced from our brain to a tool, and this is where I see really big risks.

We can all agree that producing something (piece of code, book, post, message, mail, presentation) is difficult. It requires focus, reasoning, critical thinking, and being able to iteratively refine something, until you are satisfied with the final result. The very process of working on something makes you better at it15.

Here I mean it in a really broad sense, you are getting smarter in something by the process of doing it. “Practice makes perfect” is very true. If you take away the time used for refining something, understanding why it’s not working, trying to fix it and then perfectioning it, you take away the experience that will make you better at it16.

But then, introducing human laziness, the very thing that pushed us forward as a species, that has also big downsides. We are all really good at being lazy. Human progress is often justified by “this will make us use less energy in some way”. That’s great. That’s progress. We don’t hunt anymore, we can grow crops next to our village, we live in an industrial automation world, and so on. You get the point I am making.

But now, it seems to me we are rushing toward the new worrying era of the “let’s also reduce how much we need to think, after all that requires a lot of energy!”17

“But wait, wait, wait! Calculators, computers, smartphones, there are so many examples of things that makes us think ‘less’ and are now normal tools! You really are an old man, Alberto, you are complaining about the progress!” đŸ‘ŽđŸ»

Maybe. Of course that’s a valid counterpoint. It’s also true that no longer doing math operations “by hand” or no longer writing with pen and paper is making us less proficient at those things, as some area of the brain is no longer activated. And don’t get me wrong, I am the first one to say that this is not bad, it’s making our life easier.

But, while solving a math operation with a calculator saves time, allowing for more thinking, having all the thinking done by AI tools is offsourcing several skills that are tied with producing a good result. Being able to convey properly ideas, refine them, use introspection and self-criticism before presenting a final result is completely overridden by the usage of an AI tool, while also stealing the entire learning experience.

Write a prompt, check the output summarily, write another prompt, check again the output, rewrite the prompt, and so on. The result of your work is no longer “yours”, you just pull levers and push buttons convincing a black box to vomit something that seems like a good enough result, obtaining then an output that you like and that you can appropriate as yours.

Do we really want to outsource our ability to think and to write to an external tool? We are already losing skills on several fronts18. We are spoiled by technology, that’s not bad per-se, but the trade-off is between obtaining comfort with less energy spent and being less and less autonomous in everything.

The dystopia on the spaceship in Wall-E was laughable at, quite funny and seemed so far away: humans unable to walk, being pampered by an AI entity while being fed entertainment and spending their entire existence like that. Nowadays it’s making me laugh a bit less.

Not everyone enjoys the process

My daily job involves a lot of thinking and writing. I write tasks for other people, I write documentation, I communicate with other people trying to convey my thoughts in a clear manner. It’s time consuming, it requires energy, but I like it. I would not trade it off only because an AI tool is faster than me. I don’t see in it any “progress”, but a terrible way of losing skills, while a company justifies the change with “increased efficiency” or “more profit”, as always19.

The things I wrote will not resonate with everyone of you. A lot of people don’t like their job or simply don’t care about it. If thinking time can be avoided - therefore using less energy - why not? Why feel guilty about it? If this is your opinion, I cannot do anything to change that. I am simply trying to provide my perspective, and I can only hope that the bleak future I am writing about, where people are unable to do things autonomously and they use their time to write prompts for AI agents and tools, is really really far away.

Now what? Why this post?

I don’t expect to fix the world with a post, but the mixture of corporations pushing AI tools everywhere and the amount of people lost in the hype is really concerning.

I am not using any AI tools, at all. My suggestion is for you to do the same.

If not for my main point (getting dumber), you can read again what I listed in the assumptions section. These “let’s ignore that” are very real issues, already enough on their own to re-evaluate the usage of all this AI stuff that is being fed to us.

I don’t think I’ll change my mind on this soon. I enjoy writing and thinking, in general. The idea of having a “magic shortcut to do everything” is a strong temptation, but I believe that the very act of creating and doing things on our own is what makes us get better, and it’s one of the most important things that makes us humans.

Thank you for reading.

  1. Just going briefly into the topic, LLM (models) are the language models currently running under the hood of all modern chatbots (GPT, PaLM, Gemini, LLaMA). ‘AI’ is the “common term” nowadays to include all of them 

  2. Actually, I need to rant quite a lot. I wanted to write something about my feelings on the topic since a couple of years. Recently, seeing the general hype and the bubble not bursting (yet), I decided that it was time. My reticence in writing online (thing that I don’t do anymore) is also due to the idea of this post being digested by several AI crawlers for training. That’s going to happen for sure, and I don’t like the idea, but here we are, who knows how everything I wrote will be ingested and vomited out in some way. What a time to be alive! 

  3. That being said, how much AI is concretely used in the company, it depends a lot. I am pretty sure the majority of the places that really use something AI-adjacent is just paying a huge amount of money to OpenAI or someone similar to use theirs APIs and show customers how cool their website chatbot is 

  4. About this, nowadays it’s extremely easy to use widely available AI tools to create fake accounts, and then generate and post comments/articles about how much AI is cool and the future. This is not an original idea, a theory about this, called the dead internet theory was already around in 2016. Today, with chatbots always more and more accessible, it’s quite likely to encounter AI generated content, from recipes websites up to newspapers. Notice that this is defined as a “conspiracy theory” but there is an active discussion ongoing to remove the “conspiracy” part, that I totally understand (you should take a look at the long list of linked articles). Some people can still spot when something is not written by a human, but if AI writing “improves” to be indistinguishable, are we looking forward to an internet filled only of incorrect things and fake humans? 

  5. Shhh! Never ask “yes, but how are we going to make billions?” to these companies, it’s not a nice question! The answer is obvious, no? With AI! Step 1: AI - Step 2: ??? - Step 3: PROFIT! 

  6. Recommended related read, a 1986 short story titled The Feeling of Power by Asimov. Plot summary: in the distant future, humans live in a computer-aided society and have forgotten the fundamentals of mathematics, including even the rudimentary skill of counting
 

  7. I personally don’t believe that current LLMs and AI tools have any kind of consciousness, but we could open a long discussion on the interesting question “what are we then?” and dig deeper talking about Chinese rooms, stochastic parrots and so on for hours 

  8. As you can imagine, these opinions are negative. Big surprise. Let me simply tell a now obvious and quite sad thing: when in the past someone used to say “AI will replace jobs”, nobody thought that the first ones to go would have been the artistic ones (drawers, photographers, visual artists, designers
) 

  9. And I wonder what these super good corporations use all this data for. Simply selling them for profit to ads companies? Or to profile people? Tinfoil hat mode on: is it really a stretch to imagine a corporation injecting some prompt to willingly “polarize” some people, convincing them of something, maybe extremize some ideas? Sounds like a “conspiracy theory” but it’s really easy to implement, nowadays, with people using these tools 

  10. Don’t get me wrong here, stupidity is old as the first human! We had plenty of stupid people around before AI, and we will always have them until the last human dies. My concern is the overall trend and the fact that tools like these can speed up a lot how fast people are getting dumber 

  11. Some - unfortunately - real examples: looking up for information in a file, instead of simply using “find”; writing a mail to summarize a plan/list of things to do, for which it was quite crucial that the author of the mail had clear in their mind the overall activity 

  12. Domotics still lives on, but not in the form it was hyped about in the beginning. Also, if we consider it progress, lots of people have Amazon or Google smart devices at home with always-on microphones (Echo, Alexa), sooo useful for telling you the weather saving the 10 extra seconds to pick up a smartphone - but hey, at least you will have more time to doomscroll Instagram or TikTok for hours! Great for your mental health! 

  13. This is my job and what I know better, I am sure there are plenty of other cases in which AI tools might “make the job easier”, I’m not doubting that 

  14. Regex already exist for doing advanced transformations, but this unit tests example is based on a real feedback I got from a colleague. Doing a lot of changes over several unit tests, in slightly different ways depending on the test itself, is quite easy with a well defined prompt 

  15. See also constructivism, a learning theory. “[
] the learner faces a loss of equilibrium with their previous understanding, and this demands a change in cognitive structure. This change effectively combines previous and novel information to form an improved cognitive schema” 

  16. Another software engineering tale. When I started working, ~10 years ago, the first months were a lot of trial, debugging, errors, debugging, more errors, frustration, debugging, more errors, and so on, until the issue was identified and solved. This process is for sure demanding and frustrating, but it’s essential to build a forma mentis, learning to pin down a problem. A junior software engineer in today’s world might simply use ChatGPT (or Copilot or another AI tool) to do a task without any need to understand how something is working, like this:

    You can see the role of the developer as “the (stupid) hands”. It’s a mechanical part, writing a prompt, receiving an answer, copying and pasting stuff into a terminal, checking if it’s producing errors, and repeating the process until “something works”. If you like this idea, I don’t think I can change your mind on the topic, we are completely different people 

  17. On this, I am pretty sure we are already in this era since several years. Social networks, smartphone addiction, people reading fewer books, do I need to mention something else? I am just concerned that AI tools will be another really big nail in the coffin of human ingenuity 

  18. I am the first that feels discomfort if I have to go to a new physical place without using Google Maps/Waze or any navigation system, instead of trying it out on my own. This is another skill I feel we are losing, maybe less “essential”, but it’s another example 

  19. As a side-note, another potential reason for companies to have people use AI is to train models, and maybe in the future, pushing for a “replacement” agenda, using AI agents to replace workers. After all, companies are always about the usual buzzwords: efficiency, velocity, cost optimization
 I still believe this would not work at all, but nothing will stop ideas fostered by the old usual stupid corporate mechanismsÂ