Engaging with why we might actually want inflation of text:
1) For pedagogical or explanatory purposes. For example, if I were to write:
> āxāR,x^2ā„0
I've used 10 characters to say
> For every real number x, it's square is greater than or equal to zero
For a mathematician, the first is sufficient. For someone learning, the second might be better (and perhaps as expansion of 'real number' or that 'square' is 'multiplying it by itself').
2) To make sure everything is stated and explicit. "He finally did x" implies that something has been anticipated/worked on for awhile, but "after a period of anticipation he did x" makes it more clear. This also raises the question of who was anticipating, which could be made explicit too.
As someone who spends a lot of time converting specifications to code (and explaining technical problems to non-technical people), unstated assumptions are very prevalent. And then sometimes people have different conceptions of the unstated assumption (i.e. some people might think that nobody was anticipating, it just took longer than you'd expect otherwise).
So longer text might seem like a simple expansion, but then it ends up adding detail.
I definitely agree with the authors point, I just want to argue that having a text-expander tool isn't quite as useless as 'generate garbage for me'.
This image originally came out just around the time of ChatGPT release and captures it well: https://i.imgur.com/RHGD9Tk.png
The older I get the more concise I find myself (which is not to say I'm actually concise, as my comment history will demonstrate), but LLM's have really driven home just how much noise day to day communication involves. So much filler text.
It still surprises me when I see non-technical enthusiasts get excited about LLMs drafting almost useless copy or email or whatever. So much garbage text no one reads but has to be written for some reason. Its weird.
An LLM is effectively a compressed model of its input data.
Inference is then the decompression stage where it generates text from the input prompt and the compressed model.
Now that compressing and decompressing texts is trivial with LLMs, we humans should focus - in business at least - on communicating only the core of what we want to say.
If the argument to get a new keyboard is: "i like it", then this should suffice, for inflated versions of this argument can be trivially generated.
I consider inflation a double insult. (https://ruudvanasseldonk.com/2025/llm-interactions) It says "I couldn't be bothered to spend time writing this myself, but I'm expecting you to read all the fluff."
> That we are using LLMs for inflation should not be taken as a criticism of these wonderful tools. It might, however, make us consider why we find ourselves inflating content. At best weāre implicitly rewarding obfuscation and time wasting; at worst weāre allowing a lack of clear thinking to be covered up. I think weāve all known this to be true, but LLMs allow us to see the full extent of this with our own eyes. Perhaps it will encourage us to change!
Yeah, this is the problem. Wealth distribution stopped working sometime in the late 20th century and we're fighting each others for competitive advantages. That's the core of this phenomenon.
No one needs containers full of baby sized left shoes, but proof of work must be shown. So the leathers must be cut and shoes must be sewn, only to be left in the ever growing pile in the backyard. That's kind of wrong.
Long documents in business contexts that get summarized and go mostly unread are the byproduct of a specific and common level of trust and accountability in those contexts: people don't believe someone has done enough critical thinking or has a strong enough justification for a proposal unless they've put it on the page, but if it is on the page, it's assumed that it does in fact represent critical thinking and legitimate justification.
If trust was higher, shorter documents would be more desirable. If trust was lower, or accountability higher, summarization would be used a lot more carefully.
LLMs haven't changed anything in this regard except that they've made it extremely easy to abuse trust at that specific level. The long-term result will be that trust will fall in the general case, and people will eventually become more careful about using summarization. I don't think it will be long before productized AI used in business contexts will be pretrained/fine-tuned to perform a basic level of AI content detection or include a qualitative measure of information density by default when performing summarization.
Huh, this post is not what I thought it would be! Even after the first two paragraphs!
There's a line of thought which states that intelligence rhymes with compression: Identifying patterns allows better prediction, enables better compression of the data.
However, internally, LLMs typically do the opposite: Tokenization and vectorization multiply the bit rate of the input signal. Chain of thought techniques add a lot of extra text, further increasing the bit rate.
My PM said theyād written a bunch of tickets for a project yesterday morning that we hadnāt fully scoped yet. I was pleasantly surprised because I canāt complain if they are going to get ahead of things and start scaffolding tickets.
Of course when I went to read them they were 100% slop. The funniest requirement were progress bars for actions that donāt have progress. The tickets were, even if you assume the requirements werenāt slop, at least 15 points a piece.
But ok maybe with all of these new tools we can respond by implementing these insane requirements. The real problem is what this article is discussing. Each ticket was also 500-700 words. Requirements that boil down to a single if statement were described in prose. While this is hilarious the problem is it makes them harder to understand.
I tried to explain this and they just said āok fine rewrite them thenā. Which I did in maybe 15min because there wasnāt actually much to write.
At this point Iām at a loss for how to even work with people that are so convinced these things will save time because they look at the volume of the output.
> the load on my server is reduced
isn't this the opposite? Enabling compression will INCREASE the load on your server as you need more CPU to compress/decompress the data.
Perhaps we should judge the performance of an LLM by how well it can compress arbitrary information. A higher IQ would mean more compression, after all.
LLM Inflation is an interesting choice of terminology. Like with many things in our contemporary society there is a temptation to assign absolute, quantitative value to everyday concepts but realistically we should know this to be a fallacy. Many concepts actually have no "on paper" value but still manifest significant social value. Marketing is the typical example, and yet we don't refer to advertising as inflation (though maybe we should).
This concept probably applies to lots of work in the "AI" space right now. The idea of using huge amounts of compute to generate lifelike voices for LLMs comes to mind as being recently maligned (something many users may not want). Or people upset about getting AI summaries in search that they didn't ask for. And yet, swaths of capital has been invested in these ideas and perhaps its a worthwhile use of resources. I am not sure personally. Time will tell. But I suspect its more complicated than the author is implying here.
I think the usage of LLMs will push for a societal change in how we communicate.
instead of elongated sentences, we perhaps might start seeing an increase in just communicating through the minimum constructing points of whatever meaning we hope to convey, leaving the presentation work for the LLM on the receiving side
This type of verbiage inflation was happening in business all the time anyway. LLMs are just being used as a method for doing it faster.
I saw an interesting argument recently that the reason you get this type of verbose language in corporate settings is that English lacks a formal tense. Apparently it's much less common in languages that have one. But in corporate English the verbosity is used as a signal that you took time to produce the text out of respect for the person you're communicating with.
This of course now gets weird with LLMs because I doubt it can last as a signal of respect for very long when it just means you fed some bullet points to ChatGPT.
>Creating the necessary prose is torturous for most of us, so Bob fires up the LLM du jour, types in āPlease create a 4 paragraph long business case for my manager, explaining why I need to replace my old, slow computerā and copies the result into his email.
>Bobās manager receives 4 paragraphs of dense prose and realises from the first line that heās going to have to read the whole thing carefully to work out what heās being asked for and why. Instead, he copies the email into the LLM du jour and types at the start āPlease summarise this email for me in one sentenceā. The 4 paragraphs are summarised as āThe sender needs a new computer as his current one is old and slow and makes him unproductive.ā
Sam Altman actually had a concise tweet about this blog's topic (https://x.com/sama/status/1631394688384270336)
>something very strange about people writing bullet points, having ChatGPT expand it to a polite email, sending it, and the sender using ChatGPT to condense it into the key bullet points 2:42 PM Ā· Mar 2, 2023 Ā· 1.2M Views
I call BS.
The 4 paragraphs requirement was not introduced 'because LLM'. It was there all along for what just should have been 'gimme 2 -3 bullet points'. They wanted Bob to hold back on requesting the new machine he needed, not by denying his request openly, but by making the process convoluted. Now Bob can cut through the BS, they want to blame the LMM for wasting their time and resources? BS!
One of the things that makes me hopeful for the future of LLMs is precisely this: humans are needlessly verbose, and LLMs can cut through the crap.
I expect smaller models to become incrementally better at compressing what truly matters in terms of information. Books, reports, blog posts⦠all kinds of long-form content can be synthesized in just a few words or pages. Itās no wonder that even small LLMs can provide accurate results for many queries.
The example in the article does not look like LLM Inflation, but that LLM can't reduce the waste in a bureaucratic process.
Why choose the word "inflation" to mean the opposite of compression? If you said it to a stranger, they'd assume you mean the price of LLMs is going up due to scarcity. I would call this LLM fluffing or LLM decompression
Where I work, I do the opposite: I let my colleagues know that they should write much mess and much more concise.
I actually straight up reject it when text is too inflated, and I remind people that LLMs are available to expand on request.
I've seen this happen IRL.
My former (obviously) wannabe manager used GAI to pimp our CV's before sending out to clients, pretty sure they too consulted stupid to summarize on their end.
We've also had XML inflation for 30 years.
Good point to better use it for haikus. Not sure if it that good at it though.
The problem described in this post has nothing to do with LLMs. It has everything to do with work culture and bureaucracy. Rules and laws that don't make sense remain because changing it requires time, energy and effort that most people in companies have either tried and failed or don't care enough to make a change.
This is one example of the "horseless carriage" AI solutions. I've begun questioning further that actually we're going into a generation where a lot of the things we are doing now are not even necessary.
I'll give you one more example. The whole "Office" stack of ["Word", "Excel", "Powerpoint"] can also go away. But we still use it because change is hard.
Answer me this question. In the near future if we could have LLMs that can traverse to massive amount of data why do we need to make excel sheets anymore? Will we as a society continue to make excel spreadsheets because we want the insights the sheet provides or do we make excel sheets to make excel sheets.
The current generation of LLM products I find are horseless carriages. Why would you need agents to make spreadsheets when you should just be able to ask the agent to give you answers you are looking for from the spreadsheet.
> Brevity is the soul of wit
Now that pachinko machines can create lots of prose, maybe it's time to finally learn this lesson.
Bob is highly optimistic!
> At best weāre implicitly rewarding obfuscation and time wasting; at worst weāre allowing a lack of clear thinking to be covered up.
Most people don't think very clearly. That's why rhetoric is effective. That's why most communication is fluffy social signaling. You can give people great advice and their eyes glaze over because the words didn't fill them with emotion, or something, and they do the exact opposite.
No wonder LLMs get put to work playing that stupid game.
> Bob needs a new computer for his job. In order to obtain a new work computer he has to create a 4 paragraph business case explaining why the new computer will improve his productivity.
Is this situation in any way realistic one? Because the way companies work in my beck of woods, no one wants your 4 paragraph business case essay about computer. Like, it is funny anecdote.
But, in real world, at least in my experience, pretty much everyone preferred short for emails and messages. They would skim the long ones at best, especially in situation that can be boiled down to "Tom wants a new computer and is verbose about it".
The author made up a fake situation to drive a point that doesn't exist
> One of the signal achievements of computing is data compression
Ah, yes. It is an achievement in signals in a way.
> Bob needs a new computer for his job.... In order to obtain a new work computer he has to create a 4 paragraph business case explaining why the new computer will improve his productivity.
> Bobās manager receives 4 paragraphs of dense prose and realises from the first line that heās going to have to read the whole thing carefully to work out what heās being asked for and why. Instead, he copies the email into the LLM.... The 4 paragraphs are summarised as āThe sender needs a new computer as his current one is old and slow and makes him unproductive.ā The manager approves the request.
"LLM inflation" as a "bad" thing often reflects a "bad" system.
In the case described, the bad system is the expectation that one has to write, or is more likely to obtain a favorable result from writing, a 4 paragraph business case. Since Bob inflates his words to fill 4 paragraphs and the manager deflates them to summarise, it's clear that the 4 paragraph expectation/incentive is the "bad" thing here.
This phenomenon of assigning the cause of "bad" things to LLMs is pretty rife.
In fact, one could say that the LLM is optimizing given the system requirement: it's a lot easier to get around this bad framework.